Skip to main content

Home/ Document Wars/ Group items matching "Computer" in title, tags, annotations or url

Group items matching
in title, tags, annotations or url

Sort By: Relevance | Date Filter: All | Bookmarks | Topics Simple Middle
Gary Edwards

Free Online PDF to HTML5 Converter, convert pdf to html5 flip book | pubhtml5 - 1 views

  •  
    "PUB HTML5 automatically converts your legacy content to rich and interactive eBooks. Add interactivity, audios, videos, documents, HTML activities, assessments and more to provide a rich reading experience to your readers. PUB HTML5 enables you to convert your content only once and publish them to multiple platforms like iPad, Android and Windows 8 tablets, PC/Mac and industry standard formats like HTML5 and MOBI. For PUB HTML5, the operating steps of creating fabulous digital magazines are foolproof ones. We extol minimum efforts and maximum outputting effects. Just follow the right procedures of the software, and you can totally customize your own digital magazine on your IPAD. After you have converted your PDF files, with multiple Custom Setting buttons, you get the privilege to design your own digital magazines .You may chose the template you prefer; change the background image; insert rich media including audios, video, images; add links, etc. The whole process can be easily achieved within minutes. Converting your PDFs into HTML5 in order to create iPad magazines can be a simple and worthwhile experience following the right procedures. This video provides you with a step by step procedure on how to create iPad magazines from the very beginning. Even though the PDF is great for posting reading documents like manuals on a website, it can sometimes annoy and even deter your viewers. Public or shared computers may not have a PDF viewer installed or downloading a PDF might not agree with a user's browsing habits. In order to make material in a PDF more accessible to others, converting your PDF to HTML5 file may be an alternative to consider. You can convert PDF to HTML5 free by using the Export tool in PUB HTML5. This option lets you perform different types of on-the-fly PDF conversions. After you have personalized your digital creation by using PUB HTML5 on your PC, you may easily preview your digital work on your IPAD or any other electronic devices. You ma
Gary Edwards

Online brochure maker by pub HTML5 is available online for everyone - WhaTech - 0 views

  •  
    "The press release introduces detailed capabilities of PUB HTML5's online brochure maker. The software is developed for people to prepare brochures online easily. PUB HTML5 is a leading publishing destination suitable for business owners, non-profit organizations as well as individuals. Whether you have a small scale business or a huge multinational company, you can certainly depend on PUB HTML5 for digital publishing. It is an open access HTML5 Digital Publishing Platform that allows you to convert PDF, MS Office and Open Office documents to HTML5 and jQuery based publications at no cost at all. These features make your flipbook even more attractive for the customers. Anna Lee, the chief designer of PUB HTML5 has announced a new online brochure maker (http://pubhtml5.com/digital-brochures-maker). It would make brochure designing simpler for the users. With the help of this software, you will be able to customize your catalogs, brochures and magazines online. More than six lakh users rely on PUB HTML5 for online publishing. The content published using the software program is compatible with all leading computers, laptops, tablets and mobile interfaces. Operating Systems like Android, iOS and Windows 8 also support this software quite well."
Gary Edwards

Dropbox Slashes Its Price as the Cost of a Gigabyte Nears Zero | Business | WIRED - 0 views

  • how many gigabytes you can store, and at what price.
  • The cut brings Dropbox in line, once again, with rival services at its gargantuan competitors: Google and Microsoft. But Dropbox’s decision to bury the lead signals something more important about the business it’s in:
  • in the competitive market for file storing, syncing, and sharing, gigabytes don’t matter quite as much as they did in the past.
  • ...8 more annotations...
  • The game is all about what you can do with them.
  • ChenLi Wang, Dropbox’s head of product
  • So, if Dropbox isn’t really selling storage, then what is it selling? Services.
  • The competition becomes squarely about what each competitor can do, rather than how much users can upload.
  • That’s been the approach Microsoft has taken, says Michal Gideoni, director of product management for Office.
  • Gideoni describes storage for Microsoft as just one aspect of its “holistic” approach to the cloud, an approach anchored not by file-syncing but by Office 365, the online version of its iconic productivity software.
  • As at Dropbox, Gideoni talks in terms of workflow, of data on the move, not just of a box for holding data in place.
  • Dropbox for Business also offers deep integration with Office files, but so far those features are not available with the consumer version.
  •  
    "When I talk to folks at Dropbox, they're eager to tell me about how different people are using its file-sharing service: the musician, the photographer, the professor, the startup founder. They like to talk about new features, like password-protected links and the remote wipe tool that lets you remove files from a lost computer. But what they save for the end of our meeting, almost like an afterthought, are the two numbers that traditionally meant the most for a data storage service: how many gigabytes you can store, and at what price. As it turns out, these numbers look at lot better than they used to. On Wednesday, the company slashed the price of a gigabyte by 90 percent on Dropbox Pro, the paid version of its signature consumer product. Up until now, users paid $9.99 per month to store up to 100 gigabytes of data. Now, for that same price, they can store one terabyte. The cut brings Dropbox in line, once again, with rival services at its gargantuan competitors: Google and Microsoft. But Dropbox's decision to bury the lead signals something more important about the business it's in: in the competitive market for file storing, syncing, and sharing, gigabytes don't matter quite as much as they did in the past. The game is all about what you can do with them. "It's how you get the content in and out and how does it let you do the work you want to accomplish," says ChenLi Wang, Dropbox's head of product. "We want people to rely on Dropbox as the home for all their stuff as opposed to thinking of it as a fixed storage limit." What Dropbox Is Selling"
Paul Merrell

EUROPA - Press Releases - Antitrust: Commission opens proceedings against MathWorks - 1 views

  • Brussels, 1 March 2012 - The European Commission has opened a formal investigation to assess whether The MathWorks Inc., a U.S.-based software company, has distorted competition in the market for the design of commercial control systems by preventing competitors from achieving interoperability with its products. The Commission will investigate whether by allegedly refusing to provide a competitor with end-user licences and interoperability information, the company has breached EU antitrust rules that prohibit the abuse of a dominant position. The opening of proceedings means that the Commission will examine the case as a matter of priority. It does not prejudge the outcome of the investigation. The investigation follows a complaint alleging that MathWorks had refused to provide a competitor with end-user software licences and accompanying interoperability information for its flagship products "Simulink" and "MATLAB", thereby preventing it from lawfully reverse-engineering in order to achieve interoperability with these two products.
  • As in the Microsoft case (see IP/04/382 and MEMO/04/70 and MEMO/07/359), the issue of software interoperability is central to this investigation. The Commission's investigation will focus on whether MathWorks' behaviour has prevented competitors from achieving interoperability with its widely used products and thereby hindered competition in breach of Article 102 TFEU. In this context, it is recalled that the European Directive 2009/24/EC on the legal protection of computer programmes also aims to foster interoperability by allowing for reverse-engineering for interoperability purposes provided that the software at issue was lawfully acquired.
  • Background MathWorks' "Simulink" and "MATLAB" software products are widely used for designing and simulating control systems. Control systems are deployed in many innovative industries such as in cruise control or anti-lock braking systems (ABS) for cars. Article 102 TFEU prohibits the abuse of a dominant position which may affect trade and prevent or restrict competition. The implementation of this provision is defined in the Antitrust Regulation (Council Regulation No 1/2003) which can be applied by the Commission and by the national competition authorities of EU Member States. The Commission has informed MathWorks and the Member States' competition authorities that it has formally opened proceedings in this case.
  •  
    Commission v. MIcrosoft Redux.
Gary Edwards

XML Production Workflows? Start with the Web and XHTML - 0 views

  • Challenges: Some Ugly Truths The challenges of building—and living with—an XML workflow are clear enough. The return on investment is a long-term proposition. Regardless of the benefits XML may provide, the starting reality is that it represents a very different way of doing things than the one we are familiar with. The Word Processing and Desktop Publishing paradigm, based on the promise of onscreen, WYSIWYG layout, is so dominant as to be practically inescapable. It has proven really hard to get from here to there, no matter how attractive XML might be on paper. A considerable amount of organizational effort and labour must be expended up front in order to realize the benefits. This is why XML is often referred to as an “investment”: you sink a bunch of time and money up front, and realize the benefits—greater flexibility, multiple output options, searching and indexing, and general futureproofing—later, over the long haul. It is not a short-term return proposition. And, of course, the returns you are able to realize from your XML investment are commensurate with what you put in up front: fine-grained, semantically rich tagging is going to give you more potential for searchability and recombination than a looser, more general-purpose approach, but it sure costs more. For instance, the Text Encoding Initiative (TEI) is the grand example of pouring enormous amounts of energy into the up-front tagging, with a very open-ended set of possibilities down the line. TEI helpfully defines a level to which most of us do not have to aspire.[5] But understanding this on a theoretical level is only part of the challenge. There are many practical issues that must be addressed. Software and labour are two of the most critical. How do you get the content into XML in the first place? Unfortunately, despite two decades of people doing SGML and XML, this remains an ugly question.
  • Practical Challenges In 2009, there is still no truly likeable—let alone standard—editing and authoring software for XML. For many (myself included), the high-water mark here was Adobe’s FrameMaker, substantially developed by the late 1990s. With no substantial market for it, it is relegated today mostly to the tech writing industry, unavailable for the Mac, and just far enough afield from the kinds of tools we use today that its adoption represents a significant hurdle. And FrameMaker was the best of the breed; most of the other software in decent circulation are programmers’ tools—the sort of things that, as Michael Tamblyn pointed out, encourage editors to drink at their desks. The labour question represents a stumbling block as well. The skill-sets and mind-sets that effective XML editors need have limited overlap with those needed by literary and more traditional production editors. The need to think of documents as machine-readable databases is not something that comes naturally to folks steeped in literary culture. In combination with the sheer time and effort that rich tagging requires, many publishers simply outsource the tagging to India, drawing a division of labour that spans oceans, to put it mildly. Once you have XML content, then what do you do with it? How do you produce books from it? Presumably, you need to be able to produce print output as well as digital formats. But while the latter are new enough to be generally XML-friendly (e-book formats being largely XML based, for instance), there aren’t any straightforward, standard ways of moving XML content into the kind of print production environments we are used to seeing. This isn’t to say that there aren’t ways of getting print—even very high-quality print—output from XML, just that most of them involve replacing your prepress staff with Java programmers.
  • Why does this have to be so hard? It’s not that XML is new, or immature, or untested. Remember that the basics have been around, and in production, since the early 1980s at least. But we have to take account of a substantial and long-running cultural disconnect between traditional editorial and production processes (the ones most of us know intimately) and the ways computing people have approached things. Interestingly, this cultural divide looked rather different in the 1970s, when publishers were looking at how to move to digital typesetting. Back then, printers and software developers could speak the same language. But that was before the ascendancy of the Desktop Publishing paradigm, which computerized the publishing industry while at the same time isolating it culturally. Those of us who learned how to do things the Quark way or the Adobe way had little in common with people who programmed databases or document-management systems. Desktop publishing technology isolated us in a smooth, self-contained universe of toolbars, grid lines, and laser proofs. So, now that the reasons to get with this program, XML, loom large, how can we bridge this long-standing divide?
  • ...44 more annotations...
  • Using the Web as a Production Platform The answer, I think, is right in front of you. The bridge is the Web, a technology and platform that is fundamentally based on XML, and which many publishers are by now comfortably familiar with. Perhaps not entirely comfortably, but at least most publishers are already working with the Web; they already either know or have on staff people who understand it and can work with it. The foundation of our argument is this: rather than looking at jumping to XML in its full, industrial complexity, which seems to be what the O'Reilly-backed StartWithXML initiative[6] is suggesting, publishers instead leverage existing tools and technologies—starting with the Web—as a means of getting XML workflows in place. This means making small investments and working with known tools rather than spending tens of thousands of dollars on XML software and rarefied consultants. It means re-thinking how the existing pieces of the production toolchain fit together; re-thinking the existing roles of software components already in use. It means, fundamentally, taking the Web seriously as a content platform, rather than thinking of it as something you need to get content out to, somehow. If nothing else, the Web represents an opportunity to think about editorial and production from outside the shrink-wrapped Desktop Publishing paradigm.
  • Is the Web made of Real XML? At this point some predictable objections can be heard: wait a moment, the Web isn’t really made out of XML; the HTML that makes up most of the Web is at best the bastard child of SGML, and it is far too flaky/unstructured/underpowered to be taken seriously. We counter by arguing that although HTML on the Web exists in a staggering array of different incarnations, and that the majority of it is indeed an unstructured mess, this does not undermine the general principle that basic, ubiquitous Web technologies can make a solid platform for content management, editorial process, and production workflow.
  • With the advent of a published XML standard in the late 1990s came the W3C’s adoption of XHTML: the realization of the Web’s native content markup as a proper XML document type. Today, its acceptance is almost ubiquitous, even while the majority of actual content out there may not be strictly conforming. The more important point is that most contemporary Web software, from browsers to authoring tools to content management systems (from blogs to enterprise systems), are capable of working with clean, valid XHTML. Or, to put the argument the other way around, clean, valid XHTML content plays absolutely seamlessly with everything else on the Web.[7]
  • The objection which follows, then, will be that even if we grant that XHTML is a real XML document type, that it is underpowered for “serious” content because it is almost entirely presentation (formatting) oriented; it lacks any semantic depth. In XHTML, a paragraph is a paragraph is a paragraph, as opposed to a section or an epigraph or a summary.
  • n contrast, more “serious” XML document types like DocBook[8] or DITA-derived schemas[9] are capable of making semantic distinctions about content chunks at a fine level of granularity and with a high degree of specificity.
  • So there is an argument for recalling the 80:20 rule here. If XHTML can provide 80% of the value with just 20% of the investment, then what exactly is the business case for spending the other 80% to achieve that last 20% of value? We suspect the ratio is actually quite a bit steeper than 80:20 for most publishers.
  • IDML is a well thought-out XML standard that achieves two very different goals simultaneously: it preserves all of the information that InDesign needs to do what it does; and it is broken up in a way that makes it possible for mere mortals (or at least our Master of Publishing students) to work with it.
  • XHTML, on the other hand, is supported by a vast array of quotidian software, starting with the ubiquitous Web browser. For this very reason, XHTML is in fact employed as a component part of several more specialized document types (ONIX and ePub among them).
  • Why re-invent a general-purpose prose representation when XHTML already does the job?
  • It is worth pausing for a moment to consider the role of XHTML in the ePub standard for ebook content. An ePub file is, anatomically, a simply disguised zip archive. Inside the zip archive are a few standard component parts: there are specialized files that declare metadata about the book, and about the format of the book. And then there is the book’s content, represented in XHTML. An ePub book is a Web page in a wrapper.
  • To sum up the general argument: the Web as it already exists presents incredible value to publishers, as a platform for doing XML content management with existing (and often free) tools, and without having to go blindly into the unknown. At this point, we can offer a few design guidelines: prefer existing and/or ubiquitous tools over specialized ones wherever possible; prefer free software over proprietary systems where possible; prefer simple tools controlled and coordinated by human beings over fully automated (and therefore complex) systems; play to our strengths: use Web software for storing and managing content, use layout software for layout, and keep editors and production people in charge of their own domains.
  • Putting the Pieces Together: A Prototype
  • At the SFU Master of Publishing Program, we have been chipping away at this general line of thinking for a few years. Over that time, Web content management systems have been getting more and more sophisticated, all the while getting more streamlined and easier to use. (NB: if you have a blog, you have a Web content management system.) The Web is beginning to be recognized as a writing and editing environment used by millions of people. And the ways in which content is represented, stored, and exchanged online have become increasingly robust and standardized.
  • The missing piece of the puzzle has been print production: how can we move content from its malleable, fluid form on line into the kind of high-quality print production environments we’ve come to expect after two decades of Desktop Publishing?
  • Anyone who has tried to print Web content knows that the existing methods leave much to be desired (hyphenation and justification, for starters). In the absence of decent tools for this, most publishers quite naturally think of producing the print content first, and then think about how to get material onto the Web for various purposes. So we tend to export from Word, or from Adobe, as something of an afterthought.
  • While this sort of works, it isn’t elegant, and it completely ignores the considerable advantages of Web-based content management.
  • Content managed online is stored in one central location, accessible simultaneously to everyone in your firm, available anywhere you have an Internet connection, and usually exists in a much more fluid format than Word files. If only we could manage the editorial flow online, and then go to print formats at the end, instead of the other way around. At SFU, we made several attempts to make this work by way of the supposed “XML import” capabilities of various Desktop Publishing tools, without much success.[12]
  • In the winter of 2009, Adobe solved this part of the problem for us with the introduction of its Creative Suite 4. What CS4 offers is the option of a complete XML representation of an InDesign document: what Adobe calls IDML (InDesign Markup Language).
  • The IDML file format is—like ePub—a simply disguised zip archive that, when unpacked, reveals a cluster of XML files that represent all the different facets of an InDesign document: layout spreads, master pages, defined styles, colours, and of course, the content.
  • What this represented to us in concrete terms was the ability to take Web-based content and move it into InDesign in a straightforward way, thus bridging Web and print production environments using existing tools and skillsets, with a little added help from free software.
  • Such a workflow—beginning with the Web and exporting to print—is surely more in line with the way we will do business in the 21st century, where the Web is the default platform for reaching audiences, developing content, and putting the pieces together. It is time, we suggest, for publishers to re-orient their operations and start with the Web.
  • We would take clean XHTML content, transform it to IDML-marked content, and merge that with nicely designed templates in InDesign.
  • The result is an almost push-button publication workflow, which results in a nice, familiar InDesign document that fits straight into the way publishers actually do production.
  • Tracing the steps To begin with, we worked backwards, moving the book content back to clean XHTML.
  • The simplest method for this conversion—and if you want to create Web content, this is an excellent route—was to use Adobe’s “Export to Digital Editions” option, which creates an ePub file.
  • Recall that ePub is just XHTML in a wrapper, so within the ePub file was a relatively clean XHTML document. It was somewhat cleaner (that is, the XHTML tagging was simpler and less cluttered) than InDesign’s other Web-oriented exports, possibly because Digital Editions is a well understood target, compared with somebody’s website.
  • In order to achieve our target of clean XHTML, we needed to do some editing; the XHTML produced by InDesign’s “Digital Editions” export was presentation-oriented. For instance, bulleted list items were tagged as paragraphs, with a class attribute identifying them as list items. Using the search-and-replace function, we converted such structures to proper XHTML list and list-item elements. Our guiding principle was to make the XHTML as straightforward as possible, not dependent on any particular software to interpret it.
  • We broke the book’s content into individual chapter files; each chapter could then carry its own basic metadata, and the pages conveniently fit our Web content management system (which is actually just a wiki). We assembled a dynamically generated table of contents for the 12 chapters, and created a cover page. Essentially, the book was entirely Web-based at this point.
  • When the book chapters are viewed online, they are formatted via a CSS2 stylesheet that defines a main column for content as well as dedicating screen real estate for navigational elements. We then created a second template to render the content for exporting; this was essentially a bare-bones version of the book with no navigation and minimal styling. Pages (or even the entire book) can be exported (via the “Save As...” function in a Web browser) for use in either print production or ebook conversion. At this point, we required no skills beyond those of any decent Web designer.
  • Integrating with CS4 for Print Adobe’s IDML language defines elements specific to InDesign; there is nothing in the language that looks remotely like XHTML. So a mechanical transformation step is needed to convert the XHTML content into something InDesign can use. This is not as hard as it might seem.
  • Both XHTML and IDML are composed of straightforward, well-documented structures, and so transformation from one to the other is, as they say, “trivial.” We chose to use XSLT (Extensible Stylesheet Language Transforms) to do the work. XSLT is part of the overall XML specification, and thus is very well supported in a wide variety of tools. Our prototype used a scripting engine called xsltproc, a nearly ubiquitous piece of software that we found already installed as part of Mac OS X (contemporary Linux distributions also have this as a standard tool), though any XSLT processor would work.
  • In other words, we don’t need to buy InCopy, because we just replaced it with the Web. Our wiki is now plugged directly into our InDesign layout. It even automatically updates the InDesign document when the content changes. Credit is due at this point to Adobe: this integration is possible because of the open file format in the Creative Suite 4.
  • We wrote an XSLT transformation script[18] that converted the XHTML content from the Web into an InCopy ICML file. The script itself is less than 500 lines long, and was written and debugged over a period of about a week by amateurs (again, the people named at the start of this article). The script runs in a couple of seconds, and the resulting .icml file can then be “placed” directly into an InDesign template. The ICML file references an InDesign stylesheet, so the template file can be set up with a house-styled layout, master pages, and stylesheet definitions for paragraphs and character ranges.
  • The result is very simple and easy to use. Our demonstration requires that a production editor run the XSLT transformation script manually, but there is no reason why this couldn’t be built directly into the Web content management system so that exporting the content to print ran the transformation automatically. The resulting file would then be “placed” in InDesign and proofed.
  • It should be noted that the Book Publishing 1 proof-of-concept was artificially complex; we began with a book laid out in InDesign and ended up with a look-alike book laid out in InDesign. But next time—for instance, when we publish Book Publishing 2—we can begin the process with the content on the Web, and keep it there throughout the editorial process. The book’s content could potentially be written and edited entirely online, as Web content, and then automatically poured into an InDesign template at proof time. “Just in time,” as they say. This represents an entirely new way of thinking of book production. With a Web-first orientation, it makes little sense to think of the book as “in print” or “out of print”—the book is simply available, in the first place online; in the second place in derivative digital formats; and third, but really not much more difficult, in print-ready format, via the usual InDesign CS print production system publishers are already familiar with.
  • Creating Ebook Files Creating electronic versions from XHTML source is vastly simpler than trying to generate these out of the existing print process. The ePub version is extremely easy to generate; so is online marketing copy or excerpts for the Web, since the content begins life Web-native.
  • Since an ePub file is essentially XHTML content in a special wrapper, all that is required is that we properly “wrap” our XHTML content. Ideally, the content in an ePub file is broken into chapters (as ours was) and a table of contents file is generated in order to allow easy navigation within an ebook reader. We used Julian Smart’s free tool eCub[19] to simply and automatically generate the ePub wrapper and the table of contents. The only custom development we did was to create a CSS stylesheet for the ebook so that headings and paragraph indents looked the way we wanted. Starting with XHTML content, creating ePub is almost too easy.
  • today, we are able to put the process together using nothing but standard, relatively ubiquitous Web tools: the Web itself as an editing and content management environment, standard Web scripting tools for the conversion process, and the well-documented IDML file format to integrate the layout tool.
  • Our project demonstrates that Web technologies are indeed good enough to use in an XML-oriented workflow; more specialized and expensive options are not necessarily required. For massive-scale enterprise publishing, this approach may not offer enough flexibility, and the challenge of adding and extracting extra semantic richness may prove more trouble than it's worth.
  • But for smaller firms who are looking at the straightforward benefits of XML-based processes—single source publishing, online content and workflow management, open and accessible archive formats, greater online discoverability—here is a way forward.
  • Rather than a public-facing website, our system relies on the Web as a content management platform—of course a public face could easily be added.
  • The final piece of our puzzle, the ability to integrate print production, was made possible by Adobe's release of InDesign with an open XML file format. Since the Web's XHTML is also XML, is can be easily and confidently transformed to the InDesign format.
  • Furthermore, just to get technical for a moment, XHTML is extensible in a fairly straightforward way, through the common “class” attribute on each element. Web developers have long leveraged this kind of extensibility in the elaboration of “microformats” for semantic-web applications.[10] There is no reason why publishers shouldn’t think to use XHTML’s simple extensibility in a similar way for their own ends.
  • Using the Web as a Production Platform
  •  
    I was looking for an answer to a problem Marbux had presented, and found this interesting article.  The issue was that of the upcoming conversion of the Note Case Pro (NCP) layout engine to the WebKit layout engine, and what to do about the NCP document format. My initial reaction was to encode the legacy NCP document format in XML, and run an XSLT to a universal pivot format like TEI-XML.  From there, the TEI-XML community would provide all the XSLT transformation routines for conversion to ODF, OOXML, XHTML, ePUB and HTML/CSS. Researching the problems one might encounter with this approach, I found this article.  Fascinating stuff. My take away is that TEI-XML would not be as effective a "universal pivot point" as XHTML.  Or perhaps, if NCP really wants to get aggressive; IDML - InDesign Markup Language. The important point though is that XHTML is a browser specific version of XML, and compatible with the Web Kit layout engine Miro wants to move NCP to. The concept of encoding an existing application-specific format in XML has been around since 1998, when XML was first introduced as a W3C standard, a "structured" subset of SGML. (HTML is also a subset of SGML). The multiplatform StarOffice productivity suite became "OpenOffice" when Sun purchased the company in 1998, and open sourced the code base. The OpenOffice developer team came out with a XML encoding of their existing document formats in 2000. The application specific encoding became an OASIS document format standard proposal in 2002 - also known as ODF. Microsoft followed OpenOffice with a XML encoding of their application-specific binary document formats, known as OOXML. Encoding the existing NCP format in XML, specifically targeting XHTML as a "universal pivot point", would put the NCP Outliner in the Web editor category, without breaking backwards compatibility. The trick is in the XSLT conversion process. But I think that is something much easier to handle then trying to
  •  
    I was looking for an answer to a problem Marbux had presented, and found this interesting article.  The issue was that of the upcoming conversion of the Note Case Pro (NCP) layout engine to the WebKit layout engine, and what to do about the NCP document format. My initial reaction was to encode the legacy NCP document format in XML, and run an XSLT to a universal pivot format like TEI-XML.  From there, the TEI-XML community would provide all the XSLT transformation routines for conversion to ODF, OOXML, XHTML, ePUB and HTML/CSS. Researching the problems one might encounter with this approach, I found this article.  Fascinating stuff. My take away is that TEI-XML would not be as effective a "universal pivot point" as XHTML.  Or perhaps, if NCP really wants to get aggressive; IDML - InDesign Markup Language. As an after thought, i was thinking that an alternative title to this article might have been, "Working with Web as the Center of Everything".
  •  
    I was looking for an answer to a problem Marbux had presented, and found this interesting article.  The issue was that of the upcoming conversion of the Note Case Pro (NCP) layout engine to the WebKit layout engine, and what to do about the NCP document format. My initial reaction was to encode the legacy NCP document format in XML, and run an XSLT to a universal pivot format like TEI-XML.  From there, the TEI-XML community would provide all the XSLT transformation routines for conversion to ODF, OOXML, XHTML, ePUB and HTML/CSS. Researching the problems one might encounter with this approach, I found this article.  Fascinating stuff. My take away is that TEI-XML would not be as effective a "universal pivot point" as XHTML.  Or perhaps, if NCP really wants to get aggressive; IDML - InDesign Markup Language. The important point though is that XHTML is a browser specific version of XML, and compatible with the Web Kit layout engine Miro wants to move NCP to. The concept of encoding an existing application-specific format in XML has been around since 1998, when XML was first introduced as a W3C standard, a "structured" subset of SGML. (HTML is also a subset of SGML). The multiplatform StarOffice productivity suite became "OpenOffice" when Sun purchased the company in 1998, and open sourced the code base. The OpenOffice developer team came out with a XML encoding of their existing document formats in 2000. The application specific encoding became an OASIS document format standard proposal in 2002 - also known as ODF. Microsoft followed OpenOffice with a XML encoding of their application-specific binary document formats, known as OOXML. Encoding the existing NCP format in XML, specifically targeting XHTML as a "universal pivot point", would put the NCP Outliner in the Web editor category, without breaking backwards compatibility. The trick is in the XSLT conversion process. But I think that is something much easier to handle then trying to
Gary Edwards

Microsoft trounces pro-ODF forces in state battles over open document formats - 0 views

  • Microsoft trounces pro-ODF forces in state battles over open document formats Eric Lai and Gregg Keizer   document.write(''); if (navigator.userAgent.indexOf("Gecko")==-1) { document.write(''); } document.write(''); if (document.getElementById('dclk999')) { document.getElementById('dclk999').src = 'http://ad.doubleclick.net/adi/idg.us.cpw.desktopapplications/index;pos=imu;tile=3;sz=336x280;ord=' + ord + '?'; } document.write(''); if ((!document.images && navigator.userAgent.indexOf('Mozilla/2.') >= 0)|| navigator.userAgent.indexOf("WebTV") >= 0) {document.write('');} June 03, 2007 (Computerworld)
  • Keeping it private The other problem, Mathers said, was the jargon-laden disinformation that committee members felt they were being fed by lobbyists for both IBM and Microsoft. Although lobbyists would tell the committee one thing in private, they got cold feet when asked to verify the information publicly, under oath. "Suddenly, nobody wanted to sign witness affirmation forms and testify," he said. That undermined the credibility of each side, but it particularly damaged the position of ODF proponents. After Wyne testified publicly that in Massachusetts, only a handful of computers had thus far been converted over to using ODF, IBM declined to dispute her claims, Mathers said -- despite having earlier given "gleaming" reports on the progress of ODF in Massachusetts. "That's when I really started to question the whole bill," he said.
  •  
    Uh Oh.  They got IBM dead to rights in Massachusetts.  I guess the truth about Massachusetts will be told!  Finally.  ODF failed in Massachusetts because there isn't a reasonable  means of implementing ODF.  Same in California.
  •  
    A must read.  And yes, the ODF Vendors are the reason ODF lost.  They didn't provide useful solutions.  In fact, the applicaitons they proposed were seen by government CIO's as cstly and disruptive "rip out and replace" non starters.  In California, CIO's asked if it was even possible to implement ODF!!!
Gary Edwards

But can money buy love? :: Another Microsoft Sponsored OOXML Study - 0 views

  •  
    Joe Wilson of Microsoft Watch knocks another one out of the park. Why is it that so few in the media get it? Or anyone else for that matter? Matt Assay gets it. But few understand the Vista Stack and the importance of OOXML in the transition of the monopoly base from MSOffice to the Vista Stack. No doubt the arrogance of those who dare challenge Microsoft is both a necessary blessing and guaranteed curse. Take for instance the widely held assumption that Microsoft invented MS-XML (OfficeOpenXML) in response to OpenDocument (ODf). This is false, misleading and will inevitably result in a FOSS death spiral in the face of a Vista Stack juggernaut. But it sure does feel good.

    Joe Wilson at Microsoft Watch points out the real reason for MS-XML, and why ISO approval of OOXML is so important. Microsoft needs OOXML approved as an international standard because OOXML is the binding model for the emerging Vista Stack of loosely coupled but information integrated applications.

    The Vista Stack model converges desktop, server, device and web information systems using OOXML-Smart Documents, .NET 3.0 and the XAML presentation layer as the binding components.

    The challenge for Microsoft is to migrate existing MSOffice bound business processes, line of business integrated apps, and advanced add-ons to the Exchange/SharePoint Hub. Once the existing documents, applications (MSOffice) and processes are migrated to the E/S Hub, they can be bound tightly to the rest of the Vista Stack.

    Others see OOXML as some sort of surrender or late recognition that the salad days of MSOffice are over. They jubilantly point to Web 2.0, Office 2.0 and rise of the LiNUX Desktop as having ushered in this end of monopoly for MSOffice. Like the ODf champions, these people are similarly sadly mistaken!

    While they celebrate, Microsoft is quie
Gary Edwards

Slamming the door shut on MS OOXML - 0 views

  • So your goal is a networked world where metadata is routinely trashed by apps developed by those who are too dumb or otherwise disabled to preserve metadata and only the big boys get to do interoperability, right? So if I send you a document for your editing, I can't count on getting it back with xml:id attributes intact. No thanks, Patrick. That sounds way too much like how things have worked ever since office productivity software first came on the market. In your world, interoperability belongs only to those who can map features 1:1 with the most featureful apps. And that is precisely why OpenDocument never should have been approved as a standard. Your kind of interoperability makes ODF a de facto Sun Microsystems standard wearing the clothing of a de jure standard. Why not just standardize the whole world on Microsoft apps and be done with it? Are two monopolies maintained by an interoperability barrier between them better than one? Fortunately, we don't have to debate the issue because the Directives resolve the issue. You lose under the rules of the game.
  •  
    Marbux on metadata and the language of universal interoperability: Few people are aware of the raging debate that has pushed ODF to the edge. The OASIS ODF TC is split between those who support Universal Interoperability, and those who insist on continuing with limited ODF interoperability.

    ODF (OpenDocument), formally known as Open Office XML, began it's standards life in the fall of 2002 when Sun submitted the OpenOffice file format to OASIS for consideration as a office suite XML fiel format standard. The work on ODF did not start off as a clean slate in that there were near 600 pages of application specific specification from day one of the standards work. The forces of universal interop have sought for years to separate ODF from the application specific features and implementation model of OpenOffice that began with those early specification volumes, and continues through the undue influence Sun continues to have over the ODF specification work.

    Many mistakenly believed that submission of ODF to ISO and subsequent approval as an international standard would provide an effective separation, putting ODF on the track of a truly universal file format.

    Marbux is one of those Universal Interop soldiers who has dug in his heels, cried to the heavens that enough is enough, and demanded the necessary changes to ODF interoperability language.

    This post he recently submitted to the OASIS ODF Metadata SC is a devastating rebuttal to the arguments of those who support the status quo of limited interoperability.

    In prior posts, marbux argues that ISO directives demand without compromise universal interoperability. This demand is also shared by the World Trade Organization directives regarding international trade laws and agreements. Here he brings those arguments together with the technical issues for achieving universal interop.

    It's a devastating argument.

Gary Edwards

The Age of OOXML Computing - thanks a pant load Sun! - 0 views

  • Why does Microsoft want another standard, what's the rationale? There are at least 4 good reasons why: *ODF started out and was completed as an XML format, specifically supporting OpenOffice with a tight scope around that product. *It wasn't until 2005 that the spec was offered up as a general XML office document format and consequently renamed to ODF. *No opportunity existed for Microsoft to actually participate in this full process - given the original scope, the 6 months between the re-naming of the spec to ODF, and its subsequent approval by OASIS as a standard. *The scope of the ODF spec never included even the basic requirements that Microsoft required to support a fully open format, and nor did the OASIS technical committee want to include these requirements.
  •  
    Erwin's StarOffice Tango has an exhaustive response to this Microsoft Q&A. Correcting false statements by Microsoft
Gary Edwards

FAA May Ditch Microsoft's Windows Vista And Office For Google And Linux Combo - Technology News by InformationWeek - 0 views

  • Bowen's compatibility concerns, combined with the potential cost of upgrading the FAA's 45,000 workers to Microsoft's next-generation desktop environment, could make the moratorium permanent. "We're considering the cost to deploy [Windows Vista] in our organization. But when you consider the incompatibilities, and the fact that we haven't seen much in the way of documented business value, we felt that we needed to do a lot more study," said Bowen. Because of Google Apps' sudden entry into the desktop productivity market
  •  
    The FAA issues their "NO ViSTA" mandate, hinting that it might be permanent if they can come up with MSOffice alternatives.  They are looking at Google Apps!

    Okay, so plan B does have legs.  The recent failure of ISO/IEC to stand up to the recidivist reprobate from Redmond is having repercussions.  Who would have ever thought ISO would fold so quickly without ceremony?  One day there are 20 out of 30 JTCS1 national bodies (NB's) objecting to Micrsoft's proprietary XML proposal, the MOOX Ecma 376 specfication, and the next ISO is approving without comment the placing of MOOX into the ISO fast track where approval is near certain.  With fast track, the technical objections and contradictions are assumed to be the provence of Ecma, and not the JTCS1 experts group.

    Apparently the USA Federal Government divisions had a plan B contingency for just such a case.  And why not?  Microsoft was able to purchase a presidential pardon for their illegal anti trust violations.  If they can do that, what's to stop them from purchasing an International Standard?  Piece of cake!

    But Google Apps?  And i say that as one who uses Google Docs every day.

    The problem of migrating away from MSOffice and MOOX to ODF or some other "open" XML portable file format is that there are two barriers one must cross.

    The first barrier is that of converting the billions of MS binary docuemnts into ODF XML. 

    The second is that of replacing the MSOffice bound business processes that drive critical day to day business operabions. 

    Google Apps is fine for documents that benefit from collaborative computing activities.  But there is no way one can migrate MSOffice bound business processes - the workgroup-worflow documents to Google Apps.  For one thing Google Apps is unable to facillitate important issues like XForms.  Nor can they round trip an ODF document with the needed fidelity a
Gary Edwards

Microsoft Suffers Latest Blow As NIST Bans Windows Vista - Technology News by InformationWeek - 0 views

  • In a new setback to Microsoft's public sector business, the influential National Institute of Standards and Technology has banned the software maker's Windows Vista operating system from its internal computing networks, according to an agency document obtained by InformationWeek.
  •  
    Excuse me!  Excuse me!  Does the right hand know what the left hand is doing?

    NiST, the National Institute of Standards and Technology, is authorized by the USA Department of Commerce. 

    Years ago, in conjunction with the Department of Defense (WWI), the Dept of Commerce joined with two manufacturing consortia to form ANSI.  Int eh aftemath of WWII and the formation of ISO/IEC, the US Congress, at the behest of the Department of Commerce, authorized the NiST subdiary.  NiST then authorized (and continues to oversee) ANSI to take on the USA representation at ISO/IEC.

    ANSI in turn authorized INCITS to take on the ISO/IEC document processing specific standardization issues.  It is INCiTS that represents the citizens on the ISO/IE SCT 1 workgroups (wk1) responsible for both ISO 26300 (OpenDocument - ODF) and Ecma 376 (MOOX).

    Okay, so now we have the technical staffers at NiST refusing to allow purchases of Vista, MSOffice 2007 and IE 7.0.  What's going on?  And why is this happenign near everywhere at this exact same moment in time?

    The answer is that this is clearly plan B. 

    Plan A was to force Microsoft to enable MSOffice native use of ODF.  The reasoning here is that governments could force Microsoft to implement ODF, the monopolist control over desktops would be broken, and the the threat of MS leveraging that monnopoly into servers, devices and Internet systems be averted.

    The key to this plan A was to mandate purchase requirements comply with Open Standards.  And not just any "Open Standards".  Microsoft had previously demonstrated how easy it was to use ECMA as rubber stamp for standards proposals that were anything but open.  This is why in August of 2004 the EU asked the OASIS ODF Technical Committee to submit ODF to ISO/IEC.  ISO had not yet been corrupted in the same way as the hapless money hungry ECMA.

    Plan A was going along
Gary Edwards

Microsoft playing three card monte with XML conversion, with Sun as the "outside man" working the mark (which is guess who?) - 0 views

  • In a highly informative post to his Open Stack blog Wednesday, Edwards explains how three key features are necessary for organizations to convert to open formats. These are: Conversion Fidelity - the billions of binaries problem Round Trip Fidelity - the MSOffice bound business processes, line of business integrated apps, and assistive technology type add-ons Application Interop - the cross platform, inter application, cross information domain problem
  •  
    Dana Blankenhorn posted this article back in March of 2007.  It was right at the time when the OASIS ODF TC and Metadata XML/RDF SC (Sub Committee) were going at it hammer and tong concerning three very important file format characteristics needed to fulfill a real world interoperability expectation:

    .... Compatibility - file format level interop -
    :::  backwards compatibility / compatibility with existing file formats, including the legacy of billions of binary Microsoft documents

    ....... Interoperability - application level interop-
    ::::::  application interoperability including interop with all Microsoft applications

Gary Edwards

Rough Type: Nicholas Carr's Blog: Fat Guy in Salesforce hell - Flock - 0 views

  • Second, don't underestimate the lock-in power that programs like Outlook and Excel and Quickbooks and Peachtree and their associated files still hold, particularly in smaller businesses. Someday we may have standard document formats and easily transportable data, but we don't yet. The competitive battle for the future of software is going to be fought out at the level of the Little Picture as much as at the level of the Big Picture. Lose sight of either one, and you'll be in trouble. In other words: It ain't over till the Fat Guy rants.
  •  
    Wow!  Another great quote from Nick.  When we were at the Office 2.0 Conference a few weeks ago, this was the problem every single collaborative computing initiative was facing.  Sure they had great collaborative efforts.  But these efforts were outside exisitng businesss processes and applications!  That's fine for kids and consumers.  But it's the kiss of death for enterprise, smb, and organizations with workgroup busines sprocesses based on MSOffice and Outlook.

    So no matter how innovative the WEb 2.0 - Office 2.0 - Enterprise 2.0 applications and services are, they are setting the marketplace for Microsoft to come in and take everything.  Because Microsoft and Microsoft alone ownes the interoperability - integration interfaces into MSOffice and Outlook, they are in a position to destroy any of the 2.0 players at will.  It's simply a matter of entering the space with their own 2.0 application or service.

    The more i see of this, the more convinced i am that the governemnts of the world are going to have to step in stop Microsoft's push to move from the desktop into server, device and web systems.

    ~ge~

Gary Edwards

INTERVIEW: Craig Mundie -- Microsoft's technology chief, taking over from Bill Gates - 0 views

  • In this exclusive interview with APC, Mundie says the notion of all software delivered entirely through the web browser is now widely recognised as being 'popular mythology'. He also stakes the claim that Google's existence and success was contingent on Microsoft creating Windows. He talks about what's coming down the pipeline for future versions of Windows, and his belief that Windows can get still more market share than it has today. He also discusses the issues around the recent controversy over the Office Open XML file format.
  • So Vista is in its diffusion cycle and until there is enough of it out there, you won't really see the developer community come across.
    • Gary Edwards
       
      Uh, the diffusion we really should be focused on involves the OOXML plug-in for MSOffice, IE 7.0, MSOffice 2007, and the Exchange/SharePoint Hub. 

      The Exchange/SahrePoint juggernaught is now at 65% marketshare, with Apache servers in noticeable decline.

      So it seems the improtant "diffusion" is going forward nicely.  The exploitation of the E/S Hub has also started, and here the Microsoft deelopers have an uncahllenged advantage.  Most of the business processes being migrated to the E/S Hub are coming off the MSOffice bound desktop.  Outsiders to the MS Stack do not have the requisite access to the internals that drive these MSOffice bound business processes, so they have little hope of getting into the "exploitation" cycle.

      This aspect was on full display at the recent Office 2.0 Conference in San Francisco.  The only way a O2 provider can position their service as a collaborative addon to existing business processes is to have some higher level of interop-integration into those processes beyond basic conversion to HTML.

      Most O2 operatives struggle to convince the market that an existing business process can be enhanced by stepping outside the process and putting the collaboration value elsewhere.  While this approach is disruptive and unfriendly, it tends to work until a more integrated, more interoperable coolaboration value becomes available.

      And that's the problem with O2.  Everyone is excited over the new collaboration possibilities, but the money is with the integration of collaborative computing into existing business processes.  This is a near impossible barrier for non Microsoft shops and would be competitors.  If you're Microsoft though, and you control existing formats, applications and processes, the collaboration stuff is simple value added on.  It's all low hanging fruit that Microsoft can get paid to deliver while O2 players struggle to f
  • So far, we have delivered about 60 million copies. That would represent about six per cent of the global Windows install base. So it has probably got to get up another few percentage points before you will start to see a big migration of the developer community.
    • Gary Edwards
       
      What is he talking about? Does a developer write to Vista? Or do they write to MS Stack ready .NET - OOXML-Smart Documents, XAML, Silverlight stuff?
  • ...2 more annotations...
  • Rather, what will happen is that you'll have, a seamless integration of locally running software in increasingly powerful client devices (not just desktops) and a set of services that work in conjunction with that. A lot of what we are doing with the Live platform not only allows us to provide the service component for our parts, but also gives the abilities for the developer community to perfect their composite applications and get them deployed at scale.
    • Gary Edwards
       
      Bear in mind that these "service components" are proprietary, and represent the only way to connect MS clients to the rest of the MS STack of applications.
  • Microsoft's business is not to control the platform per se, but in fact to allow it to be exploited by the world's developers. The fact that we have it out there gives us a good business, but in some ways it doesn't give us an advantage over any of the other developers in terms of being able to utilise it.
    • Gary Edwards
       
      Oh right! The anti trust restrictions will not be lifted until November. Have to be careful here. But how is it Craig that non Microsoft devlopers and service providers will be albe to access and interoperate with important "service components"?
  •  
    Great inteview, i'll comment as i make my way down the page.  Hopefully others will do the same.
Gary Edwards

Choose Your Future :: You're in good hands with Microsoft - 0 views

  • With ISO/IEC having standardized Open Document Format (ODF) 1.0 and considering standardizing Ecma Office Open XML (Open XML), there is a lot of public discussion about document formats and whether the world should begin to focus on only one format or continue to see multiple formats developed and used over time. Microsoft believes that users should be able to choose among formats and pick the one that best meets their needs. We also believe in encouraging the continued evolution of computing and data formats. And we support the ratification of Open XML in ISO/IEC. There are several reasons for these views:
  •  
    This the infamous Microsoft letter that many have misconstrued as an announcement of support for OpenDocument.  Only in a Microsoft universe can the statement, "We're not going to attack ODF" be taken as Micrsoft support for ODF.
Gary Edwards

Jeremy Allison: Einstein's definition of insanity... - 0 views

  • But standards don't rule the computing world. Today, ninety-two per cent of desktops and seventy per cent of servers run the proprietary and non-standardized Microsoft Windows OS.
Gary Edwards

A Fresh Cup » About - 0 views

shared by Gary Edwards on 09 Feb 08 - Cached
  • Then I got fed up with Microsoft. It took about fifteen years, but I finally reached the point where, no matter how good some of their software was, or how well-meaning some of the rank and file employees were, I could no longer stomach the corporate policies. You can read more about that in my posts What’s Going On Here?,  The Rest of the Story, and The Examined Software Life. Certainly, reasonable people can disagree on these things, but for me, working with Microsoft tools and technologies is no longer an option. So I quit.
    • Gary Edwards
       
      Same here. For me the 1996 JavaONE Conference was a great reunion of former Windows - Microsoft Developer Network members i had not seen since the old OS2-Windows Conference days. It was clear to me then that much of the Internet wave surge back then was charged by developers and investors dissaffected with Microsoft and the Windows monopoly. What people at that first JavaONE Confernece were looking for was a means to convert the Interent into a full fledged computational platform; augmenting the ubiquitous communications, access and exchange grid with web ready applications. They sought an alternative to Windows platform development. One of the more prominent memes back then was provided by the rather small but vocal FOSS developers community attending the conference. They had this strong belief that much of the value of the Internet was captured in this phrase, "Owned by none, used by all". For anyone familiar with the abusive monopolist's tactics, the meme was quite resonant. So much so that the openness of Java became a serious issue dooging Sun for years. I've always thought Sun paid an iincredible price for the reprehensible business practices of Microsoft. In early times, Java would have rode the same wave of application developer take up that Windows enjoyed in the early 90's. The monopolist however had changed the rules forever. And that's actually a good thing. Thanks for writing about your journey Mike. I'm looking forward to your blogervations.
  •  
    Not specific to Document formatting, but a huge example of Microsoft vs. "The right thing". Read the three articles linked in the post.
  •  
    Talk about "Switch", this is incredible. Very interesting fellow and a good
Gary Edwards

Prince: What's New - Docuemnt Publishing on Steroids with XHTML - CSS 3.0 (CDF+) - 0 views

  • Prince is a computer program that converts XML and HTML into PDF documents. Prince can read many XML formats, including XHTML and SVG. Prince formats documents according to style sheets written in CSS. Standards support HTML, CSS, SVG, MathML Web enabled Load documents, style sheets, images and fonts over HTTP Publishing features Hyphenation, crop marks, columns, page floats and footnotes Eye candy Rounded borders, small caps, CMYK and RGBA colors
Paul Merrell

Federal future cloudy for Microsoft Word -- Government Computer News - 0 views

  • “We have explained [to federal agencies] ways of moving from Microsoft Word to an i4i implementation of custom XML,” said Michael Vuple, i4i’s founder. “If agencies want custom XML, we are prepared, and we are working on a way for them to use our technology.” The company hasn’t been actively marketing that approach to government so far because it didn’t want to take advantage of the current “unfortunate situation,” he said. But with agencies likely to be asking the question, he said i4i will probably have to take a more proactive stance in the future.
« First ‹ Previous 41 - 60 of 76 Next ›
Showing 20 items per page