Skip to main content

Home/ Document Wars/ Group items matching "marbux" in title, tags, annotations or url

Group items matching
in title, tags, annotations or url

Sort By: Relevance | Date Filter: All | Bookmarks | Topics Simple Middle
Gary Edwards

Microsoft Will Support ODF! But Only If It Doesn't 'Restrict Choice Among Formats' - 0 views

  • By Marbux posted Jun 19, 2007 - 3:16 PM Asellus sez: "I will not say OOXML is easy to implement, but saying ODF is easier to implement just by looking at the ISO specification is a fallacy." I shouldn't respond to trolls, but I will this time. Asellus is simply wrong. Large hunks of Ecma 376 are simply undocumented. And what's more, absolutely no vendor has a featureful app that writes to that format. Not even Microsoft. There's a myth that Ecma 376 is the same as the Office Open XML used by Microsoft. It is not. I've spend a few hundred hours comparing the Ecma 376 specification (the version of OOXML being considered at ISO) to the information about the undocumented APIs used by MS Office 2007 that recently sprung loose in litigation. See http://www.groklaw.net/p...Rpt_Andrew_Schulman.pdf Each of those APIs *should* have corresponding metadata in the formats, but are not in the Ecma 376 specification.
  •  
    Incredible comment by Marbux!  With one swipe he takes out both Ecma 376 and ODF. 

    Microsoft has written a letter claiming that they will support ODF in MSOffice, but only if ISO approves Ecma 376 as a second office suite XML file format standard.  ODF was approved by ISO nearly a year ago.

    Criticizing Ecma 376 is easy.  It was designed to meet the needs of  a proprietary application, MSOffice, and, to meet the needs of the emerging MS Vista Stack of applications that spans desktop to server to device to web platforms.  It's filled with MS platform dependencies that make it impossibly non interoperable with anything not fully compliant with Microsoft owned API's.

    Criticizing ODF however is another matter entirely.  Marbux points to the extremely poor ODF interoperability record.  If MOOXML (not Ecma 376 - since that is a read only file format) is tied to vendor-application specific MSOffice, then ODF is similarly tied to the many vendor versions of OpenOffice/StarOffice.

    The "many vendor" aspect of OpenOffice is somewhat of a scam.  The interoperability that ODF shares across Novell Office, StarOffice, IBM WorkPlace, Red Office, and NeoOffice is entirely based on the fact that these iterations of OpenOffice are based on a single code base controlled 100% by Sun.  Which is exactly the case with MSOffice.  With this important exception - MOOXML (not Ecma 376) is interoperable across the entire Vista Stack!

    The Vista Stack is comprised of Exchange/SharePoint, MS Live, MS Dynamics, MS SQL Server, MS Internet Server, MS Grove, MS Collaboration Server, and MS Active Directory.   Behind these applications sits a an important foundation of shared assets: MOOXML, Smart Documents, XAML and .NET 3.0.  All of which can be worked into third party, Stack dependent applications through the Visual Studio .NET IDE.

    Here are some thoughts i wou
Gary Edwards

Slamming the door shut on MS OOXML - 0 views

  • So your goal is a networked world where metadata is routinely trashed by apps developed by those who are too dumb or otherwise disabled to preserve metadata and only the big boys get to do interoperability, right? So if I send you a document for your editing, I can't count on getting it back with xml:id attributes intact. No thanks, Patrick. That sounds way too much like how things have worked ever since office productivity software first came on the market. In your world, interoperability belongs only to those who can map features 1:1 with the most featureful apps. And that is precisely why OpenDocument never should have been approved as a standard. Your kind of interoperability makes ODF a de facto Sun Microsystems standard wearing the clothing of a de jure standard. Why not just standardize the whole world on Microsoft apps and be done with it? Are two monopolies maintained by an interoperability barrier between them better than one? Fortunately, we don't have to debate the issue because the Directives resolve the issue. You lose under the rules of the game.
  •  
    Marbux on metadata and the language of universal interoperability: Few people are aware of the raging debate that has pushed ODF to the edge. The OASIS ODF TC is split between those who support Universal Interoperability, and those who insist on continuing with limited ODF interoperability.

    ODF (OpenDocument), formally known as Open Office XML, began it's standards life in the fall of 2002 when Sun submitted the OpenOffice file format to OASIS for consideration as a office suite XML fiel format standard. The work on ODF did not start off as a clean slate in that there were near 600 pages of application specific specification from day one of the standards work. The forces of universal interop have sought for years to separate ODF from the application specific features and implementation model of OpenOffice that began with those early specification volumes, and continues through the undue influence Sun continues to have over the ODF specification work.

    Many mistakenly believed that submission of ODF to ISO and subsequent approval as an international standard would provide an effective separation, putting ODF on the track of a truly universal file format.

    Marbux is one of those Universal Interop soldiers who has dug in his heels, cried to the heavens that enough is enough, and demanded the necessary changes to ODF interoperability language.

    This post he recently submitted to the OASIS ODF Metadata SC is a devastating rebuttal to the arguments of those who support the status quo of limited interoperability.

    In prior posts, Marbux argues that ISO directives demand without compromise universal interoperability. This demand is also shared by the World Trade Organization directives regarding international trade laws and agreements. Here he brings those arguments together with the technical issues for achieving universal interop.

    It's a devastating argument.

Gary Edwards

XML Production Workflows? Start with the Web and XHTML - 0 views

  • Challenges: Some Ugly Truths The challenges of building—and living with—an XML workflow are clear enough. The return on investment is a long-term proposition. Regardless of the benefits XML may provide, the starting reality is that it represents a very different way of doing things than the one we are familiar with. The Word Processing and Desktop Publishing paradigm, based on the promise of onscreen, WYSIWYG layout, is so dominant as to be practically inescapable. It has proven really hard to get from here to there, no matter how attractive XML might be on paper. A considerable amount of organizational effort and labour must be expended up front in order to realize the benefits. This is why XML is often referred to as an “investment”: you sink a bunch of time and money up front, and realize the benefits—greater flexibility, multiple output options, searching and indexing, and general futureproofing—later, over the long haul. It is not a short-term return proposition. And, of course, the returns you are able to realize from your XML investment are commensurate with what you put in up front: fine-grained, semantically rich tagging is going to give you more potential for searchability and recombination than a looser, more general-purpose approach, but it sure costs more. For instance, the Text Encoding Initiative (TEI) is the grand example of pouring enormous amounts of energy into the up-front tagging, with a very open-ended set of possibilities down the line. TEI helpfully defines a level to which most of us do not have to aspire.[5] But understanding this on a theoretical level is only part of the challenge. There are many practical issues that must be addressed. Software and labour are two of the most critical. How do you get the content into XML in the first place? Unfortunately, despite two decades of people doing SGML and XML, this remains an ugly question.
  • Practical Challenges In 2009, there is still no truly likeable—let alone standard—editing and authoring software for XML. For many (myself included), the high-water mark here was Adobe’s FrameMaker, substantially developed by the late 1990s. With no substantial market for it, it is relegated today mostly to the tech writing industry, unavailable for the Mac, and just far enough afield from the kinds of tools we use today that its adoption represents a significant hurdle. And FrameMaker was the best of the breed; most of the other software in decent circulation are programmers’ tools—the sort of things that, as Michael Tamblyn pointed out, encourage editors to drink at their desks. The labour question represents a stumbling block as well. The skill-sets and mind-sets that effective XML editors need have limited overlap with those needed by literary and more traditional production editors. The need to think of documents as machine-readable databases is not something that comes naturally to folks steeped in literary culture. In combination with the sheer time and effort that rich tagging requires, many publishers simply outsource the tagging to India, drawing a division of labour that spans oceans, to put it mildly. Once you have XML content, then what do you do with it? How do you produce books from it? Presumably, you need to be able to produce print output as well as digital formats. But while the latter are new enough to be generally XML-friendly (e-book formats being largely XML based, for instance), there aren’t any straightforward, standard ways of moving XML content into the kind of print production environments we are used to seeing. This isn’t to say that there aren’t ways of getting print—even very high-quality print—output from XML, just that most of them involve replacing your prepress staff with Java programmers.
  • Why does this have to be so hard? It’s not that XML is new, or immature, or untested. Remember that the basics have been around, and in production, since the early 1980s at least. But we have to take account of a substantial and long-running cultural disconnect between traditional editorial and production processes (the ones most of us know intimately) and the ways computing people have approached things. Interestingly, this cultural divide looked rather different in the 1970s, when publishers were looking at how to move to digital typesetting. Back then, printers and software developers could speak the same language. But that was before the ascendancy of the Desktop Publishing paradigm, which computerized the publishing industry while at the same time isolating it culturally. Those of us who learned how to do things the Quark way or the Adobe way had little in common with people who programmed databases or document-management systems. Desktop publishing technology isolated us in a smooth, self-contained universe of toolbars, grid lines, and laser proofs. So, now that the reasons to get with this program, XML, loom large, how can we bridge this long-standing divide?
  • ...44 more annotations...
  • Using the Web as a Production Platform The answer, I think, is right in front of you. The bridge is the Web, a technology and platform that is fundamentally based on XML, and which many publishers are by now comfortably familiar with. Perhaps not entirely comfortably, but at least most publishers are already working with the Web; they already either know or have on staff people who understand it and can work with it. The foundation of our argument is this: rather than looking at jumping to XML in its full, industrial complexity, which seems to be what the O'Reilly-backed StartWithXML initiative[6] is suggesting, publishers instead leverage existing tools and technologies—starting with the Web—as a means of getting XML workflows in place. This means making small investments and working with known tools rather than spending tens of thousands of dollars on XML software and rarefied consultants. It means re-thinking how the existing pieces of the production toolchain fit together; re-thinking the existing roles of software components already in use. It means, fundamentally, taking the Web seriously as a content platform, rather than thinking of it as something you need to get content out to, somehow. If nothing else, the Web represents an opportunity to think about editorial and production from outside the shrink-wrapped Desktop Publishing paradigm.
  • Is the Web made of Real XML? At this point some predictable objections can be heard: wait a moment, the Web isn’t really made out of XML; the HTML that makes up most of the Web is at best the bastard child of SGML, and it is far too flaky/unstructured/underpowered to be taken seriously. We counter by arguing that although HTML on the Web exists in a staggering array of different incarnations, and that the majority of it is indeed an unstructured mess, this does not undermine the general principle that basic, ubiquitous Web technologies can make a solid platform for content management, editorial process, and production workflow.
  • With the advent of a published XML standard in the late 1990s came the W3C’s adoption of XHTML: the realization of the Web’s native content markup as a proper XML document type. Today, its acceptance is almost ubiquitous, even while the majority of actual content out there may not be strictly conforming. The more important point is that most contemporary Web software, from browsers to authoring tools to content management systems (from blogs to enterprise systems), are capable of working with clean, valid XHTML. Or, to put the argument the other way around, clean, valid XHTML content plays absolutely seamlessly with everything else on the Web.[7]
  • The objection which follows, then, will be that even if we grant that XHTML is a real XML document type, that it is underpowered for “serious” content because it is almost entirely presentation (formatting) oriented; it lacks any semantic depth. In XHTML, a paragraph is a paragraph is a paragraph, as opposed to a section or an epigraph or a summary.
  • n contrast, more “serious” XML document types like DocBook[8] or DITA-derived schemas[9] are capable of making semantic distinctions about content chunks at a fine level of granularity and with a high degree of specificity.
  • So there is an argument for recalling the 80:20 rule here. If XHTML can provide 80% of the value with just 20% of the investment, then what exactly is the business case for spending the other 80% to achieve that last 20% of value? We suspect the ratio is actually quite a bit steeper than 80:20 for most publishers.
  • IDML is a well thought-out XML standard that achieves two very different goals simultaneously: it preserves all of the information that InDesign needs to do what it does; and it is broken up in a way that makes it possible for mere mortals (or at least our Master of Publishing students) to work with it.
  • XHTML, on the other hand, is supported by a vast array of quotidian software, starting with the ubiquitous Web browser. For this very reason, XHTML is in fact employed as a component part of several more specialized document types (ONIX and ePub among them).
  • Why re-invent a general-purpose prose representation when XHTML already does the job?
  • It is worth pausing for a moment to consider the role of XHTML in the ePub standard for ebook content. An ePub file is, anatomically, a simply disguised zip archive. Inside the zip archive are a few standard component parts: there are specialized files that declare metadata about the book, and about the format of the book. And then there is the book’s content, represented in XHTML. An ePub book is a Web page in a wrapper.
  • To sum up the general argument: the Web as it already exists presents incredible value to publishers, as a platform for doing XML content management with existing (and often free) tools, and without having to go blindly into the unknown. At this point, we can offer a few design guidelines: prefer existing and/or ubiquitous tools over specialized ones wherever possible; prefer free software over proprietary systems where possible; prefer simple tools controlled and coordinated by human beings over fully automated (and therefore complex) systems; play to our strengths: use Web software for storing and managing content, use layout software for layout, and keep editors and production people in charge of their own domains.
  • Putting the Pieces Together: A Prototype
  • At the SFU Master of Publishing Program, we have been chipping away at this general line of thinking for a few years. Over that time, Web content management systems have been getting more and more sophisticated, all the while getting more streamlined and easier to use. (NB: if you have a blog, you have a Web content management system.) The Web is beginning to be recognized as a writing and editing environment used by millions of people. And the ways in which content is represented, stored, and exchanged online have become increasingly robust and standardized.
  • The missing piece of the puzzle has been print production: how can we move content from its malleable, fluid form on line into the kind of high-quality print production environments we’ve come to expect after two decades of Desktop Publishing?
  • Anyone who has tried to print Web content knows that the existing methods leave much to be desired (hyphenation and justification, for starters). In the absence of decent tools for this, most publishers quite naturally think of producing the print content first, and then think about how to get material onto the Web for various purposes. So we tend to export from Word, or from Adobe, as something of an afterthought.
  • While this sort of works, it isn’t elegant, and it completely ignores the considerable advantages of Web-based content management.
  • Content managed online is stored in one central location, accessible simultaneously to everyone in your firm, available anywhere you have an Internet connection, and usually exists in a much more fluid format than Word files. If only we could manage the editorial flow online, and then go to print formats at the end, instead of the other way around. At SFU, we made several attempts to make this work by way of the supposed “XML import” capabilities of various Desktop Publishing tools, without much success.[12]
  • In the winter of 2009, Adobe solved this part of the problem for us with the introduction of its Creative Suite 4. What CS4 offers is the option of a complete XML representation of an InDesign document: what Adobe calls IDML (InDesign Markup Language).
  • The IDML file format is—like ePub—a simply disguised zip archive that, when unpacked, reveals a cluster of XML files that represent all the different facets of an InDesign document: layout spreads, master pages, defined styles, colours, and of course, the content.
  • What this represented to us in concrete terms was the ability to take Web-based content and move it into InDesign in a straightforward way, thus bridging Web and print production environments using existing tools and skillsets, with a little added help from free software.
  • Such a workflow—beginning with the Web and exporting to print—is surely more in line with the way we will do business in the 21st century, where the Web is the default platform for reaching audiences, developing content, and putting the pieces together. It is time, we suggest, for publishers to re-orient their operations and start with the Web.
  • We would take clean XHTML content, transform it to IDML-marked content, and merge that with nicely designed templates in InDesign.
  • The result is an almost push-button publication workflow, which results in a nice, familiar InDesign document that fits straight into the way publishers actually do production.
  • Tracing the steps To begin with, we worked backwards, moving the book content back to clean XHTML.
  • The simplest method for this conversion—and if you want to create Web content, this is an excellent route—was to use Adobe’s “Export to Digital Editions” option, which creates an ePub file.
  • Recall that ePub is just XHTML in a wrapper, so within the ePub file was a relatively clean XHTML document. It was somewhat cleaner (that is, the XHTML tagging was simpler and less cluttered) than InDesign’s other Web-oriented exports, possibly because Digital Editions is a well understood target, compared with somebody’s website.
  • In order to achieve our target of clean XHTML, we needed to do some editing; the XHTML produced by InDesign’s “Digital Editions” export was presentation-oriented. For instance, bulleted list items were tagged as paragraphs, with a class attribute identifying them as list items. Using the search-and-replace function, we converted such structures to proper XHTML list and list-item elements. Our guiding principle was to make the XHTML as straightforward as possible, not dependent on any particular software to interpret it.
  • We broke the book’s content into individual chapter files; each chapter could then carry its own basic metadata, and the pages conveniently fit our Web content management system (which is actually just a wiki). We assembled a dynamically generated table of contents for the 12 chapters, and created a cover page. Essentially, the book was entirely Web-based at this point.
  • When the book chapters are viewed online, they are formatted via a CSS2 stylesheet that defines a main column for content as well as dedicating screen real estate for navigational elements. We then created a second template to render the content for exporting; this was essentially a bare-bones version of the book with no navigation and minimal styling. Pages (or even the entire book) can be exported (via the “Save As...” function in a Web browser) for use in either print production or ebook conversion. At this point, we required no skills beyond those of any decent Web designer.
  • Integrating with CS4 for Print Adobe’s IDML language defines elements specific to InDesign; there is nothing in the language that looks remotely like XHTML. So a mechanical transformation step is needed to convert the XHTML content into something InDesign can use. This is not as hard as it might seem.
  • Both XHTML and IDML are composed of straightforward, well-documented structures, and so transformation from one to the other is, as they say, “trivial.” We chose to use XSLT (Extensible Stylesheet Language Transforms) to do the work. XSLT is part of the overall XML specification, and thus is very well supported in a wide variety of tools. Our prototype used a scripting engine called xsltproc, a nearly ubiquitous piece of software that we found already installed as part of Mac OS X (contemporary Linux distributions also have this as a standard tool), though any XSLT processor would work.
  • In other words, we don’t need to buy InCopy, because we just replaced it with the Web. Our wiki is now plugged directly into our InDesign layout. It even automatically updates the InDesign document when the content changes. Credit is due at this point to Adobe: this integration is possible because of the open file format in the Creative Suite 4.
  • We wrote an XSLT transformation script[18] that converted the XHTML content from the Web into an InCopy ICML file. The script itself is less than 500 lines long, and was written and debugged over a period of about a week by amateurs (again, the people named at the start of this article). The script runs in a couple of seconds, and the resulting .icml file can then be “placed” directly into an InDesign template. The ICML file references an InDesign stylesheet, so the template file can be set up with a house-styled layout, master pages, and stylesheet definitions for paragraphs and character ranges.
  • The result is very simple and easy to use. Our demonstration requires that a production editor run the XSLT transformation script manually, but there is no reason why this couldn’t be built directly into the Web content management system so that exporting the content to print ran the transformation automatically. The resulting file would then be “placed” in InDesign and proofed.
  • It should be noted that the Book Publishing 1 proof-of-concept was artificially complex; we began with a book laid out in InDesign and ended up with a look-alike book laid out in InDesign. But next time—for instance, when we publish Book Publishing 2—we can begin the process with the content on the Web, and keep it there throughout the editorial process. The book’s content could potentially be written and edited entirely online, as Web content, and then automatically poured into an InDesign template at proof time. “Just in time,” as they say. This represents an entirely new way of thinking of book production. With a Web-first orientation, it makes little sense to think of the book as “in print” or “out of print”—the book is simply available, in the first place online; in the second place in derivative digital formats; and third, but really not much more difficult, in print-ready format, via the usual InDesign CS print production system publishers are already familiar with.
  • Creating Ebook Files Creating electronic versions from XHTML source is vastly simpler than trying to generate these out of the existing print process. The ePub version is extremely easy to generate; so is online marketing copy or excerpts for the Web, since the content begins life Web-native.
  • Since an ePub file is essentially XHTML content in a special wrapper, all that is required is that we properly “wrap” our XHTML content. Ideally, the content in an ePub file is broken into chapters (as ours was) and a table of contents file is generated in order to allow easy navigation within an ebook reader. We used Julian Smart’s free tool eCub[19] to simply and automatically generate the ePub wrapper and the table of contents. The only custom development we did was to create a CSS stylesheet for the ebook so that headings and paragraph indents looked the way we wanted. Starting with XHTML content, creating ePub is almost too easy.
  • today, we are able to put the process together using nothing but standard, relatively ubiquitous Web tools: the Web itself as an editing and content management environment, standard Web scripting tools for the conversion process, and the well-documented IDML file format to integrate the layout tool.
  • Our project demonstrates that Web technologies are indeed good enough to use in an XML-oriented workflow; more specialized and expensive options are not necessarily required. For massive-scale enterprise publishing, this approach may not offer enough flexibility, and the challenge of adding and extracting extra semantic richness may prove more trouble than it's worth.
  • But for smaller firms who are looking at the straightforward benefits of XML-based processes—single source publishing, online content and workflow management, open and accessible archive formats, greater online discoverability—here is a way forward.
  • Rather than a public-facing website, our system relies on the Web as a content management platform—of course a public face could easily be added.
  • The final piece of our puzzle, the ability to integrate print production, was made possible by Adobe's release of InDesign with an open XML file format. Since the Web's XHTML is also XML, is can be easily and confidently transformed to the InDesign format.
  • Furthermore, just to get technical for a moment, XHTML is extensible in a fairly straightforward way, through the common “class” attribute on each element. Web developers have long leveraged this kind of extensibility in the elaboration of “microformats” for semantic-web applications.[10] There is no reason why publishers shouldn’t think to use XHTML’s simple extensibility in a similar way for their own ends.
  • Using the Web as a Production Platform
  •  
    I was looking for an answer to a problem Marbux had presented, and found this interesting article.  The issue was that of the upcoming conversion of the Note Case Pro (NCP) layout engine to the WebKit layout engine, and what to do about the NCP document format. My initial reaction was to encode the legacy NCP document format in XML, and run an XSLT to a universal pivot format like TEI-XML.  From there, the TEI-XML community would provide all the XSLT transformation routines for conversion to ODF, OOXML, XHTML, ePUB and HTML/CSS. Researching the problems one might encounter with this approach, I found this article.  Fascinating stuff. My take away is that TEI-XML would not be as effective a "universal pivot point" as XHTML.  Or perhaps, if NCP really wants to get aggressive; IDML - InDesign Markup Language. The important point though is that XHTML is a browser specific version of XML, and compatible with the Web Kit layout engine Miro wants to move NCP to. The concept of encoding an existing application-specific format in XML has been around since 1998, when XML was first introduced as a W3C standard, a "structured" subset of SGML. (HTML is also a subset of SGML). The multiplatform StarOffice productivity suite became "OpenOffice" when Sun purchased the company in 1998, and open sourced the code base. The OpenOffice developer team came out with a XML encoding of their existing document formats in 2000. The application specific encoding became an OASIS document format standard proposal in 2002 - also known as ODF. Microsoft followed OpenOffice with a XML encoding of their application-specific binary document formats, known as OOXML. Encoding the existing NCP format in XML, specifically targeting XHTML as a "universal pivot point", would put the NCP Outliner in the Web editor category, without breaking backwards compatibility. The trick is in the XSLT conversion process. But I think that is something much easier to handle then trying to
  •  
    I was looking for an answer to a problem Marbux had presented, and found this interesting article.  The issue was that of the upcoming conversion of the Note Case Pro (NCP) layout engine to the WebKit layout engine, and what to do about the NCP document format. My initial reaction was to encode the legacy NCP document format in XML, and run an XSLT to a universal pivot format like TEI-XML.  From there, the TEI-XML community would provide all the XSLT transformation routines for conversion to ODF, OOXML, XHTML, ePUB and HTML/CSS. Researching the problems one might encounter with this approach, I found this article.  Fascinating stuff. My take away is that TEI-XML would not be as effective a "universal pivot point" as XHTML.  Or perhaps, if NCP really wants to get aggressive; IDML - InDesign Markup Language. As an after thought, i was thinking that an alternative title to this article might have been, "Working with Web as the Center of Everything".
  •  
    I was looking for an answer to a problem Marbux had presented, and found this interesting article.  The issue was that of the upcoming conversion of the Note Case Pro (NCP) layout engine to the WebKit layout engine, and what to do about the NCP document format. My initial reaction was to encode the legacy NCP document format in XML, and run an XSLT to a universal pivot format like TEI-XML.  From there, the TEI-XML community would provide all the XSLT transformation routines for conversion to ODF, OOXML, XHTML, ePUB and HTML/CSS. Researching the problems one might encounter with this approach, I found this article.  Fascinating stuff. My take away is that TEI-XML would not be as effective a "universal pivot point" as XHTML.  Or perhaps, if NCP really wants to get aggressive; IDML - InDesign Markup Language. The important point though is that XHTML is a browser specific version of XML, and compatible with the Web Kit layout engine Miro wants to move NCP to. The concept of encoding an existing application-specific format in XML has been around since 1998, when XML was first introduced as a W3C standard, a "structured" subset of SGML. (HTML is also a subset of SGML). The multiplatform StarOffice productivity suite became "OpenOffice" when Sun purchased the company in 1998, and open sourced the code base. The OpenOffice developer team came out with a XML encoding of their existing document formats in 2000. The application specific encoding became an OASIS document format standard proposal in 2002 - also known as ODF. Microsoft followed OpenOffice with a XML encoding of their application-specific binary document formats, known as OOXML. Encoding the existing NCP format in XML, specifically targeting XHTML as a "universal pivot point", would put the NCP Outliner in the Web editor category, without breaking backwards compatibility. The trick is in the XSLT conversion process. But I think that is something much easier to handle then trying to
Gary Edwards

Patent Ruling Against Microsoft Hinges on Meaning of Custom XML - 0 views

  •  
    Marbux discovered this gem, joining the argument with an insightful but disagreeing post.  I however agreed with the articles author, Jeff Cogswell, that both the judge and jury confused the XML pane feature set with metacode mapping claims in the now infamous i4i 449 patent.  If Marbux is right, then HTML-CSS, ODF, and RDF/XML-RDFa are also infringing on this patent.  Which i4i claims is not the case. Except: Here's one part of the ruling:  ...  Microsoft Corporation is hereby permanently enjoined from ... selling, offering to sell ... any Infringing and Future Word Products that have the capability of opening a .XML, .DOCX, or .DOCM file ("an XML file") containing custom XML. The odd wording here is "custom XML," which appears several times in the ruling. Based on the comments in response to eWEEK's articles on the ruling, as well as comments I've seen elsewhere, a great deal of people think the problem was that Microsoft uses XML as its format. But that isn't the case. The ruling focuses on the use of custom XML. The ruling is not about the fact that Word uses XML. If it did, there would be a worldwide disaster, considering how prevalent XML is. But what exactly is custom XML? To start with, let's look at the claims of the patent itself and try to make a connection. The patent, which was written back in 1994, covers a new way of providing formatting in a word processing program. To understand the claims of the patent, it's important to note the distinction between what the inventors call content and what are called metacodes (which are ultimately formatting codes).
Gary Edwards

EOOXML objections - Grokdoc - 0 views

  •  
    Marbux has done some heroic work here, using the GrokDoc Wiki.  The Title is "EOOXML Objections", and it's primary purpose is to help ISO National Body Memebers evaluate the 0ver 6,000 pages of the Microsoft - ECMA Office Open XML Specification for MSOffice. 

    On January 5th, 2007, Microsoft officially submitted EOOXML to ISO under the fast track rules.  Before EOOXML can hit the fast track though, ISO provides members with a 30 day "Contradition Review Phase".   During this brief phase, ISO NB's (national standards body members) muct evaluate the proposal and post their allegations concerning contradictions and inconsistencies with other ISO products - like ODF.

    What Marbux is assembling here is a one stop shop for ISO NB's strugglign to understand the issues at stake.  It's incredible wha the has accomplished in such a short time.  But then, the clock is ticking.  February 5th is a hard and unmovable deadline. 

    The basic contradiction is thatt EOOXML is a subset of ISO existing product, ODF.  Both attempt to do the exact same thing:  provide an XML file format for desktop productivity environments such as MSOffice, OpenOffice, and WordPerfect Office.  What seriously differentiates the two is that ODF was designed expressly to be a universal file format, application and platform independent, able to transition across many different information domains connecting the legacy of desktop productivity to near everything else.  MOOXML on the other hand was designed for MSOffice and the legacy of billions of binary documents that only Micrsoft knows the secrets to converting to XML.  As such, MOOXML is designed to be application and platform bound, with these proprietary dependencies written right into the specification.

    One of the more important elements of the Marbux arguments is that the OpenDocument Foundation's daVinci Plugin and InfoSet Engine - API prove conclusively
Gary Edwards

The real state of ODF Interoperability? There is none : Comments from the Northwest Progressive Institute Advocate Review of OpenOffice and ODF - 0 views

  •  
    Marbux nails it again in the comments section of this obscure review. In particular, he sites Shah, Rajiv C. and Kesan, Jay P., Lost in Translation: Interoperability Issues for Open Standards - ODF and OOXML as Examples (September 2008), Link to paper on SSRN (compatibility fidelity comparisons of ODF implementations testing only a very small set of word processing features). "...Switching documents, I go through similar travails with the published ODF 1.1 specification, using both the PDF and ODT versions. Bottom line: I can't get either document into WordPerfect X3 or X4 using any rich text format. So I convert the document to plain text using Symphony and get my work done. That is the real state of ODF interoperability. There is no such thing. But that does not stop the vested interests from claiming that there is. E.g.:"
Gary Edwards

The big winner from Apache OpenOffice.org | ITworld - Brian Proffitt - 1 views

  •  
    Brian is once again writing about OpenOffice and ODF, this time in the aftermath of Oracle's decision to cut OOo loose and turn it over to Apache instead of The Document Foundation.  Good discussion - features a lengthy comment from the mighty Marbux where he vigorusly corrects the river of spin coming out of IBM.  Worth a careful read! excerpt: IBM seems to maneuver itself to any open source project that suits its needs, and for whatever reason they have decided to hitch their wagon to Oracle's star (or vice versa). With this historical context, there is really little surprise in Oracle's decision to go with the Apache Software Foundation, because IBM was probably influencing the decision. My second question doesn't have a definitive answer--yet. But it needs to be answered. It is simply this: how will OpenOffice.org remain relevant to end users?
  •  
    I should have added to that comment a stronger warning for the Apache Foundation Board and developers considering joining the IBM-backed Apache OpenOffice.org incubator project in regard to the danger posed by IBM and Oracle's control of the OpenDocument Formats Technical Committee at OASIS, aptly characterized by IBM's Rob Weir: "Those who control the exchange format, can control interoperability and turn it on or off like a water faucet to meet their business objectives." Rob Weir, Those Who Forget Santayana, An Antic Disposition (20 December 2007), http://www.robweir.com/blog/2007/12/those-who-forget-santayana.html What IBM, Oracle, and others can do by manipulating the ODF specification that Apache OOo depends upon is something entirely outside the control of the Apache Foundation. And as history has taught us so well, IBM and Sun exercised that control mercilessly via their co-chairmanship of the ODF TC to block all real interoperability initiatives. That is the very reason that only ODF implementations that share the same code base can interoperate. And if one were tempted to think that IBM and Sun/Oracle would not even consider manipulating the ODF specification to their own commercial advantage, consider the fact that in writing the quoted statement above, Rob Weir was speaking from deep personal experience in in such activities. So beware, both Apache Foundation and LibreOffice developers.
Gary Edwards

Comments on 'On the Office format wars' - 0 views

  • A fatal flaw in your analysis By Marbux Posted Saturday 21st April 2007 08:15 GMT Your analysis contains a fatal flaw, Martin. That is your belief that adequate Microsoft XML <> OpenDocument translators will be available. In fact, all of the translators suck mightily and there is no prospect at all of them being perfected. The major problems are: (i) that Microsoft's XML formats seem deliberately designed to thwart their parsing with XPath, which is essential to XML transformations; (ii) that Microsoft's "XML" file formats include binary blobs, bitmasks, and multiple Windows and Microsoft dependendencies, all of which defy XML transformations; and (iii) OpenDocument assumes a richer page layout engine than Microsoft Word provides, so while DOCX can be completely mapped to ODT it is impossible to fully map in the other direction without declaring an MS Office interoperability subset of OpenDocument and ODF applications implementing a compatibility mode with reduced features. (That is more than somewhat ironic, given Microsoft's spin that it couldn't implement all of its features in OpenDocument. In fact, the exact opposite is true.) In fact, Steve Ballmer is on record as saying that the developers of the Novell-Microsoft-Clever Age plug-ins will not even attempt to achieve full fidelity file translations between the two formats. http://www.eweek.com/article2/0,1895,2050848,00.asp?kc=EWEWEMNL103006EP17A Those translators achieve at best far less conversion fidelity than existing file conversion filters between OpenDocument and Microsoft binary file formats such as the OpenOffice.org conversion filters, which achieve only about 80 per cent fidelity. The file format cognescenti know this. See e.g., the paper by Gary Edwards and Sam Hiser included in this edition of the European Journal for the Informatics Professional. http://www.upgrade-cepis.org/issues/2006/6/up7-6Hiser.pdf (PDF). (Note that I contributed to that paper.) And as also detailed in that paper, what works well enough for some of us does not necessarily work well enough for all. Anything less than full fidelity data conversions is absolutely unacceptable in the context of wholly automated business processes and is in fact illegal in various contexts, including government records. So your thesis doesn't fly. In fact, I'd go so far as to bet that you have been suckered by the Microsoft spin doctors. Another indication is your depiction of the file format wars as being waged primarily between IBM and Microsoft, a recent theme of Microsoft's public relations machine. While it is seductive to believe that the controversy is just another chapter in the war between major competitors, the pro-ODF camp is far broader than IBM. For example, nearly 20 governments recently opposed fast track processing of Microsoft's draft standard at ISO. Do you believe they were all carrying water for IBM? Government bodies in more than 50 nations have chosen to adopt ODF. http://opendocumentfellowship.org/government/precedent And dozens of developers now support the OpenDocument standard in their applications. http://opendocumentfellowship.org/applications While IBM has had a noteworthy role in proliferating the OpenDocument formats, there is a movement without a recognizable leader in the industry. When it comes to vendor influence on things relevant to ODF, Sun Microsystem's far outshines IBM. But in fact, a core group of open standards and free and open source developers and advocates -- inside and outside government -- have played a far larger role. This is a customer-driven phenomenon, not a vendor-driven effort as you portray. So I will respectfully suggest that you reexamine your position on these issues. Reasonable minds can differ, but not on the grounds you advocate.
  •  
    Here we go again.  A couple of boot lickin lackies at The Register make some moronic statements about the OpenDocument XML file format, and the portable document cognisceti experts come out of the wood work to set the record straight.  I think it's a scam to get boost hits. 

    Once again Marbux hands out a major bitch splappin to Microsoft shills who have no idea what's coming.  What a great job Marbox does, and does with a kind consideration that certainly isn't warranted given the idiocy of the main article.  Where does the man's patience come from?  I gave up long ago.

    ~ge~

Gary Edwards

What will it be for ODF? Continuation of limited interop? Or a transition to Universal Interoperability? - 0 views

    • Gary Edwards
       
      Preserving metadata! Preserving application specific information. Preserving "unknown" information inside of a document
  • Unless we add conformance requirements for the preservation of metadata and processing instructions, the less featureful apps will never  be able to round-trip documents with the more featureful apps. Our language should require that. Personally, I believe that the software-as-an-end-point client-side office suites are dinosaurs at the end of their era. They are being finished off by a thousand cuts as users spend less and less time using them and more and more time using other apps, such as web apps. ODF either develops methods for interoperability among all apps or it will die along with the office suites. E.g., Microsoft knows this and is busily migrating its Office development budget across the Sharepoint/Exchange server hubs to the network. Meanwhile, this TC fiddles with preserving the 1995 software-as-an-endpoint vision.
  •  
    Marbux is clearly at the top of his game here as he hammers the interoperability issue.
Gary Edwards

Linux Foundation Legal : Behind Putting the OpenDocument Foundation to Bed (without its supper) : Updegroove - 0 views

  • CDF is one of the very many useful projects that W3C has been laboring on, but not one that you would have been likely to have heard much about. Until recently, that is, when Gary Edwards, Sam Hiser and Marbux, the management (and perhaps sole remaining members) of the OpenDocument Foundation decided that CDF was the answer to all of the problems that ODF was designed to address. This announcement gave rise to a flurry of press attention that Sam Hiser has collected here. As others (such as Rob Weir) have already documented, these articles gave the OpenDocument Foundation’s position far more attention than it deserved. The most astonishing piece was written by ZDNet’s Mary Jo Foley. Early on in her article she stated that, “the ODF camp might unravel before Microsoft’s rival Office Open XML (OOXML) comes up for final international standardization vote early next year.” All because Gary, Sam and Marbux have decided that ODF does not meet their needs. Astonishing indeed, given that there is no available evidence to support such a prediction.
  •  
    Uh?  The ODF failure in Massachusetts doesn't count as evidence that ODF was not designed to be compatible with existing MS documents or interoperable with existing MSOffice applications?

    And it's not just the da Vinci plug-in that failed to implement ODF in Massachusetts!  Nine months later Sun delivered their ODF plug-in for MSOffice to Massachusetts.  The next day, Massachusetts threw in the towel, officially recognizing MS-OOXML (and the MS-OOXML Compatibility Pack plug-in) as a standard format for the future.

    Worse, the Massachusetts recognition of MS-OOXML came just weeks before the September 2nd ISO vote on MS-OOXML.  Why not wait a few more weeks?  After all, Massachusetts had conducted a year long pilot study to implement ODF using ODF desktop office sutie alternatives to MSOffice.  Not only did the rip out and replace approach fail, but they were also unable to integrate OpenOffice ODF desktops into existing MSOffice bound workgroups.

    The year long pilot study was followed by another year long effort trying to implement ODF using the plug-in approach.  That too failed with Sun's ODF plug-in the final candidate to prove the difficulty of implementing ODF in situations where MSOffice workgroups dominate.

    California and the EU-IDABC were closely watching the events in Massachusetts, as was most every CIO in government and private enterprise.  Reasoning that if Massachusetts was unable to implement ODF, California CIO's totally refused IBM and Sun's effort to get a pilot study underway.

    Across the pond, in the aftermath of Massachusetts CIO Louis Guiterrez resignation on October 4th, 2006, the EU-IDABC set about developing their own file format, ODEF.  The Open Document Exchange Format splashed into the public discussion on February 28th, 2007 at the "Open Document Exchange Workshop" held in Berlin, Germany.

    Meanwhile, the Sun ODF plug-in is fl
Gary Edwards

Can IBM save OpenOffice.org from itself? - 0 views

  • OpenOffice.org's biggest foe may be Microsoft Office, but critics say the open-source organization has, from its inception, also been one of its application suite's own worst enemies -- a victim of a development culture that differs radically from the open-source norm. Observers now wonder if IBM's entry into OpenOffice.org can make the necessary changes.
    • Gary Edwards
       
      good article from Eric Lai of ComputerWorld. Written on the eve of the infamous Barcelona OpenOffice.org developers conference, Eric argues that OOo isn't a real open source community. Instead, OOo is a owned and operated by Sun.
      One of the more important control points Sun insists on is that of commit rights and project managers. Only Sun employees have these rights and can hold these important positions..
      The more important point is made by Marbux (below: ODF is an application specific format designed exactly for OpenOffice. While other applications might partially implement ODF, interoperability and successful ODF document exchange require the OOo code base!
      From Marbux: This is the only article I've found to date where IBM (Heintzman) flat out says IBM wants changes in OOo licensing, more in line with the Eclipse and Apache licenses. See pg. 2. Significant because it feeds the meme that IBM's own ODF-based development goal is proprietary closed source built on the OOo code base, e.g., Symphony, et cet.
      And that has huge signficance once you realize that ODF is not the real standard; the OOo code base is the real ODF standard. Look around the world and you see that ODF adoption decisions by governments are all in reality decisions to go with StarOffice, OOo, or OOo clones. I haven't, for example, seen a single instance where a government decided to ride with KOffice. Why would they, with the interop issues between KOffice and OOo? The fact that OOo's code base is the real ODF standard will figure strongly in the comments. Couple it with Sun's iron-fisted control of the OOo code base, and you have vendor lock-in with a Microsoft partner.
      But with 70 developers committed in China, where developers salaries are inexpensive, IBM will soon be in a position to threaten to fork the OOo code base using proprietary extensions. Is that their real tactic to force changes in licensing and gov
Gary Edwards

Google: OOXML 'insufficient and unnecessary' - marbux - ge comments | ZDNet UK - 0 views

  • Google's technical analysis of the OOXML specification — which notoriously runs to 6,000 pages of code, compared with ODF's 860 pages — has led the company to believe that "OOXML would be an insufficient and unnecessary standard, designed purely around the needs of Microsoft Office", Bhorat claimed.
  • ODF and OOXML are standards in... Marbux
  • Interoperability and the binary ODF conversion di... garyedwards Sorry, the comment was cut short. Here'... garyedwards
Gary Edwards

PlexNex: Analyzing the Microsoft Office Open XML License - 0 views

  • There are many other warts in the Microsoft covenant not to sue. E.g., the covenant applies only to Ecma Office Open XML; it does not apply to any future version, including a version that might be approved by ISO or a variant that might be actually implemented by Microsoft in MS Office. So Microsoft makes no guarantee that it will not move the goal posts at any time.
  •  
    Whoa!  This has already happened.  In his blog titled, "The Formats of Excel 2007",  XML expert Rob Weir demonstrates for us that MSOffice 2007 Excel has a new file format.  Rob demonstrates that there are four file format choices in Excel; EOOXML, Legacy XLS binary, and two  new binary extensions of EOOXML: "Excel Macro-Enabled Workbook" - xlsxm, and "Excel Binary Workbook" - xlsb.

    The new binaries are proprietary extensions to EOOXML.  xlsb in particular looks to be something known as a XML Binary InfoSet..  XBiS is a compressed form of an XML file used in situations where bandwidth and device cpu constraints demand such an extreme.  We can't be sure about xlsb, but it looks like a duck, walks like a duck, quacks like a duck and thherefore....

    This must be some kind of record.  EOOXML isn't yet 30 days old and Micrsoft has eXtended it with a proprietary binary representation not available to the rest of the world.  And XBiS was designed so that implementations would be open and application and platform independent.  But that's not what we see with Microsoft's xlsb.

    What Marbux is pointing out here is that only Micrsoft has the legal rights to do this proprietary eXtension of EOOXML.  Beat the drums.  Sound the alarms.  Hide the women and children.  Nothing has changed.  The longboats are fancier, there are more of them. The swords of the pillagers remain just as sharp.  Their determination and drive just as strong.

    Some quick backgroud references:  Compression, XML</b
  • ...2 more comments...
  •  
    Whoa!  This has already happened.  In his blog titled, "The Formats of Excel 2007",  XML expert Rob Weir demonstrates for us that MSOffice 2007 Excel has a new file format.  Rob demonstrates that there are four file format choices in Excel; EOOXML, Legacy XLS binary, and two  new binary extensions of EOOXML: "Excel Macro-Enabled Workbook" - xlsxm, and "Excel Binary Workbook" - xlsb.

    The new binaries are proprietary extensions to EOOXML.  xlsb in particular looks to be something known as a XML Binary InfoSet..  XBiS is a compressed form of an XML file used in situations where bandwidth and device cpu constraints demand such an extreme.  We can't be sure about xlsb, but it looks like a duck, walks like a duck, quacks like a duck and thherefore....

    This must be some kind of record.  EOOXML isn't yet 30 days old and Micrsoft has eXtended it with a proprietary binary representation not available to the rest of the world.  And XBiS was designed so that implementations would be open and application and platform independent.  But that's not what we see with Microsoft's xlsb.

    What Marbux is pointing out here is that only Micrsoft has the legal rights to do this proprietary eXtension of EOOXML.  Beat the drums.  Sound the alarms.  Hide the women and children.  Nothing has changed.  The longboats are fancier, there are more of them. The swords of the pillagers remain just as sharp.  Their determination and drive just as strong.

    Some quick backgroud references:  Compression, XML</b
  •  
    There are many other warts in the Microsoft covenant not to sue. E.g., the covenant applies only toEcmaOffice Open XML; it does not apply to any future version, including a version that might be approved by ISO or a variant that might be actually imple
  •  
    There are many other warts in the Microsoft covenant not to sue. E.g., the covenant applies only toEcmaOffice Open XML; it does not apply to any future version, including a version that might be approved by ISO or a variant that might be actually imple
  •  
    There are many other warts in the Microsoft covenant not to sue. E.g., the covenant applies only toEcmaOffice Open XML; it does not apply to any future version, including a version that might be approved by ISO or a variant that might be actually imple
Gary Edwards

BetaNews | Microsoft Will Support ODF If It Doesn't 'Restrict Choice Among Formats' - 0 views

  • None of this is to say that OpenDocument is perfect. Far from it. OpenDocument at present is crippled from an interoperability standpoint. I'm a member of the OASIS OpenDocument Technical Committee and I think the resistance of the big vendors to fixing the interoperability warts is simply outrageous, particularly because they are fairly trivial changes. But the advancement of software users' interests are not advanced by painting OOXML as other than deeply flawed. It is vendor-specific and far from "open." The lesser of the two evils is clearly OpenDocument, which is at least open even if not yet interoperable. The sooner folks can start discussing practical methods of convergence, the better. See e.g., http://ec.europa.eu/idabc/servlets/Doc?id=27956 That set of slides summarizing a conference of some 20 European national governments' IT types says a lot more about the future of office document formats than Mr. Asellus has to offer.
    • Gary Edwards
       
      Marbux hits a homerun! Right ON!
Gary Edwards

Microsoft Support for ODF - the Q&A - 0 views

  • Hi Gary,I am a technology journalist with Asia's ONLY Linux-focused magazine, LINUX For You. I am working on a story revolving the recent development of Microsoft supporting ODF Format. I want to understand the equation of the whole development, would you please help me understand: Q1. What do you think drove Microsoft to support the ODF format?
  •  
    This is the full response to Swapnil's seven questions.  It's long.  But we hold back nothing!  Thanks again to Marbux.  He is a peach!
Gary Edwards

The Interoperability Wars :: ODF vs. OOXML - 0 views

  •  
    This discussion is part of a much larger response to a series of questions from a LiNUX magazine in India.  The mighty Marbux helped answer the questions, so please don't mistake his eloquence for my customary thundering.
Gary Edwards

The trap is set! Phony OASIS Messaging standard OK'd for Web services - 0 views

  • OASIS considers WS-ReliableMessaging critical for SOA, in that it can handle a variety of SOA requirements. WS-ReliableMessaging can be extended to enable integration with capabilities such as security. SOAP binding for interoperability is included. WS-ReliableMessaging is part of a series of Web services specifications dubbed WS-* that have been championed by companies such as Microsoft.
  •  
    Check out the comments below this article.  There are links to Ian Foster of Globus and to the Marbux license disseration posted with the OpenDocument Foundation's NO VOTE at OASIS.
Gary Edwards

An Antic Disposition: Cracks in the Foundation - IBM takes over ODF - 0 views

  • You must admire their tenacity. Gary Edwards and the pseudonymous "Marbux". The mythology of Silicon Valley is filled with stories of two guys and a garage founding great enterprises. And here we have two guys, and through blogs, interviews, and constant attendance at conferences, they have become some of the most-heard voices on ODF. Maybe it is partly due to the power of the name? The "OpenDocument Foundation" sounds so official. Although it has no official role in the ODF standard, this name opens doors. The ODF Alliance , the ODF Fellowship, the OASIS ODF TC, ODF Adoption TC (and many other groups without "ODF" in their name) have done far more to promote and improve ODF, yet the OpenDocument Foundation, Inc. seems to score the panel invites. Not bad for two guys without a garage.
  •  
    An eMail went out today, October 24th, 2007, nominating IBM's Rob "Show me your garage!" Weir to be the new Co Chairman of OASIS ODF TC.  So it's looks like it's true; IBM is moving to take over ODF and OpenOffice.

    Not that that's bad.  In the long run this is perhaps the best thing that ever happened to ODF and OpenOffice.  There is no way IBM's Lotus Notes business plan for ODF-OOo could be any worse than Sun's plan has turned out to be. 

    ~ge~

Gary Edwards

ODF Civil War: Bulll Run - Suggested Changes on the Metadata proposal - OASIS ODF - 0 views

  • From our perspective it would be better to aim for doing the job in ODF 1.2, even if that requires delay. We will oppose ODF 1.2 at ISO unless the interoperability warts are cleaned up. What the market requires is no longer in doubt. See the slides linked above and further presentations linked from this page, &lt; http://ec.europa.eu/idabc/en/document/6474/5935&gt;. Substantial progress toward those goals would seem to be mandatory to maintain Europe's preference for a harmonized set of file formats that uses ODF to provide the common functionality. Delaying commencement of such work enhances the likelihood that governments will tire of waiting for ODF to become interoperable with MS Office and simply go with MOOXML. We may not be able to force Microsoft to participate in the harmonization work, but we will be in a far better position if we have done everything we can in aid of that interoperability without Microsoft's assistance. As the situation stands, we have what is known in the U.S. as a "Mexican stand-off," where neither side has taken a solitary step toward what Europe has requested. We have decided to do that work via a fork of ODF; it is up to this TC whether it wishes to cooperate in that effort.
  •  
    This is the famous marbux response to Sun regarding Sun's attempt to partially implement ODF 1.2 XML-RDF metadata.  It's a treasure.

    There is one problem with marbux's statement though.  We had decided long ago not to fork ODF even if the five iX "interoperability enhancement" proposals were refused by the OASIS ODF TC.   This assurance was provided to Massachusetts CIO Louis Gutierrez witht he the first ODF iX proposal submitted on July 12th, 2006.  Louis ended up signing off on three iX proposals before his resignation October 4th, 2006.

    The ODF iX enhancements were essential to saving ODF in Massachusetts.  Without them, there was no way our da Vinci plug-in could convert existing MSOffice documents and processes to ODF with the needed round trip fidelity.

    For nearly a year we tried to push through some semblance of the needed iX enhancements.  We also tried to push through a much needed Interoperability Framework, which will be critical to any ISO approval of ODF 1.2.

    Our critics are correct in that every iX effort was defeated, with Sun providing the primary opposition. 

    Still rather than fork ODF, we are simply going to move on. 

    On October 4th, 2006, all work on ODF da Vinci ended - not to be resumed unless and until we had the ODF iX enhancements we needed to crack the MSOffice bound workgroup-workflow business process barrier.

    In April of 2007, with our OASIS membership officially shredded by OASIS management, bleeding from the List Enhancement Proposal doonybrook, and totally defeated with our hope - the metadata XML-RDF work, we threw in the towel.

    Since then we've moved on to CDF, the W3C Compound Document format.  Incredibly, CDF is able to do what ODF can not.  With CDF we can solve the three primary problems confronting governments and MSOffice bound workgroups everywhere. 

    The challenge for these g
Gary Edwards

ConsortiumInfo.org - Putting the OpenDocument Foundation to Bed (without its supper) - The OASIS Lawyer - 0 views

  • Here's what Chris Lilley had to say, reconstructed from my notes (in other words, this is not a direct quote): So we were in a meeting when these articles about the Foundation and CDF started to appear, and we were really puzzled.&nbsp; CDF isn't anything like ODF at all – it's an "interoperability agreement," mainly focused on two other specifications - XHTML and SVG.&nbsp; You'd need to use another W3C specification, called Web Interactive Compound Document (WICD, pronounced "wicked"), for exporting, and even then you could only view, and not edit the output.&nbsp; The one thing I'd really want your readers to know is that CDF (even together with WICD) was not created to be, and isn't suitable for use, as an office format.&nbsp; Here are some other takeaways from my conversation with Chris: Although they would be welcome to become members, Neither Gary, Sam nor Marbux are members of W3C or the CDF working group The W3C has never been contacted by anyone from the Foundation about CDF.&nbsp; After the articles began appearing, the W3C sent an inquiry to the Foundation, and received only a general reply in response The CDF working group was not chartered to achieve conversion between formats Although he hasn't spent a lot of time trying to unravel what Gary has written on the subject, he can't make any sense out of why the Foundation thinks that CDF makes sense as a substitute for ODF
1 - 20 of 30 Next ›
Showing 20 items per page