Challenges: Some Ugly Truths
The challenges of building—and living with—an XML workflow are clear enough. The return on investment is a long-term proposition.
Regardless of the benefits XML may provide, the starting reality is that it represents a very different way of doing things than the one we are familiar with. The Word Processing and Desktop Publishing paradigm, based on the promise of onscreen, WYSIWYG layout, is so dominant as to be practically inescapable. It has proven really hard to get from here to there, no matter how attractive XML might be on paper.
A considerable amount of organizational effort and labour must be expended up front in order to realize the benefits. This is why XML is often referred to as an “investment”: you sink a bunch of time and money up front, and realize the benefits—greater flexibility, multiple output options, searching and indexing, and general futureproofing—later, over the long haul. It is not a short-term return proposition.
And, of course, the returns you are able to realize from your XML investment are commensurate with what you put in up front: fine-grained, semantically rich tagging is going to give you more potential for searchability and recombination than a looser, more general-purpose approach, but it sure costs more. For instance, the Text Encoding Initiative (TEI) is the grand example of pouring enormous amounts of energy into the up-front tagging, with a very open-ended set of possibilities down the line. TEI helpfully defines a level to which most of us do not have to aspire.[5]
But understanding this on a theoretical level is only part of the challenge. There are many practical issues that must be addressed. Software and labour are two of the most critical. How do you get the content into XML in the first place? Unfortunately, despite two decades of people doing SGML and XML, this remains an ugly question.
Group items matching
in title, tags, annotations or urlXML Production Workflows? Start with the Web and XHTML - 0 views
-
-
Practical Challenges In 2009, there is still no truly likeable—let alone standard—editing and authoring software for XML. For many (myself included), the high-water mark here was Adobe’s FrameMaker, substantially developed by the late 1990s. With no substantial market for it, it is relegated today mostly to the tech writing industry, unavailable for the Mac, and just far enough afield from the kinds of tools we use today that its adoption represents a significant hurdle. And FrameMaker was the best of the breed; most of the other software in decent circulation are programmers’ tools—the sort of things that, as Michael Tamblyn pointed out, encourage editors to drink at their desks. The labour question represents a stumbling block as well. The skill-sets and mind-sets that effective XML editors need have limited overlap with those needed by literary and more traditional production editors. The need to think of documents as machine-readable databases is not something that comes naturally to folks steeped in literary culture. In combination with the sheer time and effort that rich tagging requires, many publishers simply outsource the tagging to India, drawing a division of labour that spans oceans, to put it mildly. Once you have XML content, then what do you do with it? How do you produce books from it? Presumably, you need to be able to produce print output as well as digital formats. But while the latter are new enough to be generally XML-friendly (e-book formats being largely XML based, for instance), there aren’t any straightforward, standard ways of moving XML content into the kind of print production environments we are used to seeing. This isn’t to say that there aren’t ways of getting print—even very high-quality print—output from XML, just that most of them involve replacing your prepress staff with Java programmers.
-
Why does this have to be so hard? It’s not that XML is new, or immature, or untested. Remember that the basics have been around, and in production, since the early 1980s at least. But we have to take account of a substantial and long-running cultural disconnect between traditional editorial and production processes (the ones most of us know intimately) and the ways computing people have approached things. Interestingly, this cultural divide looked rather different in the 1970s, when publishers were looking at how to move to digital typesetting. Back then, printers and software developers could speak the same language. But that was before the ascendancy of the Desktop Publishing paradigm, which computerized the publishing industry while at the same time isolating it culturally. Those of us who learned how to do things the Quark way or the Adobe way had little in common with people who programmed databases or document-management systems. Desktop publishing technology isolated us in a smooth, self-contained universe of toolbars, grid lines, and laser proofs. So, now that the reasons to get with this program, XML, loom large, how can we bridge this long-standing divide?
- ...44 more annotations...
-
I was looking for an answer to a problem Marbux had presented, and found this interesting article. The issue was that of the upcoming conversion of the Note Case Pro (NCP) layout engine to the WebKit layout engine, and what to do about the NCP document format. My initial reaction was to encode the legacy NCP document format in XML, and run an XSLT to a universal pivot format like TEI-XML. From there, the TEI-XML community would provide all the XSLT transformation routines for conversion to ODF, OOXML, XHTML, ePUB and HTML/CSS. Researching the problems one might encounter with this approach, I found this article. Fascinating stuff. My take away is that TEI-XML would not be as effective a "universal pivot point" as XHTML. Or perhaps, if NCP really wants to get aggressive; IDML - InDesign Markup Language. The important point though is that XHTML is a browser specific version of XML, and compatible with the Web Kit layout engine Miro wants to move NCP to. The concept of encoding an existing application-specific format in XML has been around since 1998, when XML was first introduced as a W3C standard, a "structured" subset of SGML. (HTML is also a subset of SGML). The multiplatform StarOffice productivity suite became "OpenOffice" when Sun purchased the company in 1998, and open sourced the code base. The OpenOffice developer team came out with a XML encoding of their existing document formats in 2000. The application specific encoding became an OASIS document format standard proposal in 2002 - also known as ODF. Microsoft followed OpenOffice with a XML encoding of their application-specific binary document formats, known as OOXML. Encoding the existing NCP format in XML, specifically targeting XHTML as a "universal pivot point", would put the NCP Outliner in the Web editor category, without breaking backwards compatibility. The trick is in the XSLT conversion process. But I think that is something much easier to handle then trying to
-
I was looking for an answer to a problem Marbux had presented, and found this interesting article. The issue was that of the upcoming conversion of the Note Case Pro (NCP) layout engine to the WebKit layout engine, and what to do about the NCP document format. My initial reaction was to encode the legacy NCP document format in XML, and run an XSLT to a universal pivot format like TEI-XML. From there, the TEI-XML community would provide all the XSLT transformation routines for conversion to ODF, OOXML, XHTML, ePUB and HTML/CSS. Researching the problems one might encounter with this approach, I found this article. Fascinating stuff. My take away is that TEI-XML would not be as effective a "universal pivot point" as XHTML. Or perhaps, if NCP really wants to get aggressive; IDML - InDesign Markup Language. The important point though is that XHTML is a browser specific version of XML, and compatible with the Web Kit layout engine Miro wants to move NCP to. The concept of encoding an existing application-specific format in XML has been around since 1998, when XML was first introduced as a W3C standard, a "structured" subset of SGML. (HTML is also a subset of SGML). The multiplatform StarOffice productivity suite became "OpenOffice" when Sun purchased the company in 1998, and open sourced the code base. The OpenOffice developer team came out with a XML encoding of their existing document formats in 2000. The application specific encoding became an OASIS document format standard proposal in 2002 - also known as ODF. Microsoft followed OpenOffice with a XML encoding of their application-specific binary document formats, known as OOXML. Encoding the existing NCP format in XML, specifically targeting XHTML as a "universal pivot point", would put the NCP Outliner in the Web editor category, without breaking backwards compatibility. The trick is in the XSLT conversion process. But I think that is something much easier to handle then trying to
The new UI wars: Why there's no Flash on iPhone 2.0 « counternotions - 0 views
-
- publishers of Flash apps have to port their apps to native Web apps if they want to run inside a Web browser going forward because the Web has moved off the PC, you can’t accessorize it with PC software anymore, WebKit is so small and light and cross-platform that it is the plug-in now, inside iPhone, iPod, Nokia, Android, iTunes and other Mac and Windows apps - publishers of Flash video have to deploy MPEG-4 H.264/AAC if they want to run inside an audio-video player (on any device) going forward, the decoder chips for this are already in EVERYTHING, from iPod to Blu-Ray to NVIDIA GPU’s Most of the world has already done both of the above, including Google and Apple. This is not the beginning of the end for Flash, it is the end of the end.
-
Notice he doesn’t say at all that Flash is running natively on the ARM CPU inside the iPhone. And once again, as I point out in the article above, technical problems may be solved by Adobe, but cross-platform runtime compatibility and multi-touch UI frameworks remain as serious impediments.
-
the direction Apple is taking in WebKit with canvas, downloadable fonts, SVG, CSS animation, CSS transformations, faster JavaScript, HTML5 audio/video embedding, exposure of multi-touch to JS and so on is precisely to create an open source alternative to the Flash runtime engine, without having to download a proprietary plugin:
-
this article takes the RiA discussion to an entirely new level - the battle between Apple, Adobe and Microsoft to control the future user interface (UI). Adobe Flash extends the aging WiMP model, trying to create a "UI Convergence" across many platforms through the Flash RiA. With iPhone, Apple introduces the patented "gestures UI", running off the WebKit RiA. Microsoft presumably is copying the Flash RiA with the XAML rich WPF Silverlight RiA. Unfortunately, counternotions doe snot cover Silverlight. This incredible discussion is limited to Adobe and Apple.
-
Most quality online stores. Know whether you are a trusted online retailer in the world. Whatever we can buy very good quality. and do not hesitate. Everything is very high quality. Including clothes, accessories, bags, cups. Highly recommended. This is one of the trusted online store in the world. View now www.retrostyler.com
Is the W3C to Blame for the Breaking of the Web? | Continuing Intermittent Incoherency » Power vs. Authority - 0 views
-
Consider the recent CSS features added by WebKit: transformations, animations, gradients, masks, et cetera. They’ve very nearly _run out_ of standards to implement, so they’re starting to implement the wouldn’t-it-be-cool-if stuff. If I’m not mistaken, this is the exact sort of thing you’re wishing for.
-
Changing the renderer (which is what we’re taking about when we talk about upgrading “the web”) goes hand-in-hand today with upgrading the *rest* of the browser as well, which requires the user to care…and users (to a one) don’t give a flying leap about CSS 2.1 support.
-
Another article taking up the issue of "Blame the W3C" for what increasingly looks like a proprietary Web future. The author is an Ajax-DOJO supporter, and he tries to defend the W3C by saying it's not their job, they don't have the "power" or the "authority" to push the Web forward. About the best they can do is, at the end of the day, try to corral big vendors into agreement. Meanwhile, the Web has become the wild wild west with browser vendors innovating into their corporate web stacks where vast profits and future monopolies rest. For me, WebKit represents the best effort insisting that the Web remain Open. It's OSS with excellent big vendor support. And they are pushing the envelope. Finally!
The cross-platform option: Web apps for smartphones | InfoWorld | Analysis | 2009-01-20 | By Peter Wayner - 0 views
-
Tailoring a Web application for iPhone, BlackBerry, and other top devices has many advantages over costly native development, and new techniques and approaches are closing the functionality gap. This article walks through the many different issues confronting Web developers and, the many innovations developers are using to close the gap between the Web and increasingly visual devices. Excellent article! Covers the iPhone WebKit - CSS transformations, iUI and PhoneGAP
1 - 5 of 5
Showing 20▼ items per page
".....Yes, that's right a bulging docked menu, with no javascript. Just so you remember, there no javascript in the demo. Check out the Javascript free OSX Dock Menu Demo.
This demo actually proves an important point Tom Yager made earlier about Ajax; Will JavaScript inconsistencies break the Web?
Taking AJAX literally makes lousy Web apps: "As little as possible should be the rule for JavaScript, which must play a supporting role to CSS and HTML". Tom concludes that it's best to follow the WebKit model, putting everything possible into first CSS4, then HTML5, and then JavaScript. I would argue that the proliferation of JavaScript libraries is a good hedge against the non interoperable future Yager warns of. But hey, why stop the guy when he's on a roll. CSS4! I guess the webkit-transforms have been officially christened. Thanks Tom.
~ge~