Skip to main content

Home/ Open Web/ Group items tagged Dynamic

Rss Feed Group items tagged

Gary Edwards

Roger Black : "We save trees" - The Story of TreeSaver - 0 views

  •  
    Roger Black is a publication designer, and Filipe Fortes was the project manager for Microsoft's Windows Presentation Foundation - Silverlight project.  Today he specializes in dynamic layout algorithms written in Open Web JavaScript.  Roger is a reknown publication designer, and describes here the genesis behind TreeSaver.  Fascinating story certain to become a key explanation of how digital media ran away with the print publishing industry.   I'm wondering what kind of authoring tools will evolve that can publish directly into the TS JavaScript templates?   My first inclination would be to adapt OOo Impress.  It has an outline view, a notes editing capability, and provides a decent visual canvas.  The problem is that it's locked into "slides".  Can Impress be unlocked and flowing?  That might work. excerpt:  when Microsoft put together a dynamic page layout for The New York Times Reader, did they know that it was the future? It certainly wasn't the immediate present, since they couldn't pry the WPF visual layer off of Windows, leaving it a single-OS solution. (The Times' Reader later was taken up by Adobe, which at least got it to work on both Mac and PC.) Filipe Fortes, PM on the MS news client project, knew. I'd met him when the group invited me out to Redmond to help design the first templates for the Times. Later I saw him at the 2007 Mix conference in Vegas, and I asked him how to make the dynamic page size idea work multi-platform. He said, "We could do it in HTML."
Gary Edwards

Microsoft Reinvents its Cloud Strategy - 0 views

  •  
    Microsoft announces a new plan, "Microsoft Dynamics", to accelerate the transition to replace their traditional / conventional software systems with cloud-ready infrastructure. Dynamics also serves as a migration guide, providing "seamless integration" of services from old IT to new Cloud. Competitors mentioned include SalesForce.com, SAP and Oracle. Lots of focus on integrating CRM into the full line of business applications. This is somewhat similar to the Visual Productivity challenge of integrating gMail into a working productivity system based on the Google Apps platform. Embedding the full range of Visual Web Communications into productivity apps and services is the key at all levels. One thing to consider in this article is that the only Cloud - Productivity - Business systems contenders are SalesForce, Oracle, SAP and Microsoft. Google Apps isn't mentioned. Nor is Amazon, RackSpace or VMware. Apple, Cisco, HP and Facebook are also left out.
Paul Merrell

Feds Claim They Can Enter a House and Demand Fingerprints to Unlock Everyone's Phones - 0 views

  • Under the Fourth Amendment, Americans are protected from unreasonable searches and seizures, but according to one group of federal prosecutors, just being in the wrong house at the wrong time is cause enough to make every single person inside provide their fingerprints and unlock their phones.Back in 2014, a Virginia Circuit Court ruled that while suspects cannot be forced to provide phone passcodes, biometric data like fingerprints doesn’t have the same constitutional protection. Since then, multiple law enforcement agencies have tried to force individual suspects to unlock their phones with their fingers, but none have claimed the sweeping authority found in a Justice Department memorandum recently uncovered by Forbes.
  • In the court document filed earlier this year, federal prosecutors in California argued that a warrant for a mass finger-unlocking was constitutionally sound even though “the government does not know ahead of time the identity of every digital device or every fingerprint (or indeed, every other piece of evidence) that it will find in the search” because “it has demonstrated probable cause that evidence may exist at the search location.” Criminal defense lawyer Marina Medvin, however, disagreed. Advertisement Advertisement “They want the ability to get a warrant on the assumption that they will learn more after they have a warrant,” Medvin told Forbes. “This would be an unbelievably audacious abuse of power if it were permitted.”Unfortunately, other documents related to the case were not publicly available, so its unclear if the search was actually executed. Even so, Medvin believes the memorandum sets a deeply troubling precedent, using older case law regarding the collection of fingerprint evidence to request complete access to the “amazing amount of information” found on a cellphone.
Gary Edwards

Government Market Drags Microsoft Deeper into the Cloud - 0 views

  •  
    Nice article from Scott M. Fulton describing Microsoft's iron fisted lock on government desktop productivity systems and the great transition to a Cloud Productivity Platform.  Keep in mind that in 2005, Massachusetts tried to do the same thing with their SOA effort.  Then Governor Romney put over $1 M into a beta test that produced the now infamous 300 page report written by Sam Hiser.  The details of this test resulted in the even more infamous da Vinci ODF plug-in for Microsoft Office desktops.   The lessons of Massachusetts are simple enough; it's not the formats or office suite applications.  It's the business process!  Conversion of documents not only breaks the document.  It also breaks the embedded "business process". The mystery here is that Microsoft owns the client side of client/server computing.  Compound documents, loaded with intertwined OLE, ODBC, ActiveX, and other embedded protocols and interface dependencies connecting data sources with work flow, are the fuel of these client/server business productivity systems.  Break a compound document and you break the business process.   Even though Massachusetts workers were wonderfully enthusiastic and supportive of an SOA based infrastructure that would include Linux servers and desktops as well as OSS productivity applications, at the end of the day it's all about getting the work done.  Breaking the business process turned out to be a show stopper. Cloud Computing changes all that.  The reason is that the Cloud is rapidly replacing client/server as the target architecture for new productivity developments; including data centers and transaction processing systems.  There are many reasons for the great transition, but IMHO the most important is that the Web combines communications with content, data, and collaborative computing.   Anyone who ever worked with the Microsoft desktop productivity environment knows that the desktop sucks as a communication device.  There was
Gary Edwards

That's All Folks: Why the Writing Is on the Wall at Microsoft - Forbes - 1 views

  •  
    Control vs. Creativity.  As the ulitmate control freak, Ballmer was guanteed to crush the life out of Microsoft's most creative individuals and teams.  Gates was focused on Windows and MSOffice, and Ballmer on control.  That one-two punch made certain that Microsoft would not be a player in the next great wave of computing; The Cloud.   excerpt: This is yet another example of what I like to call the Wile E. Coyote syndrome. Like the unfortunate character in the old Warner Bros. cartoons, Microsoft now seems to be a company that has long since run off the cliff but, with legs spinning for all they are worth, doesn't know yet that it is ready to drop. Yet drop it most certainly will. Microsoft Win8 Tablet Is NOT a Game Changer Adam Hartung Contributor Snapshot: Steve Ballmer Follow (93) #44 Billionaires IDC Analyst: Microsoft's Surface Sizzle Needs Win8 Steak Daniel Nye Griffiths Contributor To understand how this happens, take a look at the work of Arnold J. Toynbee, a historian who studied the rise and fall of civilizations. He argued that a civilization flourishes when it motivates insiders and attracts outsiders with its creative dynamism and culture. The civilization breaks down when its leadership loses this creative capacity and gives way to, or transforms itself into, a dominant minority. When this happens, the driver of the civilization becomes control, not attraction. And it's precisely this switch from attraction to control that is the source of the breakdown. Interestingly, Toynbee says that the consequences may not be immediately apparent. A civilization can keep up momentum because the controls it puts in place generate some short-term efficiency. But eventually it will run its course and collapse, because no amount of control can replace the loss of collective creativity. Observe this in the corporate world by looking at the example of General Motors. G.M.'s 2009 bankruptcy came at the end of a long decline dating back to the early 197
Gary Edwards

Google Wave expands, in search of a clear use-case scenario | Web Apps News - Betanews - 0 views

  •  
    Excellent Wave coverage from Scott M. Fulton III.  The Hollywood Movie industry use case scenario is very interesting.  But Scott is one of the few people to draw an analogy between Google Wave convergence of concurrent communications and collaborative content and the early days of the Microsoft Office Productivity Platform where we saw DDE, OLE and MAPI rise and rule. Excerpt: Today, Google is expected to invite as many as 100,000 more participants into the private beta of its concurrent communications system, called Wave. As that happens, many more participants will be able to not only communicate with one another in a more granular form of real-time, but potentially collaborate on work and projects. It's that latter part of the program that's supposed to congeal at some point into a collective sense of purpose. But this time, unlike Microsoft's first experiments with Dynamic Data Exchange between applications on the same computer three decades ago, there isn't yet a clear, single purpose for the system. No question it could bring individuals as close together as people separated by indefinite distance could become; but as to the question of what they do with one another once they do get together, Google is hoping this question -- like so many others it puts out there in the open -- resolves itself. Yesterday, Google offered links to a number of different independent assessments of the possible, eventual purpose of Google Wave, though it offered them as use-case scenarios rather than projections of possible goals for the product...which is what many actually were.
Gary Edwards

Official Google Docs Blog: What's different about the new Google Docs? - 0 views

  •  
    The new Google documents To get around these problems, the new Google document editor doesn't use the browser to handle editable text. We wrote a brand new editing surface and layout engine, entirely in JavaScript. A new editing surface Let's start by talking about the editing surface, which processes all user input and makes the application feel like a regular editor. To you, the new editor looks like a fairly normal text box. But from the browser's perspective, it's a webpage with JavaScript that responds to any user action by dynamically changing what to display on each line. For example, the cursor you see is actually a thin, 2 pixel-wide div element that we manually place on the screen. When you click somewhere, we find the x and y coordinates of your click and draw the cursor at that position. This lets us do basic things like slanting the cursor for italicized text, and it also allows more powerful capabilities like showing multiple collaborators' cursors simultaneously, in the same document.
Gary Edwards

Adeptol Viewing Technology Features - 0 views

  •  
    Quick LinksGet a TrialEnterprise On DemandEnterprise On PremiseFAQHelpContact UsWhy Adeptol?Document SupportSupport for more than 300 document types out of boxNot a Virtual PrinterMultitenant platform for high end document viewingNo SoftwaresNo need to install any additional softwares on serverNo ActiveX/PluginsNo plugins or active x or applets need to be downloaded on client side.Fully customizableAdvanced API offers full customization and UI changes.Any OS/Any Prog LanguageInstall Adeptol Server on any OS and integrate with any programming language.AwardsAdeptol products receive industry awards and accolades year after year  View a DemoAttend a WebcastContact AdeptolView a Success StoryNo ActiveX, No Plug-in, No Software's to download. Any OS, Any Browser, Any Programming Language. That is the Power of Adeptol. Adeptol can help you retain your customers and streamline your content integration efforts. Leverage Web 2.0 technologies to get a completely scalable content viewer that easily handles any type of content in virtually unlimited volume, with additional capabilities to support high-volume transaction and archive environments. Our enterprise-class infrastructure was built to meet the needs of the world's most demanding global enterprises. Based on AJAX technology you can easily integrate the viewer into your application with complete ease. Support for all Server PlatformsCan be installed on Windows   (32bit/64bit) Server and Linux   (32bit/64bit) Server. Click here to see technical specifications.Integrate with any programming languageWhether you work in .net, c#, php, cold fusion or JSP. Adeptol Viewer can be integrated easily in any programming language using the easy API set. It also comes with sample code for all languages to get you started.Compatibility with more than 99% of the browsersTested & verified for compatibility with 99% of the various browsers on different platforms. Click here to see browser compatibility report.More than 300 Document T
Gary Edwards

The Future of Collaborative Networks : Aaron Fulkerson of MindTouch - 0 views

  •  
    MindTouch was by far and away the hottest property at the 2009 Web 2.0 Conference. And for good reason. They have figured out how to tap into the productivity value of enterprise collaborative networks. Most their underlying stuff is based on REST based data objects and services, but they also allow for proprietary data bindings. The key to MindTouch seemd to be the easy to fall into and use collaborative interface: imagine a workgroup project centered around a Web page filled with data objects, graphics and content, with each object also having a collabortaive conversation attached to it. Sounds complicated, but that's where the magic of MindTouch kicks in. It's simple. One the things that most impressed me was an interactive graph placed on one of the wiki project pages. The graph was being fed data from a local excel spreadsheet, and could be interacted with in real time. It was simple to change from a pie chart to a bar graph and so on. It was also possible to interact with the data itself and create what-if scenario's. Great stuff. With considerable persistence though, i was able to discover from Aaron that this interactivity and graphical richness was due to a Silverlight plug-in! From the article: "..... Rather than focusing on socialization, one to one interactions and individual enrichment, businesses must be concerned with creating an information fabric within their organizations. This information fabric is a federation of content from the multiplicity of data and application silos utilized on a daily basis; such as, ERP, CRM, file servers, email, databases, web-services infrastructures, etc. When you make this information fabric easy to edit between groups of individuals in a dynamic, secure, governed and real-time manner, it creates a Collaborative Network." "This is very different from social networks or social software, which is focused entirely on enabling conversations. Collaborative Networks are focused on groups accessing and organiz
Maluvia Haseltine

OpenNebula :: The Open Source Toolkit for Building Clouds - 0 views

  •  
    OpenNebula is a Virtual Infrastructure Manager that orchestrates storage, network and virtualization technologies to enable the dynamic placement of multi-tier services (groups of interconnected virtual machines) on distributed infrastructures, combining both data center resources and remote cloud resources, according to allocation policies. OpenNebula provides internal and Cloud administration and user interfaces for the full management of the Cloud platform.
Gary Edwards

Aspose.Word for NET 5.3.0 (Windows), from Aspose Pty Ltd - Software Downloads - TechRep... - 0 views

  •  
    Overview: Aspose.Word is a powerful .NET component that allows any .NET application to work with Microsoft Word documents without utilizing Microsoft Word and Office Automation. Aspose.Word supports a wide array of features including new document creation, document manipulation, powerful mail merge abilities, exporting to multiple formats (DOC, PDF & HTML) and much, much more. Aspose.Word is truly the most affordable, fastest, feature rich Word component on the market. Aspose.Word application programming interface (API) is powerful and easy to use at the same time. Classes, properties and method names are easy to remember and understand: for example, Document represents a Word document, DocumentBuilder is responsible for building a document dynamically, and so on. Although this guilde provides many code samples, intuitively understandable interface allows you create working document producing applications having minimal knowledge of C# or VB.NET and Microsoft Word features. $899
Gary Edwards

Death of The Document - CIO Central - CIO Network - Forbes - 0 views

  •  
    Well, not quite.  More IBM happy talk about interoperability and easy document interchange.  While i agree with the static versus interactive - collaborative document perspective, it's far more complicated. Today we have a world of "native"  docs and "visual" docs.   Native docs are bound to their authoring productivity environment, and are stubbornly NOT interchangeable.  Even for ODF and OOXML formats. Visual documents are spun from natives, and they are highly interchangeable, but interactively limited.  They lack the direct interaction of native authoring environments.  The Visual document phenomenon starts with PDF and the virtual print driver.  Any authoring application(s) in a productivity environment can print a PDF using the magic of the virtual print driver.   In 2008, when ISO stamped PDF with "accessibility tags", a new, highly interactive version of PDF was offically recognized.  We know this as "Tagged PDF".  And it has led the sweeping revolution of wide implementation of the paperless transaction process. The Visual Document phenomenon doesn't stop there.  The highly mobile WebKit revolution ushered in by the 2008 iPhone phenomenon led to wide acceptance of highly interactive and collaborative, but richly visual versions of SVG and HTML5-CSS3-JSON-JavaScript documents. Today we have SVG-HTML+ type visually immersive documents spun out of Server side publication presses such as FlipBoard, Cognito cComics, QWiki, Needle, Sports Illustrated, Push Pop Press, and TreeSaver to name but a few.   Clearly the visually immersive category of documents is exploding, but not for business - productivity documents.  Adobe has proposed a "CSS Regions" standard for richly immersive layout that might change that.  But mostly i think the problem for business documents, reports and forms is that they are "compound documents" bound to desktop productivity environments and workgroups. The great transition from desktop/workgroup productivity environme
Paul Merrell

M of A - Assad Says The "Boy In The Ambulance" Is Fake - This Proves It - 0 views

  • Re: Major net hack - its not necessarily off topic. .gov is herding web sites into it's own little DNS animal farms so it can properly protect the public from that dangerous 'information' stuff in time of emergency. CloudFlare is the biggest abattoir... er, animal farm. CloudFlare is kind of like a protection racket. If you pay their outrageous fees, you will be 'protected' from DDoS attacks. Since CloudFlare is the preferred covert .gov tool of censorship and content control (when things go south), they are trying to drive as many sites as possible into their digital panopticons. Who the hell is Cloudflare? ISUCKER: BIG BROTHER INTERNET CULTURE On top of that, CloudFlare’s CEO Matthew Prince made a weird, glib admission that he decided to start the company only after the Department of Homeland Security gave him a call in 2007 and suggested he take the technology behind Project Honey Pot one step further… And that makes CloudFlare a whole different story: People who sign up for the service are allowing CloudFlare to monitor, observe and scrutinize all of their site’s traffic, which makes it much easier for intel or law enforcement agencies to collect info on websites and without having to hack or request the logs from each hosting company separately. But there’s more. Because CloudFlare doesn’t just passively monitor internet traffic but works like a dynamic firewall to selectively block traffic from sources it deems to be “hostile,” website operators are giving it a whole lotta power over who gets to see their content. The whole point of CloudFlare is to restrict access to websites from specific locations/IP addresses on the fly, without notifying or bothering the website owner with the details. It’s all boils down to a question of trust, as in: do you trust a shady company with known intel/law enforcement connections to make that decision?
  • And here is an added bonus for the paranoid: Because CloudFlare partially caches websites and delivers them to web surfers via its own servers, the company also has the power to serve up redacted versions of the content to specific users. CloudFlare is perfect: it can implement censorship on the fly, without anyone getting wise to it! Right now CloudFlare says it monitors nearly 1/5 of all Internet visits. [<-- this] An astounding claim for a company most people haven’t even heard of. And techie bloggers seem very excited about getting as much Internet traffic routed through them as possible! See? Plausable deniability. A couple of degrees of separation. Yet when the Borg Queen wants to start WWIII next year, she can order the DHS Stazi to order outfits like CloudFlare to do the proper 'shaping' of internet traffic to filter out unwanted information. How far is any expose of propaganda like Dusty Boy going to happen if nobody can get to sites like MoA? You'll be able to get to all kinds of tweets and NGO sites crying about Dusty Boy 2.0, but you won't see a tweet or a web site calling them out on their lies. Will you even know they interviewed Assad? Will you know the activist 'photographer' is a paid NGO shill or that he's pals with al Zenki? Nope, not if .gov can help it.
Paul Merrell

Dropbox: Condoleeza Rice appointment won't alter privacy pledge - CNET - 0 views

  • Dropbox CEO Drew Houston sought to quell the uproar over the appointment of former US Secretary of State Condoleezza Rice to the company's board of directors, saying in a blog post Friday that Rice's appointment won't change its stance on privacy. "There's nothing more important to us than keeping your stuff safe and secure. It's why we've been fighting for transparency and government surveillance reform, and why we've been vocal and public with our principles and values," Houston wrote. "We should have been clearer that none of this is going to change with Dr. Rice's appointment to our Board." The cloud storage service is trying to grow its international presence -- something Rice should be able to help with. However, after Dropbox announced her appointment earlier this week, a Web site dedicated to the "Drop Dropbox" movement called her selection "deeply disturbing" and said her board role was "problematic on a number of deeper levels, and invites serious concerns" about management's commitment "to freedom, openness, and ethics."
  • The movement said it objected to her role in the US decision to go to war in Iraq, as well as her position on the use of torture against prisoners. What's more, they said Rice supported the George W. Bush administration's "warrantless wiretap program and expansive domestic surveillance program."Houston responded in his brief note today, saying that Dropbox "should have been clearer that none of this is going to change" in the aftermath of Rice's appointment. "Our commitment to your rights and your privacy is at the heart of every decision we make, and this will continue," he wrote.
  • "We're honored to have Dr. Rice join our board -- she brings an incredible amount of experience and insight into international markets and the dynamics that define them," Houston wrote. "As we continue to expand into new countries, we need that type of insight to help us reach new users and defend their rights. Dr. Rice understands our stance on these issues and fully supports our commitments to our users."In her only public comments about Dropbox since being named to the board, Rice didn't get very detailed in speaking with Bloomberg on Wednesday. "As a country, we are having a great national conversation and debate about exactly how to manage privacy concerns," Rice said in the interview. "I look forward to helping Dropbox navigate it."
  •  
    The straw that broke this camel's back. On top of having an absolutely horrible security model, Dropbox elects Condi Rice to its board of directors. I just completed transfer of my files to another service (in the E.U. where U.S. court orders don't reach) and deleted my Dropbox account.  
Paul Merrell

This project aims to make '404 not found' pages a thing of the past - 0 views

  • The Internet is always changing. Sites are rising and falling, content is deleted, and bad URLs can lead to '404 Not Found' errors that are as helpful as a brick wall. A new project proposes an do away with dead 404 errors by implementing new HTML code that will help access prior versions of hyperlinked content. With any luck, that means that you’ll never have to run into a dead link again. The “404-No-More” project is backed by a formidable coalition including members from organizations like the Harvard Library Innovation Lab, Los Alamos National Laboratory, Old Dominion University, and the Berkman Center for Internet & Society. Part of the Knight News Challenge, which seeks to strengthen the Internet for free expression and innovation through a variety of initiatives, 404-No-More recently reached the semifinal stage. The project aims to cure so-called link rot, the process by which hyperlinks become useless overtime because they point to addresses that are no longer available. If implemented, websites such as Wikipedia and other reference documents would be vastly improved. The new feature would also give Web authors a way provide links that contain both archived copies of content and specific dates of reference, the sort of information that diligent readers have to hunt down on a website like Archive.org.
  • While it may sound trivial, link rot can actually have real ramifications. Nearly 50 percent of the hyperlinks in Supreme Court decisions no longer work, a 2013 study revealed. Losing footnotes and citations in landmark legal decisions can mean losing crucial information and context about the laws that govern us. The same study found that 70 percent of URLs within the Harvard Law Review and similar journals didn’t link to the originally cited information, considered a serious loss surrounding the discussion of our laws. The project’s proponents have come up with more potential uses as well. Activists fighting censorship will have an easier time combatting government takedowns, for instance. Journalists will be much more capable of researching dynamic Web pages. “If every hyperlink was annotated with a publication date, you could automatically view an archived version of the content as the author intended for you to see it,” the project’s authors explain. The ephemeral nature of the Web could no longer be used as a weapon. Roger Macdonald, a director at the Internet Archive, called the 404-No-More project “an important contribution to preservation of knowledge.”
  • The new feature would come in the form of introducing the mset attribute to the <a> element in HTML, which would allow users of the code to specify multiple dates and copies of content as an external resource. For instance, if both the date of reference and the location of a copy of targeted content is known by an author, the new code would like like this: The 404-No-More project’s goals are numerous, but the ultimate goal is to have mset become a new HTML standard for hyperlinks. “An HTML standard that incorporates archives for hyperlinks will loop in these efforts and make the Web better for everyone,” project leaders wrote, “activists, journalists, and regular ol’ everyday web users.”
Gary Edwards

XML Production Workflows? Start with the Web and XHTML - 1 views

  • Challenges: Some Ugly Truths The challenges of building—and living with—an XML workflow are clear enough. The return on investment is a long-term proposition. Regardless of the benefits XML may provide, the starting reality is that it represents a very different way of doing things than the one we are familiar with. The Word Processing and Desktop Publishing paradigm, based on the promise of onscreen, WYSIWYG layout, is so dominant as to be practically inescapable. It has proven really hard to get from here to there, no matter how attractive XML might be on paper. A considerable amount of organizational effort and labour must be expended up front in order to realize the benefits. This is why XML is often referred to as an “investment”: you sink a bunch of time and money up front, and realize the benefits—greater flexibility, multiple output options, searching and indexing, and general futureproofing—later, over the long haul. It is not a short-term return proposition. And, of course, the returns you are able to realize from your XML investment are commensurate with what you put in up front: fine-grained, semantically rich tagging is going to give you more potential for searchability and recombination than a looser, more general-purpose approach, but it sure costs more. For instance, the Text Encoding Initiative (TEI) is the grand example of pouring enormous amounts of energy into the up-front tagging, with a very open-ended set of possibilities down the line. TEI helpfully defines a level to which most of us do not have to aspire.[5] But understanding this on a theoretical level is only part of the challenge. There are many practical issues that must be addressed. Software and labour are two of the most critical. How do you get the content into XML in the first place? Unfortunately, despite two decades of people doing SGML and XML, this remains an ugly question.
  • Practical Challenges In 2009, there is still no truly likeable—let alone standard—editing and authoring software for XML. For many (myself included), the high-water mark here was Adobe’s FrameMaker, substantially developed by the late 1990s. With no substantial market for it, it is relegated today mostly to the tech writing industry, unavailable for the Mac, and just far enough afield from the kinds of tools we use today that its adoption represents a significant hurdle. And FrameMaker was the best of the breed; most of the other software in decent circulation are programmers’ tools—the sort of things that, as Michael Tamblyn pointed out, encourage editors to drink at their desks. The labour question represents a stumbling block as well. The skill-sets and mind-sets that effective XML editors need have limited overlap with those needed by literary and more traditional production editors. The need to think of documents as machine-readable databases is not something that comes naturally to folks steeped in literary culture. In combination with the sheer time and effort that rich tagging requires, many publishers simply outsource the tagging to India, drawing a division of labour that spans oceans, to put it mildly. Once you have XML content, then what do you do with it? How do you produce books from it? Presumably, you need to be able to produce print output as well as digital formats. But while the latter are new enough to be generally XML-friendly (e-book formats being largely XML based, for instance), there aren’t any straightforward, standard ways of moving XML content into the kind of print production environments we are used to seeing. This isn’t to say that there aren’t ways of getting print—even very high-quality print—output from XML, just that most of them involve replacing your prepress staff with Java programmers.
  • Why does this have to be so hard? It’s not that XML is new, or immature, or untested. Remember that the basics have been around, and in production, since the early 1980s at least. But we have to take account of a substantial and long-running cultural disconnect between traditional editorial and production processes (the ones most of us know intimately) and the ways computing people have approached things. Interestingly, this cultural divide looked rather different in the 1970s, when publishers were looking at how to move to digital typesetting. Back then, printers and software developers could speak the same language. But that was before the ascendancy of the Desktop Publishing paradigm, which computerized the publishing industry while at the same time isolating it culturally. Those of us who learned how to do things the Quark way or the Adobe way had little in common with people who programmed databases or document-management systems. Desktop publishing technology isolated us in a smooth, self-contained universe of toolbars, grid lines, and laser proofs. So, now that the reasons to get with this program, XML, loom large, how can we bridge this long-standing divide?
  • ...44 more annotations...
  • Using the Web as a Production Platform The answer, I think, is right in front of you. The bridge is the Web, a technology and platform that is fundamentally based on XML, and which many publishers are by now comfortably familiar with. Perhaps not entirely comfortably, but at least most publishers are already working with the Web; they already either know or have on staff people who understand it and can work with it. The foundation of our argument is this: rather than looking at jumping to XML in its full, industrial complexity, which seems to be what the O'Reilly-backed StartWithXML initiative[6] is suggesting, publishers instead leverage existing tools and technologies—starting with the Web—as a means of getting XML workflows in place. This means making small investments and working with known tools rather than spending tens of thousands of dollars on XML software and rarefied consultants. It means re-thinking how the existing pieces of the production toolchain fit together; re-thinking the existing roles of software components already in use. It means, fundamentally, taking the Web seriously as a content platform, rather than thinking of it as something you need to get content out to, somehow. If nothing else, the Web represents an opportunity to think about editorial and production from outside the shrink-wrapped Desktop Publishing paradigm.
  • Is the Web made of Real XML? At this point some predictable objections can be heard: wait a moment, the Web isn’t really made out of XML; the HTML that makes up most of the Web is at best the bastard child of SGML, and it is far too flaky/unstructured/underpowered to be taken seriously. We counter by arguing that although HTML on the Web exists in a staggering array of different incarnations, and that the majority of it is indeed an unstructured mess, this does not undermine the general principle that basic, ubiquitous Web technologies can make a solid platform for content management, editorial process, and production workflow.
  • With the advent of a published XML standard in the late 1990s came the W3C’s adoption of XHTML: the realization of the Web’s native content markup as a proper XML document type. Today, its acceptance is almost ubiquitous, even while the majority of actual content out there may not be strictly conforming. The more important point is that most contemporary Web software, from browsers to authoring tools to content management systems (from blogs to enterprise systems), are capable of working with clean, valid XHTML. Or, to put the argument the other way around, clean, valid XHTML content plays absolutely seamlessly with everything else on the Web.[7]
  • The objection which follows, then, will be that even if we grant that XHTML is a real XML document type, that it is underpowered for “serious” content because it is almost entirely presentation (formatting) oriented; it lacks any semantic depth. In XHTML, a paragraph is a paragraph is a paragraph, as opposed to a section or an epigraph or a summary.
  • n contrast, more “serious” XML document types like DocBook[8] or DITA-derived schemas[9] are capable of making semantic distinctions about content chunks at a fine level of granularity and with a high degree of specificity.
  • So there is an argument for recalling the 80:20 rule here. If XHTML can provide 80% of the value with just 20% of the investment, then what exactly is the business case for spending the other 80% to achieve that last 20% of value? We suspect the ratio is actually quite a bit steeper than 80:20 for most publishers.
  • Furthermore, just to get technical for a moment, XHTML is extensible in a fairly straightforward way, through the common “class” attribute on each element. Web developers have long leveraged this kind of extensibility in the elaboration of “microformats” for semantic-web applications.[10] There is no reason why publishers shouldn’t think to use XHTML’s simple extensibility in a similar way for their own ends.
  • XHTML, on the other hand, is supported by a vast array of quotidian software, starting with the ubiquitous Web browser. For this very reason, XHTML is in fact employed as a component part of several more specialized document types (ONIX and ePub among them).
  • Why re-invent a general-purpose prose representation when XHTML already does the job?
  • It is worth pausing for a moment to consider the role of XHTML in the ePub standard for ebook content. An ePub file is, anatomically, a simply disguised zip archive. Inside the zip archive are a few standard component parts: there are specialized files that declare metadata about the book, and about the format of the book. And then there is the book’s content, represented in XHTML. An ePub book is a Web page in a wrapper.
  • To sum up the general argument: the Web as it already exists presents incredible value to publishers, as a platform for doing XML content management with existing (and often free) tools, and without having to go blindly into the unknown. At this point, we can offer a few design guidelines: prefer existing and/or ubiquitous tools over specialized ones wherever possible; prefer free software over proprietary systems where possible; prefer simple tools controlled and coordinated by human beings over fully automated (and therefore complex) systems; play to our strengths: use Web software for storing and managing content, use layout software for layout, and keep editors and production people in charge of their own domains.
  • Putting the Pieces Together: A Prototype
  • At the SFU Master of Publishing Program, we have been chipping away at this general line of thinking for a few years. Over that time, Web content management systems have been getting more and more sophisticated, all the while getting more streamlined and easier to use. (NB: if you have a blog, you have a Web content management system.) The Web is beginning to be recognized as a writing and editing environment used by millions of people. And the ways in which content is represented, stored, and exchanged online have become increasingly robust and standardized.
  • The missing piece of the puzzle has been print production: how can we move content from its malleable, fluid form on line into the kind of high-quality print production environments we’ve come to expect after two decades of Desktop Publishing?
  • Anyone who has tried to print Web content knows that the existing methods leave much to be desired (hyphenation and justification, for starters). In the absence of decent tools for this, most publishers quite naturally think of producing the print content first, and then think about how to get material onto the Web for various purposes. So we tend to export from Word, or from Adobe, as something of an afterthought.
  • While this sort of works, it isn’t elegant, and it completely ignores the considerable advantages of Web-based content management.
  • Content managed online is stored in one central location, accessible simultaneously to everyone in your firm, available anywhere you have an Internet connection, and usually exists in a much more fluid format than Word files. If only we could manage the editorial flow online, and then go to print formats at the end, instead of the other way around. At SFU, we made several attempts to make this work by way of the supposed “XML import” capabilities of various Desktop Publishing tools, without much success.[12]
  • In the winter of 2009, Adobe solved this part of the problem for us with the introduction of its Creative Suite 4. What CS4 offers is the option of a complete XML representation of an InDesign document: what Adobe calls IDML (InDesign Markup Language).
  • The IDML file format is—like ePub—a simply disguised zip archive that, when unpacked, reveals a cluster of XML files that represent all the different facets of an InDesign document: layout spreads, master pages, defined styles, colours, and of course, the content.
  • IDML is a well thought-out XML standard that achieves two very different goals simultaneously: it preserves all of the information that InDesign needs to do what it does; and it is broken up in a way that makes it possible for mere mortals (or at least our Master of Publishing students) to work with it.
  • What this represented to us in concrete terms was the ability to take Web-based content and move it into InDesign in a straightforward way, thus bridging Web and print production environments using existing tools and skillsets, with a little added help from free software.
  • We would take clean XHTML content, transform it to IDML-marked content, and merge that with nicely designed templates in InDesign.
  • The result is an almost push-button publication workflow, which results in a nice, familiar InDesign document that fits straight into the way publishers actually do production.
  • Tracing the steps To begin with, we worked backwards, moving the book content back to clean XHTML.
  • The simplest method for this conversion—and if you want to create Web content, this is an excellent route—was to use Adobe’s “Export to Digital Editions” option, which creates an ePub file.
  • Recall that ePub is just XHTML in a wrapper, so within the ePub file was a relatively clean XHTML document. It was somewhat cleaner (that is, the XHTML tagging was simpler and less cluttered) than InDesign’s other Web-oriented exports, possibly because Digital Editions is a well understood target, compared with somebody’s website.
  • In order to achieve our target of clean XHTML, we needed to do some editing; the XHTML produced by InDesign’s “Digital Editions” export was presentation-oriented. For instance, bulleted list items were tagged as paragraphs, with a class attribute identifying them as list items. Using the search-and-replace function, we converted such structures to proper XHTML list and list-item elements. Our guiding principle was to make the XHTML as straightforward as possible, not dependent on any particular software to interpret it.
  • We broke the book’s content into individual chapter files; each chapter could then carry its own basic metadata, and the pages conveniently fit our Web content management system (which is actually just a wiki). We assembled a dynamically generated table of contents for the 12 chapters, and created a cover page. Essentially, the book was entirely Web-based at this point.
  • When the book chapters are viewed online, they are formatted via a CSS2 stylesheet that defines a main column for content as well as dedicating screen real estate for navigational elements. We then created a second template to render the content for exporting; this was essentially a bare-bones version of the book with no navigation and minimal styling. Pages (or even the entire book) can be exported (via the “Save As...” function in a Web browser) for use in either print production or ebook conversion. At this point, we required no skills beyond those of any decent Web designer.
  • Integrating with CS4 for Print Adobe’s IDML language defines elements specific to InDesign; there is nothing in the language that looks remotely like XHTML. So a mechanical transformation step is needed to convert the XHTML content into something InDesign can use. This is not as hard as it might seem.
  • Both XHTML and IDML are composed of straightforward, well-documented structures, and so transformation from one to the other is, as they say, “trivial.” We chose to use XSLT (Extensible Stylesheet Language Transforms) to do the work. XSLT is part of the overall XML specification, and thus is very well supported in a wide variety of tools. Our prototype used a scripting engine called xsltproc, a nearly ubiquitous piece of software that we found already installed as part of Mac OS X (contemporary Linux distributions also have this as a standard tool), though any XSLT processor would work.
  • In other words, we don’t need to buy InCopy, because we just replaced it with the Web. Our wiki is now plugged directly into our InDesign layout. It even automatically updates the InDesign document when the content changes. Credit is due at this point to Adobe: this integration is possible because of the open file format in the Creative Suite 4.
  • We wrote an XSLT transformation script[18] that converted the XHTML content from the Web into an InCopy ICML file. The script itself is less than 500 lines long, and was written and debugged over a period of about a week by amateurs (again, the people named at the start of this article). The script runs in a couple of seconds, and the resulting .icml file can then be “placed” directly into an InDesign template. The ICML file references an InDesign stylesheet, so the template file can be set up with a house-styled layout, master pages, and stylesheet definitions for paragraphs and character ranges.
  • Rather than a public-facing website, our system relies on the Web as a content management platform—of course a public face could easily be added.
  • It should be noted that the Book Publishing 1 proof-of-concept was artificially complex; we began with a book laid out in InDesign and ended up with a look-alike book laid out in InDesign. But next time—for instance, when we publish Book Publishing 2—we can begin the process with the content on the Web, and keep it there throughout the editorial process. The book’s content could potentially be written and edited entirely online, as Web content, and then automatically poured into an InDesign template at proof time. “Just in time,” as they say. This represents an entirely new way of thinking of book production. With a Web-first orientation, it makes little sense to think of the book as “in print” or “out of print”—the book is simply available, in the first place online; in the second place in derivative digital formats; and third, but really not much more difficult, in print-ready format, via the usual InDesign CS print production system publishers are already familiar with.
  • Creating Ebook Files Creating electronic versions from XHTML source is vastly simpler than trying to generate these out of the existing print process. The ePub version is extremely easy to generate; so is online marketing copy or excerpts for the Web, since the content begins life Web-native.
  • Since an ePub file is essentially XHTML content in a special wrapper, all that is required is that we properly “wrap” our XHTML content. Ideally, the content in an ePub file is broken into chapters (as ours was) and a table of contents file is generated in order to allow easy navigation within an ebook reader. We used Julian Smart’s free tool eCub[19] to simply and automatically generate the ePub wrapper and the table of contents. The only custom development we did was to create a CSS stylesheet for the ebook so that headings and paragraph indents looked the way we wanted. Starting with XHTML content, creating ePub is almost too easy.
  • today, we are able to put the process together using nothing but standard, relatively ubiquitous Web tools: the Web itself as an editing and content management environment, standard Web scripting tools for the conversion process, and the well-documented IDML file format to integrate the layout tool.
  • Our project demonstrates that Web technologies are indeed good enough to use in an XML-oriented workflow; more specialized and expensive options are not necessarily required. For massive-scale enterprise publishing, this approach may not offer enough flexibility, and the challenge of adding and extracting extra semantic richness may prove more trouble than it's worth.
  • But for smaller firms who are looking at the straightforward benefits of XML-based processes—single source publishing, online content and workflow management, open and accessible archive formats, greater online discoverability—here is a way forward.
  • The result is very simple and easy to use. Our demonstration requires that a production editor run the XSLT transformation script manually, but there is no reason why this couldn’t be built directly into the Web content management system so that exporting the content to print ran the transformation automatically. The resulting file would then be “placed” in InDesign and proofed.
  • The final piece of our puzzle, the ability to integrate print production, was made possible by Adobe's release of InDesign with an open XML file format. Since the Web's XHTML is also XML, is can be easily and confidently transformed to the InDesign format.
  • Such a workflow—beginning with the Web and exporting to print—is surely more in line with the way we will do business in the 21st century, where the Web is the default platform for reaching audiences, developing content, and putting the pieces together. It is time, we suggest, for publishers to re-orient their operations and start with the Web.
  • Using the Web as a Production Platform
  •  
    I was looking for an answer to a problem Marbux had presented, and found this interesting article.  The issue was that of the upcoming conversion of the Note Case Pro (NCP) layout engine to the WebKit layout engine, and what to do about the NCP document format. My initial reaction was to encode the legacy NCP document format in XML, and run an XSLT to a universal pivot format like TEI-XML.  From there, the TEI-XML community would provide all the XSLT transformation routines for conversion to ODF, OOXML, XHTML, ePUB and HTML/CSS. Researching the problems one might encounter with this approach, I found this article.  Fascinating stuff. My take away is that TEI-XML would not be as effective a "universal pivot point" as XHTML.  Or perhaps, if NCP really wants to get aggressive; IDML - InDesign Markup Language. As an after thought, i was thinking that an alternative title to this article might have been, "Working with Web as the Center of Everything".
Gary Edwards

Why a JavaScript hater thinks everyone needs to learn JavaScript in the next year - O'R... - 1 views

  • some extremely important game-changers: jQuery, JSON, Node.js, and HTML5.
  • .js has the potential to revolutionize web development. It is a framework for building high performance web applications: applications that can respond very quickly and efficiently to a high volume of incoming requests.
  • Google has started a revolution in JavaScript performance.
  • ...11 more annotations...
  • the number of JavaScript developers is huge.
  • HTML5 is about JavaScript
  • The power of HTML5 lies in what these tags allow you to create in JavaScript.
  • HTML5, then, isn't really a major advance in angle-bracket-based tagging; it's about enabling JavaScr
  • pt to do more powerful things
  • JavaScript has long been the workhorse for implementing dynamic features in HTML. But there have always been two problems: browser incompatibilities, and the awkwardness of working directly with the DOM. The JQuery library has elegantly solved both problems, and is the basis for modern client-side browser development.
  • The use of JavaScript has also exploded in databases.
  • document databases
  • for all three databases, a "document" means a JSON document, not a Word or Excel file.
  • JSON is really just a format for serializing JavaScript objects.
  • Web servers, rich web client libraries, HTML5, databases, even JavaScript-based languages: I see JavaScript everywhere.
  •  
    OK, this article gets my vote as the most important read of the year.  We all know that the the Web is the future of both computing and communications/connectivity.  But wha tis the future of the Web?  Uber coder Mike Loukides says it's JavaScript, and what a compelling case he builds.  This is a must read.  Key concepts are diigo highlighted :) excerpt: JavaScript has "grown up." I'm sure there are many JavaScript developers who would take issue with that judgement, and argue that JavaScript has been a capable, mature, and under-appreciated language all along. They may be right, though you can write any program in any complete programming language, including awful things like BASIC. What makes a language useful is some combination of the language's expressiveness and the libraries and tools available. JavaScript clearly passed the expressiveness barrier a long time ago, even if the ceremony required for creating objects is distasteful. But recently, we've seen some extremely important game-changers: jQuery, JSON, Node.js, and HTML5. JavaScript may have been a perfectly adequate language in the past, but these changes (and a few others that I'll point out) have made JavaScript a language that is essential for every developer to know. If there's one language you need to learn in the next year, it's JavaScript. Insightful comment: HTML5 is a JavaScript API, introducing new elements but significantly redefining ALL elements as objects or classes.  Elements can be expressed with tags.  Or, you can use DOM JavaScripting to create elements. 
Paul Merrell

Google to slip SVG into Internet Explorer * The Register - 0 views

  • Microsoft might be hesitating on Scalable Vector Graphics (SVG) in Internet Explorer 8, but Google's pressing on. The search giant's engineers are building a JavaScript library to render static and dynamic SVG in Microsoft's browser. Google promised that the library, a Javascript shim, will simply drop into IE.
  • SVG has a huge presence on the web. This facet of the World Wide Web Consortium's HTML 5 spec is supported in Firefox, Safari, Opera, Chrome, and Apple's iPhone, and is used in Google Maps and Google Docs. It also topped a list of features wanted by developers in a OpenAJAX browser wish list last year.
  • There's suspicion, though, that the reason has more to do with Microsoft's internal politics, with the company wanting graphics and drawing in IE done using Silverlight instead. SVG Web is more than an answer to Microsoft's foot-dragging, however. Google has declared for HTML 5 on the web, proclaiming last week that the web programming model has "won". Support for graphics capabilities in HTML 5 should also be seen as Google's partial answer to Adobe Systems' Flash. Google has complained that Flash is not open source and its development is not driven by the community. Google said the benefit of SVG Web is that it would sit inside the DOM whereas Flash "sits on top of the web, it's not part of the web"
Paul Merrell

News from The Associated Press - 0 views

  • (AP) -- Federal regulators are urging consumers to go through their phone bills line by line after they accused T-Mobile US of wrongly charging customers for premium services, like horoscope texts and quirky ringtones, the customers never authorized. The Federal Trade Commission announced Tuesday that it is suing T-Mobile in a federal court in Seattle with the goal of making sure every unfairly charged customer sees a full refund. The lawsuit, the first of its kind against a mobile provider, is the result of months of stalled negotiations with T-Mobile, which says it is already offering refunds. "It's wrong for a company like T-Mobile to profit from scams against its customers when there were clear warning signs the charges it was imposing were fraudulent," FTC Chair Edith Ramirez in a statement.
  • The practice is called "cramming": A third party stuffs a customer's bill with bogus charges such as $10-per-month horoscopes or updates on celebrity gossip. In this case, the FTC said, T-Mobile was working with third-party vendors being investigated by regulators and known to be the subject of numerous customer complaints. T-Mobile then made it difficult for customers to notice the added charge to their bill and pocketed up to 40 percent of the total, according to the FTC.
  • The FTC told reporters in a conference call Tuesday that it had been in negotiations with T-Mobile for months in an attempt to guarantee refunds would be provided to customers but that the two sides couldn't reach an agreement. T-Mobile appears to have been laying the groundwork to head off the federal complaint. Last November, the company announced that it would no longer allow premium text services because they were waning in popularity and not all vendors had acted responsibly. In June, it announced it would reach out to consumers to provide refunds. But the FTC says that in many cases, the refunds are only partial and T-Mobile often refers customer complaints to the third-party vendors.
1 - 20 of 23 Next ›
Showing 20 items per page