Skip to main content

Home/ Future of the Web/ Group items tagged hadoop

Rss Feed Group items tagged

Paul Merrell

The New York Times Archives + Amazon Web Services = TimesMachine - Open - Code - New Yo... - 0 views

  • TimesMachine is a collection of full-page image scans of the newspaper from 1851–1922 (i.e., the public domain archives). Organized chronologically and navigated by a simple calendar interface, TimesMachine provides a unique way to traverse the historical archives of The New York Times.
  • Using Amazon Web Services, Hadoop and our own code, we ingested 405,000 very large TIFF images, 3.3 million articles in SGML and 405,000 xml files mapping articles to rectangular regions in the TIFF’s. This data was converted to a more web-friendly 810,000 PNG images (thumbnails and full images) and 405,000 JavaScript files — all of it ready to be assembled into a TimesMachine. By leveraging the power of AWS and Hadoop, we were able to utilize hundreds of machines concurrently and process all the data in less than 36 hours.
  •  
    Like this http://www.hdfilmsaati.net Film,dvd,download,free download,product... ppc,adword,adsense,amazon,clickbank,osell,bookmark,dofollow,edu,gov,ads,linkwell,traffic,scor,serp,goggle,bing,yahoo.ads,ads network,ads goggle,bing,quality links,link best,ptr,cpa,bpa
Gonzalo San Gil, PhD.

Startup Crunches 100 Terabytes of Data in a Record 23 Minutes | WIRED - 0 views

  •  
    "There's a new record holder in the world of "big data." On Friday, Databricks-a startup spun out of the University California, Berkeley-announced that it has sorted 100 terabytes of data in a record 23 minutes using a number-crunching tool called Spark, eclipsing the previous record held by Yahoo and the popular big-data tool Hadoop."
1 - 2 of 2
Showing 20 items per page