Skip to main content

Home/ Memedia/ Group items tagged feed

Rss Feed Group items tagged

isaac Mao

Youtube被墙 - 0 views

  • 说什么来骂某些机构都不为过,草了。。。 只是一个视频分享网,只是一个图片分享网,只是一个feed合烧。。。草了,这些网站好好的,干着你们蛋事了! 一个个的封杀。。你们TM有没有真正为网民们想想。。。 我日了,已经不知道怎么组织自己的语言!心里只想找地方好好发泄一下!
isaac Mao

Digg,维基百科以及Web 2.0民主的迷思 - 译言翻译 - 0 views

  • 这并不是Digg和维基百科所想倡导的那种“人们一起工作”的场景。当然,维基百科需要一定程度的管理,否则网站仅应付对乔治·W·布什页面内容的添加、删除就不堪重负了。但这解释不了网站页面上的这种现象,即1%的贡献者主导了创作。难道运作一个开放的网站必然产生这种现象吗?或者是否可能在不给予精英用户和“神秘酱汁”太多权力的情况下建立一个高质量的、用户产生内容的场所呢?
shi zhao

走近天安门2.0 - 0 views

  • 天安门2.0是什么?是独立性,是良心,是依靠自己做简单好事,是天天坐在家,日复一日坐在电脑前面,跟老婆玩Wii,making babies,相信自己相信的事。也是一种healthy eating,很多写博客,作一个有人性的极客,了解极客的工具箱里有什么,怎样使用那些工具,知道为什么好奇不杀猫也不杀人,一个常识为王的状态,一个以rss feed为武器的kingdom,https为矛盾的空间,而Tor是曾金燕驾驶的车型的国家:
feng37

…My heart's in Accra » Studying Twitter and the Moldovan protests - 0 views

  • At some point on Friday, we hit a peak tweet density - 410 of 100,000 tweets included the #pman tag. Had I been scraping results by iterating 100,000 tweets at a time, I would have had four pages of new results - my script is only looking at the first page, so I’d be dropping results. If I ran the script again, I’d try to figure out the maximum tweet density by looking for the moment where the meme was most hyped, try to do a back of the envelope calculation as to an optimum step size and then halve it - that would probably have me using 20,000 steps for this set.
  • Density of tweets charted against blocks of 100,000 tweets
  • http://search.twitter.com/search?max_id=1511783811&page=2&q=%23pman&rpp=100 Picking apart the URL: max_id=1511783811 - Only return results up to tweet #1511783811 in the database page=2 - Hand over the second page of results q=%23pman - The query is for the string #pman, encoded to escape the hash rpp=100 - Give the user 100 results per page While you can manipulate these variables to your heart’s content, you can’t get more than 100 results per page. And if you retrieve 100 results per page, your results will stop at around 15 pages - the engine, by default, wants to give you only 1500 results on any search. This makes sense from a user perspective - it’s pretty rare that you actually want to read the last 1500 posts that mention the fail whale - but it’s a pain in the ass for researchers.
  • ...3 more annotations...
  • What you need to do is figure out the approximate tweet ID number that was current when the phenomenon you’re studying was taking place. If you’re a regular twitterer, go to your personal timeline, find a tweet you posted on April 7th, and click on the date to get the ID of the tweet. In the early morning (GMT) of the 7th, the ID for a new tweet was roughly 1468000000 - the URL http://search.twitter.com/search?max_id=1468000000&q=%23pman&rpp=100 retrieves the first four tweets to use the tag #pman, including our Ur-tweet: evisoft: neata, propun sa utilizam tag-ul #pman pentru mesajele din piata marii adunari nationale My Romanian’s a little rusty, but Vitalie Eşanu appears to be suggesting we use the tag #pman - short for Piata Marii Adunari Nationale, the main square in Chisinau where the protests were slated to begin - in reference to posts about the protests. His post is timestamped 4:40am GMT, suggesting that there were at least some discussions about promoting the protests on Twitter before protesters took to the streets.
  • Now the key is to grab URLs from Twitter, increasing the max_id variable in steps so that we’re getting all results from the start tweet ID to the current tweet ID. My perl script to do this steps by 10,000 results at a time, scraping the results I get from Twitter (using the Atom feed, not the HTML) and dumping novel results into a database. This seems like a pretty fine-toothed comb to use… but if you want to be comprehensive, it’s important to figure out what maximum “tweet density” is before running your code.
  •  
    http://search.twitter.com/search?max_id=1511783811&page=2&q=%23pman&rpp=100 Picking apart the URL: max_id=1511783811 - Only return results up to tweet #1511783811 in the database page=2 - Hand over the second page of results q=%23pman - The query is for the string #pman, encoded to escape the hash rpp=100 - Give the user 100 results per page While you can manipulate these variables to your heart's content, you can't get more than 100 results per page. And if you retrieve 100 results per page, your results will stop at around 15 pages - the engine, by default, wants to give you only 1500 results on any search. This makes sense from a user perspective - it's pretty rare that you actually want to read the last 1500 posts that mention the fail whale - but it's a pain in the ass for researchers. What you need to do is figure out the approximate tweet ID number that was current when the phenomenon you're studying was taking place. If you're a regular twitterer, go to your personal timeline, find a tweet you posted on April 7th, and click on the date to get the ID of the tweet. In the early morning (GMT) of the 7th, the ID for a new tweet was roughly 1468000000 - the URL http://search.twitter.com/search?max_id=1468000000&q=%23pman&rpp=100 retrieves the first four tweets to use the tag #pman, including our Ur-tweet: evisoft: neata, propun sa utilizam tag-ul #pman pentru mesajele din piata marii adunari nationale My Romanian's a little rusty, but Vitalie Eşanu appears to be suggesting we use the tag #pman - short for Piata Marii Adunari Nationale, the main square in Chisinau where the protests were slated to begin - in reference to posts about the protests. His post is timestamped 4:40am GMT, suggesting that there were at least some discussions about promoting the protests on Twitter before protesters took to the streets. Now the key is to grab URLs from Twitter, increasing the max_id variable in steps so that we're getting all results from the st
« First ‹ Previous 41 - 44 of 44
Showing 20 items per page