Skip to main content

Home/ beyondwebct/ Group items tagged data

Rss Feed Group items tagged

Barbara Lindsey

News: The Promise of Digital Humanities - Inside Higher Ed - 0 views

  • Amid financial crises, humanities departments at many public universities have been razed. But even amid cuts, there has been a surge in interest in the digital humanities -- a branch of scholarship that takes the computational rigor that has long undergirded the sciences and applies it the study of history, language, and culture.
  • The NEH held a symposium on Tuesday for 60 recipients of its 2011 Digital Humanities Start-Up Grants, most of whom were given between $25,000 and $50,000. They were allowed two minutes each to describe their projects.
  • “While we have been anguishing over the fate of the humanities, the humanities have been busily moving into, and even colonizing, the fields that were supposedly displacing them,”
  • ...9 more annotations...
  • role-playing games
  • enabling learners to “experience” historical events or places instead of reading off a page.
  • visual representations
  • of data
  • One recurring theme in the presentations was the need for “linked open data” — types of research data that are tagged and stored in such a way that they can integrate with other research.
  • If one researcher had architectural data about New York City, and another had demographic data about the city, and each were able to cross-reference the other’s data with her own, it would deepen the context and understanding for both.
  • With linked open data on the rise, the same could soon happen with research data
  • “Linked open data is a very technical infrastructure, but the result of that is information that’s shared widely for free. A lot of scholarly data over the last hundred years or so is locked up in expensive journals that the public could never afford to subscribe to.
  • That could be the key to winning back support for the humanities
  •  
    Could the creative use of technology help humanities scholars win back public support?
Barbara Lindsey

Dr. Mashup; or, Why Educators Should Learn to Stop Worrying and Love the Remix | EDUCAU... - 0 views

  • A classroom portal that presents automatically updated syndicated resources from the campus library, news sources, student events, weblogs, and podcasts and that was built quickly using free tools.
  • Increasingly, it's not just works of art that are appropriated and remixed but the functionalities of online applications as well.
  • mashups involve the reuse, or remixing, of works of art, of content, and/or of data for purposes that usually were not intended or even imagined by the original creators.
  • ...31 more annotations...
  • hat, exactly, constitutes a valid, original work? What are the implications for how we assess and reward creativity? Can a college or university tap the same sources of innovative talent and energy as Google or Flickr? What are the risks of permitting or opening up to this activity?
    • Barbara Lindsey
       
      Good discussion point
  • Remix is the reworking or adaptation of an existing work. The remix may be subtle, or it may completely redefine how the work comes across. It may add elements from other works, but generally efforts are focused on creating an alternate version of the original. A mashup, on the other hand, involves the combination of two or more works that may be very different from one another. In this article, I will apply these terms both to content remixes and mashups, which originated as a music form but now could describe the mixing of any number of digital media sources, and to data mashups, which combine the data and functionalities of two or more Web applications.
  • Harper's article "The Ecstasy of Influence," the novelist Jonathan Lethem imaginatively reviews the history of appropriation and recasts it as essential to the act of creation.3
  • Lethem's article is a must-read for anyone with an interest in the history of ideas, creativity, and intellectual property. It brilliantly synthesizes multiple disciplines and perspectives into a wonderfully readable and compelling argument. It is also, as the subtitle of his article acknowledges, "a plagiarism." Virtually every passage is a direct lift from another source, as the author explains in his "Key," which gives the source for every line he "stole, warped, and cobbled together." (He also revised "nearly every sentence" at least slightly.) Lethem's ideas noted in the paragraph above were appropriated from Siva Vaidhyanathan, Craig Baldwin, Richard Posner, and George L. Dillon.
  • Reading Walter Benjamin's highly influential 1936 essay "The Work of Art in the Age of Mechanical Reproduction,"4 it's clear that the profound effects of reproductive technology were obvious at that time. As Gould argued in 1964 (influenced by theorists such as Marshall McLuhan5), changes in how art is produced, distributed, and consumed in the electronic age have deep effects on the character of the art itself.
  • Yet the technology developments of the past century have clearly corresponded with a new attitude toward the "aura" associated with a work of invention and with more aggressive attitudes toward appropriation. It's no mere coincidence that the rise of modernist genres using collage techniques and more fragmented structures accompanied the emergence of photography and audio recording.
  • Educational technologists may wonder if "remix" or "content mashup" are just hipper-sounding versions of the learning objects vision that has absorbed so much energy from so many talented people—with mostly disappointing results.
  • The question is, why should a culture of remix take hold when the learning object economy never did?
  • when most learning object repositories were floundering, resource-sharing services such as del.icio.us and Flickr were enjoying phenomenal growth, with their user communities eagerly contributing heaps of useful metadata via simple folksonomy-oriented tagging systems.
  • the standards/practices relationship implicit in the learning objects model has been reversed. With only the noblest of intentions, proponents of learning objects (and I was one of them) went at the problem of promoting reuse by establishing an arduous and complex set of interoperability standards and then working to persuade others to adopt those standards. Educators were asked to take on complex and ill-defined tasks in exchange for an uncertain payoff. Not surprisingly, almost all of them passed.
  • Discoverable Resources
  • Educators might justifiably argue that their materials are more authoritative, reliable, and instructionally sound than those found on the wider Web, but those materials are effectively rendered invisible and inaccessible if they are locked inside course management systems.
  • It's a dirty but open secret that many courses in private environments use copyrighted third-party materials in a way that pushes the limits of fair use—third-party IP is a big reason why many courses cannot easily be made open.
  • The potential payoff for using open and discoverable resources, open and transparent licensing, and open and remixable formats is huge: more reuse means that more dynamic content is being produced more economically, even if the reuse happens only within an organization. And when remixing happens in a social context on the open web, people learn from each other's process.
  • Part of making a resource reusable involves making the right choices for file formats.
  • To facilitate the remixing of materials, educators may want to consider making the source files that were used to create a piece of multimedia available along with the finished result.
  • In addition to choosing the right file format and perhaps offering the original sources, another issue to consider when publishing content online is the critical question: "Is there an RSS feed available?" If so, conversion tools such as Feed2JS (http://www.feed2JS.org) allow for the republication of RSS-ified content in any HTML Web environment, including a course management system, simply by copying and pasting a few lines of JavaScript code. When an original source syndicated with RSS is updated, that update is automatically rendered anywhere it has been republished.
  • Jack Schofield
  • Guardian Unlimited
  • "An API provides an interface and a set of rules that make it much easier to extract data from a website. It's a bit like a record company releasing the vocals, guitars and drums as separate tracks, so you would not have to use digital processing to extract the parts you wanted."1
  • What's new about mashed-up application development? In a sense, the factors that have promoted this approach are the same ones that have changed so much else about Web culture in recent years. Essential hardware and software has gotten more powerful and for the most part cheaper, while access to high-speed connectivity and the enhanced quality of online applications like Google Docs have improved to the point that Tim O'Reilly and others can talk of "the emergent Internet operating system."15 The growth of user-centered technologies such as blogs have fostered a DIY ("do it yourself") culture that increasingly sees online interaction as something that can be personalized and adapted on the individual level. As described earlier, light syndication and service models such as RSS have made it easier and faster than ever to create simple integrations of diverse media types. David Berlind, executive editor of ZDNet, explains: "With mashups, fewer technical skills are needed to become a developer than ever. Not only that, the simplest ones can be done in 10 or 15 minutes. Before, you had to be a pretty decent code jockey with languages like C++ or Visual Basic to turn your creativity into innovation. With mashups, much the same way blogging systems put Web publishing into the hands of millions of ordinary non-technical people, the barrier to developing applications and turning creativity into innovation is so low that there's a vacuum into which an entire new class of developers will be sucked."16
  • The ability to "clone" other users' mashups is especially exciting: a newcomer does not need to spend time learning how to structure the data flows but can simply copy an existing framework that looks useful and then make minor modifications to customize the result.19
    • Barbara Lindsey
       
      This is the idea behind the MIT repository--remixing content to suit local needs.
  • As with content remixing, open access to materials is not just a matter of some charitable impulse to share knowledge with the world; it is a core requirement for participating in some of the most exciting and innovative activity on the Web.
  • "My Maps" functionality
  • For those still wondering what the value proposition is for offering an open API, Google's development process offers a compelling example of the potential rewards.
    • Barbara Lindsey
       
      Wikinomics
  • Elsewhere, it is difficult to point to significant activity suggesting that the mashup ethos is taking hold in academia the way it is on the wider Web.
  • Yet for the most part, the notion of the data mashup and the required openness is not even a consideration in discussions of technology strategy in higher educational institutions. "Data integration" across campus systems is something that is handled by highly skilled professionals at highly skilled prices.
  • Revealing how a more adventurous and inclusive online development strategy might look on campus, Raymond Yee recently posted a comprehensive proposal for his university (UC Berkeley), in which he outlined a "technology platform" not unlike the one employed by Amazon.com (http://aws.amazon.com/)—resources and access that would be invaluable for the institution's programmers as well as for outside interests to build complementary services.
  • All too often, college and university administrators react to this type of innovation with suspicion and outright hostility rather than cooperation.
  • those of us in higher education who observe the successful practices in the wider Web world have an obligation to consider and discuss how we might apply these lessons in our own contexts. We might ask if the content we presently lock down could be made public with a license specifying reasonable terms for reuse. When choosing a content management system, we might consider how well it supports RSS syndication. In an excellent article in the March/April 2007 issue of EDUCAUSE Review, Joanne Berg, Lori Berquam, and Kathy Christoph listed a number of campus activities that could benefit from engaging social networking technologies.26
  • What might happen if we allow our campus innovators to integrate their practices in these areas in the same way that social networking application developers are already integrating theirs? What is the mission-critical data we cannot expose, and what can we expose with minimal risk? And if the notion of making data public seems too radical a step, can APIs be exposed to selected audiences, such as on-campus developers or consortia partners?
Barbara Lindsey

Happy Data Privacy Day! - Digits - WSJ - 0 views

  •  
    January 28 is Data Privacy Day in the U.S., Canada and 27 European countries. Designed to make people aware of issues related to data privacy.
Barbara Lindsey

The New Gold Mine: Your Personal Information & Tracking Data Online - WSJ.com - 0 views

  • the tracking of consumers has grown both far more pervasive and far more intrusive than is realized by all but a handful of people in the vanguard of the industry. • The study found that the nation's 50 top websites on average installed 64 pieces of tracking technology onto the computers of visitors, usually with no warning. A dozen sites each installed more than a hundred. The nonprofit Wikipedia installed none.
  • the Journal found new tools that scan in real time what people are doing on a Web page, then instantly assess location, income, shopping interests and even medical conditions. Some tools surreptitiously re-spawn themselves even after users try to delete them. • These profiles of individuals, constantly refreshed, are bought and sold on stock-market-like exchanges that have sprung up in the past 18 months.
  • Advertisers once primarily bought ads on specific Web pages—a car ad on a car site. Now, advertisers are paying a premium to follow people around the Internet, wherever they go, with highly specific marketing messages.
  • ...22 more annotations...
  • "It is a sea change in the way the industry works," says Omar Tawakol, CEO of BlueKai. "Advertisers want to buy access to people, not Web pages."
  • The Journal found that Microsoft Corp.'s popular Web portal, MSN.com, planted a tracking file packed with data: It had a prediction of a surfer's age, ZIP Code and gender, plus a code containing estimates of income, marital status, presence of children and home ownership, according to the tracking company that created the file, Targus Information Corp.
  • Tracking is done by tiny files and programs known as "cookies," "Flash cookies" and "beacons." They are placed on a computer when a user visits a website. U.S. courts have ruled that it is legal to deploy the simplest type, cookies, just as someone using a telephone might allow a friend to listen in on a conversation. Courts haven't ruled on the more complex trackers.
  • tracking companies sometimes hide their files within free software offered to websites, or hide them within other tracking files or ads. When this happens, websites aren't always aware that they're installing the files on visitors' computers.
  • Often staffed by "quants," or math gurus with expertise in quantitative analysis, some tracking companies use probability algorithms to try to pair what they know about a person's online behavior with data from offline sources about household income, geography and education, among other things. The goal is to make sophisticated assumptions in real time—plans for a summer vacation, the likelihood of repaying a loan—and sell those conclusions.
  • Consumer tracking is the foundation of an online advertising economy that racked up $23 billion in ad spending last year. Tracking activity is exploding. Researchers at AT&T Labs and Worcester Polytechnic Institute last fall found tracking technology on 80% of 1,000 popular sites, up from 40% of those sites in 2005.
  • The Journal found tracking files that collect sensitive health and financial data. On Encyclopaedia Britannica Inc.'s dictionary website Merriam-Webster.com, one tracking file from Healthline Networks Inc., an ad network, scans the page a user is viewing and targets ads related to what it sees there.
    • Barbara Lindsey
       
      Tracking you an targeting ads to you on a popular dictionary site!
  • Beacons, also known as "Web bugs" and "pixels," are small pieces of software that run on a Web page. They can track what a user is doing on the page, including what is being typed or where the mouse is moving.
  • The majority of sites examined by the Journal placed at least seven beacons from outside companies. Dictionary.com had the most, 41, including several from companies that track health conditions and one that says it can target consumers by dozens of factors, including zip code and race.
  • After the Journal contacted the company, it cut the number of networks it uses and beefed up its privacy policy to more fully disclose its practices.
  • Flash cookies can also be used by data collectors to re-install regular cookies that a user has deleted. This can circumvent a user's attempt to avoid being tracked online. Adobe condemns the practice.
  • Most sites examined by the Journal installed no Flash cookies. Comcast.net installed 55.
  • Wittingly or not, people pay a price in reduced privacy for the information and services they receive online. Dictionary.com, the site with the most tracking files, is a case study.
  • Think about how these technologies and the associated analytics can be used in other industries and social settings (e.g. education) for real beneficial impacts. This is nothing new for the web, the now that it has matured, it can be a positive game-changer.
  • Media6Degrees Inc., whose technology was found on three sites by the Journal, is pitching banks to use its data to size up consumers based on their social connections. The idea is that the creditworthy tend to hang out with the creditworthy, and deadbeats with deadbeats.
  • "There are applications of this technology that can be very powerful," says Tom Phillips, CEO of Media6Degrees. "Who knows how far we'd take it?"
  • Hidden inside Ashley Hayes-Beaty's computer, a tiny file helps gather personal details about her, all to be put up for sale for a tenth of a penny.
  • "We can segment it all the way down to one person," says Eric Porres, Lotame's chief marketing officer.
  • One of the fastest-growing businesses on the Internet, a Wall Street Journal investigation has found, is the business of spying on Internet users.
  • Yahoo Inc.'s ad network,
  • "Every time I go on the Internet," she says, she sees weight-loss ads. "I'm self-conscious about my weight," says Ms. Reid, whose father asked that her hometown not be given. "I try not to think about it…. Then [the ads] make me start thinking about it."
  • Information about people's moment-to-moment thoughts and actions, as revealed by their online activity, can change hands quickly. Within seconds of visiting eBay.com or Expedia.com, information detailing a Web surfer's activity there is likely to be auctioned on the data exchange run by BlueKai, the Seattle startup.
  •  
    a New York company that uses sophisticated software called a "beacon" to capture what people are typing on a website
Barbara Lindsey

Convenience, Communications, and Control: How Students Use Technology | Resources | EDU... - 0 views

  • They are characterized as preferring teamwork, experiential activities, and the use of technology
  • Doing is more important than knowing, and learning is accomplished through trial and error as opposed to a logical and rule-based approach.2 Similarly, Paul Hagner found that these students not only possess the skills necessary to use these new communication forms, but there is an ever increasing expectation on their part that these new communication paths be used
  • Much of the work to date, while interesting and compelling, is intuitive and largely based on qualitative data and observation.
  • ...34 more annotations...
  • There is an inexorable trend among college students to universal ownership, mobility, and access to technology.
  • Students were asked about the applications they used on their electronic devices. They reported that they use technology first for educational purposes, followed by communication.
    • Barbara Lindsey
       
      All self-reported. Would have been powerful if could have actually tracked a representative sample and compared actual use with reported use.
  • presentation software was driven primarily by the requirements of the students' major and the curriculum.
  • Communications and entertainment are very much related to gender and age.
  • From student interviews, a picture emerged of student technology use driven by the demands of the major and the classes that students take. Seniors reported spending more time overall on a computer than do freshmen, and they reported greater use of a computer at a place of employment. Seniors spent more hours on the computer each week in support of their educational activities and also more time on more advanced applications—spreadsheets, presentations, and graphics.
  • Confirming what parents suspect, students with the lowest grade point averages (GPAs) spend significantly more time playing computer games; students with the highest GPAs spend more hours weekly using the computer in support of classroom activities. At the University of Minnesota, Crookston, students spent the most hours on the computer in support of classroom activities. This likely reflects the deliberate design of the curriculum to use a laptop extensively. In summary, the curriculum's technology requirements are major motivators for students to learn to use specialized software.
  • The interviews indicated that students are skilled with basic office suite applications but tend to know just enough technology functionality to accomplish their work; they have less in-depth application knowledge or problem solving skills.
  • According to McEuen, student technology skills can be likened to writing skills: Students come to college knowing how to write, but they are not developed writers. The analogy holds true for information technology, and McEuen suggested that colleges and universities approach information technology in the same way they approach writing.6
  • he major requires the development of higher-level skill sets with particular applications.
    • Barbara Lindsey
       
      Not really quantitative--self-reported data back by selected qualitative interviews
  • The comparative literature on student IT skill self-assessment suggests that students overrate their skills; freshmen overrate their skills more than seniors, and men overrate their skills more than women.7 Our data supports these conclusions. Judy Doherty, director of the Student Technologies Resource Group at Colgate University, remarked on student skill assessment, "Students state in their job applications that they are good if not very good, but when tested their skills are average to poor, and they need a lot of training."8
  • Mary Jane Smetanka of the Minneapolis–St. Paul Star Tribune reported that some students are so conditioned by punch-a-button problem solving on computers that they approach problems with a scattershot impulsiveness instead of methodically working them through. In turn, this leads to problem-solving difficulties.
  • We expected to find that the Net Generation student prefers classes that use technology. What we found instead is a bell curve with a preference for a moderate use of technology in the classroom (see Figure 1).
    • Barbara Lindsey
       
      More information needs to be given to find out why--may be tool and method not engaging.
  • It is not surprising that if technology is used well by the instructor, students will come to appreciate its benefits.
  • A student's major was also an important predictor of preferences for technology in the classroom (see Table 3), with engineering students having the highest preference for technology in the classroom (67.8 percent), followed by business students (64.3 percent).
  • Humanities 7.7% 47.9% 40.2
  • he highest scores were given to improved communications, followed by factors related to the management of classroom activities. Lower impact activities had to do with comprehension of classroom materials (complex concepts).
  • I spend more time engaged in course activities in those courses that require me to use technology.
  • The instructors' use of technology in my classes has increased my interest in the subject matter. 3.25 Classes that use information technology are more likely to focus on real-world tasks and examples.
  • Interestingly, students do not feel that use of information technology in classes greatly increases the amount of time engaged with course activities (3.22 mean).12 This is in direct contrast to faculty perceptions reported in an earlier study, where 65 percent of faculty reported they perceived that students spend more time engaged with course materials
  • Only 12.7 percent said the most valuable benefit was improved learning; 3.7 percent perceived no benefit whatsoever. Note that students could only select one response, so more than 12.7 percent may have felt learning was improved, but it was not ranked highest. These findings compare favorably with a study done by Douglas Havelka at the University of Miami in Oxford, Ohio, who identified the top six benefits of the current implementation of IT as improving work efficiency, affecting the way people behave, improving communications, making life more convenient, saving time, and improving learning ability.14
    • Barbara Lindsey
       
      Would have been good to know exactly what kinds of technologies were meant here.
  • Our data suggest that we are at best at the cusp of technologies being employed to improve learning.
  • The interactive features least used by faculty were the features that students indicated contributed the most to their learning.
  • he students in this study called our attention to performance by noting an uneven diffusion of innovation using this technology. This may be due, in part, to faculty or student skill. It may also be due to a lack of institutional recognition of innovation, especially as the successful use of course management systems affects or does not affect faculty tenure, promotion, and merit decisions
  • we found that many of the students most skilled in the use of technology had mixed feelings about technology in the classroom.
  • What we found was that many necessary skills had to be learned at the college or university and that the motivation for doing so was very much tied to the requirements of the curriculum. Similarly, the students in our survey had not gained the necessary skills to use technology in support of academic work outside the classroom. We found a significant need for further training in the use of information technology in support of learning and problem-solving skills.
  • Course management systems were used most by both faculty and students for communication of information and administrative activities and much less in support of learning.
  • In 1997, Michael Hooker proclaimed, "higher education is on the brink of a revolution." Hooker went on to note that two of the greatest challenges our institutions face are those of "harnessing the power of digital technology and responding to the information revolution."18 Hooker and many others, however, did not anticipate the likelihood that higher education's learning revolution would be a journey of a thousand miles rather than a discrete event. Indeed, a study of learning's last great revolution—the invention of moveable type—reveals, too, a revolution conducted over centuries leading to the emergence of a publishing industry, intellectual property rights law, the augmentation of customized lectures with textbooks, and so forth.
  • Both the ECAR study on faculty use of course management systems and this study of student experiences with information technology concluded that, while information technology is indeed making important inroads into classroom and learning activities, to date the effects are largely in the convenience of postsecondary teaching and learning and do not yet constitute a "learning revolution." This should not surprise us. The invention of moveable type enhanced, nearly immediately, access to published information and reduced the time needed to produce new publications. This invention did not itself change literacy levels, teaching styles, learning styles, or other key markers of a learning revolution. These changes, while catalyzed by the new technology, depended on slower social changes to institutions. I believe that is what we are witnessing in higher education today.
  • The institutions chosen represent a nonrepresentative mix of the different types of higher education institution in the United States, in terms of Carnegie class as well as location, source of funding, and levels of technology emphasis. Note, however, that we consider our findings to be instructive rather than conclusive of student experiences at different types of Carnegie institutions.
  • Qualitative data were collected by means of focus groups and individual interviews. We interviewed undergraduate students, administrators, and individuals identified as experts in the field of student technology use in the classroom. Student focus groups and interviews of administrators were conducted at six of the thirteen schools participating in the study.
Barbara Lindsey

A New First In Mobile: Data Traffic Outstripped Voice Traffic Last Year | paidContent - 0 views

  •  
    On a global scale, data traffic exceeds voice traffic on mobile phones.
Barbara Lindsey

If San Francisco Crime were Elevation | Doug McCune - 0 views

  • Really nice. Be great to see the two combined – heatmaps and topography or atleast some kind of colour banding added to the topography. That would open up all kinds of possibilities – you could slice horizontally along the bands and create layers of different ranges. In fact mixing colour and topography would also give you a way of showing two sets of data concurrently – topography for prostitution and some kind of colour banding for wealth for example.
  • Makes the numbers come alive. G
  • Brilliant work! Can you cross this data with the physical typography? I’ve always been curious if safer neighborhoods are uphill.
  • ...5 more annotations...
  • It would be interesting to pull the data in from previous decades and see how the elevation has changed in different areas.
  • @adrian – it’s just raw totals, grouped geographically. These aren’t scientific by any means, I basically took the underlying pattern and extruded it out and smoothed it a bit to make it look “pretty”. But basically each image is the aggregate numbers for a single year of crime data.
  • @richard – yes, there is some smoothing in effect, which means that the ridge along Shotwell St (for the prostitution map) is indeed a bit smoothed between peaks. That’s not to say that there are only two peaks at Shotwell and 19th and Shotwell and 17th. There are incidents in between as well, but the big peaks at those major intersections does mean that the ridge between them appears higher than the actual incidents along those blocks support. A lot of people have commented on the usefulness of maps like these. I want to stress once again: this was done as an art project much more than a useful visualization. My goal was not to provide useful information that one could act on.
  • “one trick pony. these maps add nothing of value to a standard color plot.” I disagree: allowing for a third dimension of elevation makes the reality of concentration clearer – and half the point of crime mapping is to measure concentration, not simply “intensity.”
  • Great idea and nice work on the graphics, but there are at least three improvements you should make to reveal *true* patterns. Forgive me if you already did these. 1) Availability bias – normalize for population density (i.e. per capita activity) 2) Sampling bias – normalize for the number of cops on the beat (geographic and crime type) 2) Frame bias – break it up by daytime and night time
  •  
    Visual representation of various crime stats from San Francisco
Barbara Lindsey

Online version: Open Data, Democracy and Public Sector Reform : Tim's Blog - 0 views

  •  
    Over the weeks since I handed in my MSc Dissertation I've been trying to work out how best to share the final version. Each time I've started to edit it for release I've found more areas where I want to develop the argument further, or where I recognise that points I thought were conclusions are in fact the start of new questions. After trying out a few options, I settled on the fantastic Digress.it platform to put a copy of the report online - giving each paragraph it's own URL and space for comments and trackbacks. Hopefully this can help turn a static dissertation into something more dynamic as a tool for helping take forward thinking about the impacts of open government data. All comments, feedback, reflections and thinking aloud on the document welcome.
Barbara Lindsey

Web 2.0: What does it constitute? | 11 Feb 2008 | ComputerWeekly.com - 0 views

  • O'Reilly identified Google as "the standard bearer for Web 2.0", and pointed out the differences between it and predecessors such as Netscape, which tried to adapt for the web the business model established by Microsoft and other PC software suppliers.
  • Google "began its life as a native web application, never sold or packaged, but delivered as a service, with customers paying, directly or indirectly.
  • perpetual beta, as O'Reilly later dubbed it
  • ...13 more annotations...
  • Perhaps the most important breakthrough was Google's willingness to relinquish control of the user-end of the transaction, instead of trying to lock them in with proprietary technology and restrictive licensing
  • O'Reilly took a second Web 2.0 principle from Peer-to-Peer pioneer BitTorrent, which works by completely decentralising the delivery of files, with every client also functioning as a server. The more popular a file, is, the faster it can be served, since there are more users providing bandwidth and fragments of the file. Thus, "the service automatically gets better the more people use it".
  • Taking another model from open source, users are treated as "co-developers", actively encouraged to contribute, and monitored in real time to see what they are using, and how they are using it.
  • "Until Web 2.0 the learning curve to creating websites was quite high, complex, and a definite barrier to entry," says the third of our triumvirate of Tims, Tim Bray, director of Web Technologies at Sun Microsystems.
  • Web 2.0 takes some of its philosophical underpinning from James Surowiecki's book The Wisdom of Crowds, which asserts that the aggregated insights of large groups of diverse people can provide better answers and innovations than individual experts.
  • In practice, even fewer than 1% of people may be making a useful contribution - but these may be the most energetic and able members of a very large community. In 2006 1,000 people, just 0.003% of its users, contributed around two-thirds of Wikipedia's edits.
  • Ajax speeds up response times by enabling just part of a page to be updated, instead of downloading a whole new page. Nielsen's objections include that this breaks the "back" button - the ability to get back to where you've been, which Nielsen says is the second most used feature in Web navigation.
  • "Everybody who has a Web browser has got that platform," says Berners-Lee, in a podcast available on IBM's developerWorks site. "So the nice thing about it is when you do code up an Ajax implementation, other people can take it and play with it."
  • Web 2.0 is a step on the way to the Semantic Web, a long-standing W3C initiative to create a standards-based framework able to understand the links between data which is related in the real world, and follow that data wherever it resides, regardless of application and database boundaries.
  • The problem with Web 2.0, Pemberton says, is that it "partitions the web into a number of topical sub-webs, and locks you in, thereby reducing the value of the network as a whole."
  • How do you decide which social networking site to join? he asks. "Do you join several and repeat the work?" With the Semantic Web's Resource Description Framework (RDF), you won't need to sign up to separate networks, and can keep ownership of your data. "You could describe it as a CSS for meaning: it allows you to add a small layer of markup to your page that adds machine-readable semantics."
  • The problems with Web 2.0 lock-in which Pemberton describes, were illustrated when a prominent member of the active 1%, Robert Scoble, ran a routine called Plaxo to try to extract details of his 5,000 contacts from Facebook, in breach of the site's terms of use, and had his account disabled. Although he has apparently had his account reinstated, the furore has made the issue of Web 2.0 data ownership and portability fiercely topical.
  • when Google announced its OpenSocial set of APIs, which will enable developers to create portable applications and bridges between social networking websites, Facebook was not among those taking part. Four years after O'Reilly attempted to define Web 2.0, Google, it seems, remains the standard-bearer, while others are forgetting what it was supposed to be about.
Barbara Lindsey

What we learned from 5 million books | Video on TED.com - 0 views

    • Barbara Lindsey
       
      From YouTube version of this talk: "[Google's digtized books] are very practical and extremely awesome." Erez Lieberman Aiden and Jean-Baptiste Michel from Harvard University use the 15 million books scanned and digitized by Google to show how a visual and quantitative analysis of text can provide insights about fields as diverse as lexicography, the evolution of grammar, collective memory, the adoption of technology, the pursuit of fame, censorship, and historical epidemiology.
  • ELA: There are more sobering notes among the n-grams. For instance, here's the trajectory of Marc Chagall, an artist born in 1887. And this looks like the normal trajectory of a famous person. He gets more and more and more famous, except if you look in German. If you look in German, you see something completely bizarre, something you pretty much never see, which is he becomes extremely famous and then all of a sudden plummets, going through a nadir between 1933 and 1945, before rebounding afterward. And of course, what we're seeing is the fact Marc Chagall was a Jewish artist in Nazi Germany. Now these signals are actually so strong that we don't need to know that someone was censored. We can actually figure it out using really basic signal processing. Here's a simple way to do it. Well, a reasonable expectation is that somebody's fame in a given period of time should be roughly the average of their fame before and their fame after. So that's sort of what we expect. And we compare that to the fame that we observe. And we just divide one by the other to produce something we call a suppression index. If the suppression index is very, very, very small, then you very well might be being suppressed. If it's very large, maybe you're benefiting from propaganda.
  • Now when Google digitizes a book, they put it into a really nice format. Now we've got the data, plus we have metadata. We have information about things like where was it published, who was the author, when was it published. And what we do is go through all of those records and exclude everything that's not the highest quality data. What we're left with is a collection of five million books, 500 billion words, a string of characters a thousand times longer than the human genome -- a text which, when written out, would stretch from here to the Moon and back 10 times over -- a veritable shard of our cultural genome.
  • ...4 more annotations...
  • we're going to release statistics about the books. So take for instance "A gleam of happiness." It's four words; we call that a four-gram. We're going to tell you how many times a particular four-gram appeared in books in 1801, 1802, 1803, all the way up to 2008. That gives us a time series of how frequently this particular sentence was used over time. We do that for all the words and phrases that appear in those books, and that gives us a big table of two billion lines that tell us about the way culture has been changing.
  • You might also want to have a look at this particular n-gram, and that's to tell Nietzsche that God is not dead, although you might agree that he might need a better publicist.
  • JM: Now you can actually look at the distribution of suppression indexes over whole populations. So for instance, here -- this suppression index is for 5,000 people picked in English books where there's no known suppression -- it would be like this, basically tightly centered on one. What you expect is basically what you observe. This is distribution as seen in Germany -- very different, it's shifted to the left. People talked about it twice less as it should have been. But much more importantly, the distribution is much wider. There are many people who end up on the far left on this distribution who are talked about 10 times fewer than they should have been. But then also many people on the far right who seem to benefit from propaganda. This picture is the hallmark of censorship in the book record.
  • ELA: So culturomics is what we call this method. It's kind of like genomics. Except genomics is a lens on biology through the window of the sequence of bases in the human genome. Culturomics is similar. It's the application of massive-scale data collection analysis to the study of human culture. Here, instead of through the lens of a genome, through the lens of digitized pieces of the historical record. The great thing about culturomics is that everyone can do it. Why can everyone do it? Everyone can do it because three guys, Jon Orwant, Matt Gray and Will Brockman over at Google, saw the prototype of the Ngram Viewer, and they said, "This is so fun. We have to make this available for people." So in two weeks flat -- the two weeks before our paper came out -- they coded up a version of the Ngram Viewer for the general public. And so you too can type in any word or phrase that you're interested in and see its n-gram immediately -- also browse examples of all the various books in which your n-gram appears.
  •  
    fall 2012 syllabus
Barbara Lindsey

Educator's Voice: Data is the Foundation for Progress | Pearson Academic Executives - 0 views

  • which provide us with endless sources of information on student and faculty behaviors. This data can then be mined for clues on in-course retention, program persistence, quality of student learning, and admission demographics correlated to student success.
    • Barbara Lindsey
       
      Does anyone else find this disconcerting?
Barbara Lindsey

Colleges Awakening to the Opportunities of Data Mining - NYTimes.com - 0 views

  •  
    fall 2012 syllabus
Barbara Lindsey

Opportunities for Creating the Future of Learning - 2020 Forecast: Creating the Future ... - 0 views

  • It remains to be seen whether new learning agents and traditionally certified teachers will cooperate or compete.
  • Secondly, it emphasizes the need for learning to be an ongoing process whereby we all become engaged citizens of a global society. T
  • By embracing technologies of cooperation, prototyping new models of learning, and cultivating open and collaborative approaches to leadership, “amplified” educators and learners will become the organizational “superheroes” of schools and districts.
  • ...5 more annotations...
  • The globalization of open learning systems characterized by cooperative resource creation, evaluation, and sharing will change how educational institutions view their roles and will offer new forms of value in the global learning ecosystem.
  • The result will be an emerging toolset for designing personalized, learner-centered experiences and environments that reflect the differentiation among learners instead of forcing compliance to an average learning style and level of performance.
  • As the hierarchical structure of education splinters, traditional top-down movements of authority, knowledge, and power will unravel. Before new patterns get established, it will seem as if a host of new species has been introduced into the learning ecosystem. Authority will be a hotly contested resource, and there will be the potential for conflict and distrust.
  • Learning geographies will be accessible to communities through a range of key tools, such as data aggregated from disparate sources, geo-coded data linking learning resources and educational information to specific community locations, and visualization tools that help communicate such information in easily understood visual and graphic forms. Such information will often contain multiple layers of data (for example, school performance statistics, poverty rates, and the degree of access to fresh food).
  • These new dimensions of learning geographies will require new core skills. Among them will be navigating new visual cartographies, identifying learning resources in previously unexpected places, leveraging networks to take advantage of learning opportunities, and creating flexible educational infrastructures that can make use of dispersed community resources. Through enhanced visibility and accessibility, learning geographies will bring new transparency to issues of equity in learning.
  •  
    By embracing technologies of cooperation, prototyping new models of learning, and cultivating open and collaborative approaches to leadership, "amplified" educators and learners will become the organizational "superheroes" of schools and districts.
Barbara Lindsey

AOL Proudly Releases Massive Amounts of Private Data - 0 views

  •  
    Anonymous searches aren't really...
Barbara Lindsey

What is Pivot? - 0 views

  •  
    Pivot makes it easier to interact with massive amounts of data in ways that are powerful, informative, and fun. We tried to step back and design an interaction model that accommodates the complexity and scale of information rather than the traditional structure of the Web.
Barbara Lindsey

For Teachers - Gapminder.org - 0 views

  •  
    Great set of resources for interdisciplinary activities that make world data accessible.
Barbara Lindsey

Data Governance - Data Intensive Science - by IdeaScale - 0 views

  •  
    fall 2011 syllabus
1 - 20 of 87 Next › Last »
Showing 20 items per page