Skip to main content

Home/ Instructional & Media Services at Dickinson College/ Group items tagged search

Rss Feed Group items tagged

Ed Webb

The Generative AI Race Has a Dirty Secret | WIRED - 0 views

  • The race to build high-performance, AI-powered search engines is likely to require a dramatic rise in computing power, and with it a massive increase in the amount of energy that tech companies require and the amount of carbon they emit.
  • Every time we see a step change in online processing, we see significant increases in the power and cooling resources required by large processing centres
  • third-party analysis by researchers estimates that the training of GPT-3, which ChatGPT is partly based on, consumed 1,287 MWh, and led to emissions of more than 550 tons of carbon dioxide equivalent—the same amount as a single person taking 550 roundtrips between New York and San Francisco
  • ...3 more annotations...
  • There’s also a big difference between utilizing ChatGPT—which investment bank UBS estimates has 13 million users a day—as a standalone product, and integrating it into Bing, which handles half a billion searches every day.
  • Data centers already account for around one percent of the world’s greenhouse gas emissions, according to the International Energy Agency. That is expected to rise as demand for cloud computing increases, but the companies running search have promised to reduce their net contribution to global heating. “It’s definitely not as bad as transportation or the textile industry,” Gómez-Rodríguez says. “But [AI] can be a significant contributor to emissions.”
  • The environmental footprint and energy cost of integrating AI into search could be reduced by moving data centers onto cleaner energy sources, and by redesigning neural networks to become more efficient, reducing the so-called “inference time”—the amount of computing power required for an algorithm to work on new data.
Ed Webb

Oxford University Press launches the Anti-Google - 0 views

  • he Anti-Google: Oxford Bibliographies Online (OBO)
  • essentially a straightforward, hyperlinked collection of professionally-produced, peer-reviewed bibliographies in different subject areas—sort of a giant, interactive syllabus put together by OUP and teams of scholars in different disciplines
  • "You can't come up with a search filter that solves the problem of information overload," Zucca told Ars. OUP is betting that the solution to the problem lies in content, which is its area of expertise, and not in technology, which is Google's and Microsoft's.
  • ...3 more annotations...
  • at least users can see exactly how the sausage is made. Contrast this to Google or Bing, where the search algorithm that produces results is a closely guarded secret.
  • The word that Zucca used a number of times in our chat was "authority," and OUP is betting that individual and institutional users will value the authority enough that they'll be willing to pay for access to the service
  • This paywall is the only feature of OBO that seems truly unfortunate, given that the competition (search and Wikipedia) is free. High school kids and motivated amateurs will be left slumming it with whatever they can get from the public Internet, and OBO's potential reach and impact will be severely limite
Ed Webb

Search Engine Helps Users Connect In Arabic : NPR - 0 views

  • new technology that is revolutionizing the way Arabic-speaking people use the Internet
  • Abdullah says that of her 500 Egyptian students, 78 percent have never typed in Arabic online, a fact that greatly disturbed Habib Haddad, a Boston-based software engineer originally from Lebanon. "I mean imagine [if] 78 percent of French people don't type French," Haddad says. "Imagine how destructive that is online."
  • "The idea is, if you don't have an Arabic keyboard, you can type Arabic by spelling your words out phonetically," Jureidini says. "For example ... when you're writing the word 'falafel,' Yamli will convert that to Arabic in your Web browser. We will go and search not only the Arabic script version of that search query, but also for all the Western variations of that keyword."
  • ...1 more annotation...
  • At a recent "new" technology forum at MIT, Yamli went on to win best of show — a development that did not escape the attention of Google, which recently developed its own search and transliteration engine. "I guess Google recognizes a good idea when it sees it," Jureidini says. He adds, "And the way we counter it is by being better. We live and breathe Yamli every day, and we're constantly in the process of improving how people can use it." Experts in Arabic Web content say that since its release a year ago, Yamli has helped increase Arabic content on the Internet just by its use. They say that bodes well for the Arabic Web and for communication between the Arab and Western worlds.
Ed Webb

Professors Find Ways to Keep Heads Above 'Exaflood' of Data - Wired Campus - The Chroni... - 0 views

  • Google, a major source of information overload, can also help manage it, according to Google's chief economist. Hal Varian, who was a professor at the University of California at Berkeley before going to work for the search-engine giant, showed off an analytic tool called Google Insights for Search.
  • accurately tagging data and archiving it
Ed Webb

Google pushes journalists to create G+ profiles · kegill · Storify - 0 views

  • linking search results with Google+ was like Microsoft bundling Internet Explore with Windows
  • Market strength in one place being used to leverage sub optimal products in another.
  • It's time to tell both Google and Bing that we want to decide for ourselves, thank you very much, if content is credible, instead of their making those decisions for us, decisions made behind hidden -- and suspicious -- algorithms.
Ed Webb

The powerful and mysterious brain circuitry that makes us love Google, Twitter, and tex... - 0 views

  • For humans, this desire to search is not just about fulfilling our physical needs. Panksepp says that humans can get just as excited about abstract rewards as tangible ones. He says that when we get thrilled about the world of ideas, about making intellectual connections, about divining meaning, it is the seeking circuits that are firing.
  • Our internal sense of time is believed to be controlled by the dopamine system. People with hyperactivity disorder have a shortage of dopamine in their brains, which a recent study suggests may be at the root of the problem. For them even small stretches of time seem to drag.
  • When we get the object of our desire (be it a Twinkie or a sexual partner), we engage in consummatory acts that Panksepp says reduce arousal in the brain and temporarily, at least, inhibit our urge to seek.
  • ...3 more annotations...
  • But our brains are designed to more easily be stimulated than satisfied. "The brain seems to be more stingy with mechanisms for pleasure than for desire," Berridge has said. This makes evolutionary sense. Creatures that lack motivation, that find it easy to slip into oblivious rapture, are likely to lead short (if happy) lives. So nature imbued us with an unquenchable drive to discover, to explore. Stanford University neuroscientist Brian Knutson has been putting people in MRI scanners and looking inside their brains as they play an investing game. He has consistently found that the pictures inside our skulls show that the possibility of a payoff is much more stimulating than actually getting one.
  • all our electronic communication devices—e-mail, Facebook feeds, texts, Twitter—are feeding the same drive as our searches. Since we're restless, easily bored creatures, our gadgets give us in abundance qualities the seeking/wanting system finds particularly exciting. Novelty is one. Panksepp says the dopamine system is activated by finding something unexpected or by the anticipation of something new. If the rewards come unpredictably—as e-mail, texts, updates do—we get even more carried away. No wonder we call it a "CrackBerry."
  • If humans are seeking machines, we've now created the perfect machines to allow us to seek endlessly. This perhaps should make us cautious. In Animals in Translation, Temple Grandin writes of driving two indoor cats crazy by flicking a laser pointer around the room. They wouldn't stop stalking and pouncing on this ungraspable dot of light—their dopamine system pumping. She writes that no wild cat would indulge in such useless behavior: "A cat wants to catch the mouse, not chase it in circles forever." She says "mindless chasing" makes an animal less likely to meet its real needs "because it short-circuits intelligent stalking behavior." As we chase after flickering bits of information, it's a salutary warning.
Ed Webb

It's Time To Hide The Noise - 0 views

  • the noise is worse than ever. Indeed, it is being magnified every day as more people pile onto Twitter and Facebook and new apps yet to crest like Google Wave. The data stream is growing stronger, but so too is the danger of drowning in all that information.
  • the fact that Seesmic or TweetDeck or any of these apps can display 1,200 Tweets at once is not a feature, it’s a bug
  • if you think Twitter is noisy, wait until you see Google Wave, which doesn’t hide anything at all.  Imagine that Twhirl image below with a million dialog boxes on your screen, except you see as other people type in their messages and add new files and images to the conversation, all at once as it is happening.  It’s enough to make your brain explode.
  • ...2 more annotations...
  • all I need is two columns: the most recent Tweets from everyone I follow (the standard) and the the most interesting tweets I need to pay attention to.  Recent and Interesting.  This second column is the tricky one.  It needs to be automatically generated and personalized to my interests at that moment.
  • search is broken on Twitter.  Unless you know the exact word you are looking for, Tweets with related terms won’t show up.  And there is no way to sort searches by relevance, it is just sorted by chronology.
Ed Webb

Lacktribution: Be Like Everyone Else - CogDogBlog - 0 views

  • What exactly are the issues about attributing? Why is it good to not have to attribute? Is it a severe challenge to attribute? Does it hurt? Does it call for technical or academic skills beyond reach? Does it consume great amounts of time, resources? Why, among professional designers and technologists is it such a good thing to be free of this odious chore? I can translate this typical reason to use public domain content, “I prefer to be lazy.”
  • There is a larger implication when you reuse content and choose not to attribute. Out in the flow of all other information, it more or less says to readers, “all images are free to pilfer. Just google and take them all. Be like me.”
  • It’s not about the rules of the license, it’s about maybe, maybe, operating in this mechanized place as a human, rather than a copy cat.
  • ...4 more annotations...
  • Google search results gives more weight to pxhere.com where the image has a mighty 4 views (some of which are me) over the original image, with almost 5000 views.
  • What kind of algorithm is that? It’s one that does not favor the individual. Image search results will favor sites like Needpix, Pixsels, Pixnio, Peakpx, Nicepic, and they still favor the really slimy maxpixel which is a direct rip off of pixabay.
  • did you know that the liberating world of “use any photo you want w/o the hassle of attribution” is such a bucket of questionable slime? And that Google, with all of their algorithmic prowess, gives more favorable results to sites that lift photos than to the ones where the originals exist?
  • So yes, just reuse photos without taking all of the severe effort to give credit to the source, because “you don’t have to.” Be a copycat. Show your flag of Lacktribution. Like everyone else. I will not. I adhere to Thanktribution.
Ed Webb

Google and Meta moved cautiously on AI. Then came OpenAI's ChatGPT. - The Washington Post - 0 views

  • The surge of attention around ChatGPT is prompting pressure inside tech giants including Meta and Google to move faster, potentially sweeping safety concerns aside
  • Tech giants have been skittish since public debacles like Microsoft’s Tay, which it took down in less than a day in 2016 after trolls prompted the bot to call for a race war, suggest Hitler was right and tweet “Jews did 9/11.”
  • Some AI ethicists fear that Big Tech’s rush to market could expose billions of people to potential harms — such as sharing inaccurate information, generating fake photos or giving students the ability to cheat on school tests — before trust and safety experts have been able to study the risks. Others in the field share OpenAI’s philosophy that releasing the tools to the public, often nominally in a “beta” phase after mitigating some predictable risks, is the only way to assess real world harms.
  • ...8 more annotations...
  • Silicon Valley’s sudden willingness to consider taking more reputational risk arrives as tech stocks are tumbling
  • A chatbot that pointed to one answer directly from Google could increase its liability if the response was found to be harmful or plagiarized.
  • AI has been through several hype cycles over the past decade, but the furor over DALL-E and ChatGPT has reached new heights.
  • Soon after OpenAI released ChatGPT, tech influencers on Twitter began to predict that generative AI would spell the demise of Google search. ChatGPT delivered simple answers in an accessible way and didn’t ask users to rifle through blue links. Besides, after a quarter of a century, Google’s search interface had grown bloated with ads and marketers trying to game the system.
  • Inside big tech companies, the system of checks and balances for vetting the ethical implications of cutting-edge AI isn’t as established as privacy or data security. Typically teams of AI researchers and engineers publish papers on their findings, incorporate their technology into the company’s existing infrastructure or develop new products, a process that can sometimes clash with other teams working on responsible AI over pressure to see innovation reach the public sooner.
  • Chatbots like OpenAI routinely make factual errors and often switch their answers depending on how a question is asked
  • To Timnit Gebru, executive director of the nonprofit Distributed AI Research Institute, the prospect of Google sidelining its responsible AI team doesn’t necessarily signal a shift in power or safety concerns, because those warning of the potential harms were never empowered to begin with. “If we were lucky, we’d get invited to a meeting,” said Gebru, who helped lead Google’s Ethical AI team until she was fired for a paper criticizing large language models.
  • Rumman Chowdhury, who led Twitter’s machine-learning ethics team until Elon Musk disbanded it in November, said she expects companies like Google to increasingly sideline internal critics and ethicists as they scramble to catch up with OpenAI.“We thought it was going to be China pushing the U.S., but looks like it’s start-ups,” she said.
Ed Webb

Keep the 'Research,' Ditch the 'Paper' - Commentary - The Chronicle of Higher Education - 1 views

  • we need to construct meaningful opportunities for students to actually engage in research—to become modest but real contributors to the research on an actual question. When students write up the work they’ve actually performed, they create data and potential contributions to knowledge, contributions that can be digitally published or shared with a target community
  • Schuman’s critique of traditional writing instruction is sadly accurate. The skill it teaches most students is little more than a smash-and-grab assault on the secondary literature. Students open a window onto a search engine or database. They punch through to the first half-dozen items. Snatching random gems that seem to support their preconceived thesis, they change a few words, cobble it all together with class notes in the form of an argument, and call it "proving a thesis."
  • What happens when a newly employed person tries to pass off quote-farmed drivel as professional communication?
  • ...6 more annotations...
  • Generally these papers are just pumped-up versions of the five-paragraph essay, with filler added. Thesis-driven, argumentative, like the newspaper editorials the genre is based on, this "researched writing" promises to solve big questions with little effort: "Reproductive rights resolved in five pages!"
  • Actual writing related to research is modest, qualified, and hesitant
  • our actual model involves elaborately respectful conversation, demonstrating sensitivity to the most nuanced claims of previous researchers
  • Academic, legal, medical, and business writing has easily understandable conventions. We responsibly survey the existing literature, formally or informally creating an annotated bibliography. We write a review of the literature, identifying a "blank" spot ignored by other scholars, or a "bright" spot where we see conflicting evidence. We describe the nature of our research in terms of a contribution to the blank or bright spot in that conversation. We conclude by pointing to further questions.
  • Millions of pieces of research writing that aren’t essays usefully circulate in the profession through any number of sharing technologies, including presentations and posters; grant and experiment proposals; curated, arranged, translated, or visualized data; knowledgeable dialogue in online media with working professionals; independent journalism, arts reviews, and Wikipedia entries; documentary pitches, scripts and storyboards; and informative websites.
  • real researchers don’t write a word unless they have something to contribute. We should teach our students to do the same
Ed Webb

MillionShort - 1 views

test and tagging

Excellent Test and Tagging in Adelaide - 1 views

I have been looking for a reliable electrical safety specialist to check on my electrical equipment which we have been using in my restaurant in Adelaide. After a week of searching, I finally found...

test and tagging

started by test and tagging on 24 Nov 11 no follow-up yet
Ed Webb

Bad News : CJR - 0 views

  • Students in Howard Rheingold’s journalism class at Stanford recently teamed up with NewsTrust, a nonprofit Web site that enables people to review and rate news articles for their level of quality, in a search for lousy journalism.
  • the News Hunt is a way of getting young journalists to critically examine the work of professionals. For Rheingold, an influential writer and thinker about the online world and the man credited with coining the phrase “virtual community,” it’s all about teaching them “crap detection.”
  • last year Rheingold wrote an important essay about the topic for the San Francisco Chronicle’s Web site
  • ...3 more annotations...
  • What’s at stake is no less than the quality of the information available in our society, and our collective ability to evaluate its accuracy and value. “Are we going to have a world filled with people who pass along urban legends and hoaxes?” Rheingold said, “or are people going to educate themselves about these tools [for crap detection] so we will have collective intelligence instead of misinformation, spam, urban legends, and hoaxes?”
  • I previously called fact-checking “one of the great American pastimes of the Internet age.” But, as Rheingold noted, the opposite is also true: the manufacture and promotion of bullshit is endemic. One couldn’t exist without the other. That makes Rheingold’s essay, his recent experiment with NewsTrust, and his wiki of online critical-thinking tools” essential reading for journalists. (He’s also writing a book about this topic.)
  • I believe if we want kids to succeed online, the biggest danger is not porn or predators—the biggest danger is them not being able to distinguish truth from carefully manufactured misinformation or bullshit
  •  
    As relevant to general education as to journalism training
Ed Webb

Google - 0 views

Ed Webb

Official Google Blog: Finding more high-quality sites in search - 0 views

    • Ed Webb
       
      I kinda love Google
Ryan Burke

Wolfram|Alpha - 0 views

  •  
    Today's Wolfram|Alpha is the first step in an ambitious, long-term project to make all systematic knowledge immediately computable by anyone. You enter your question or calculation, and Wolfram|Alpha uses its built-in algorithms and growing collection of data to compute the answer
Ed Webb

Social Media is Killing the LMS Star - A Bootleg of Bryan Alexander's Lost Presentation... - 0 views

  • Note that this isn’t just a technological alternate history. It also describes a different set of social and cultural practices.
  • CMSes lumber along like radio, still playing into the air as they continue to gradually shift ever farther away on the margins. In comparison, Web 2.0 is like movies and tv combined, plus printed books and magazines. That’s where the sheer scale, creative ferment, and wife-ranging influence reside. This is the necessary background for discussing how to integrate learning and the digital world.
  • These virtual classes are like musical practice rooms, small chambers where one may try out the instrument in silent isolation. It is not connectivism but disconnectivism.
  • ...11 more annotations...
  • CMSes shift from being merely retrograde to being actively regressive if we consider the broader, subtler changes in the digital teaching landscape. Web 2.0 has rapidly grown an enormous amount of content through what Yochai Benkler calls “peer-based commons production.” One effect of this has been to grow a large area for informal learning, which students (and staff) access without our benign interference. Students (and staff) also contribute to this peering world; more on this later. For now, we can observe that as teachers we grapple with this mechanism of change through many means, but the CMS in its silo’d isolation is not a useful tool.
  • those curious about teaching with social media have easy access to a growing, accessible community of experienced staff by means of those very media. A meta-community of Web 2.0 academic practitioners is now too vast to catalogue. Academics in every discipline blog about their work. Wikis record their efforts and thoughts, as do podcasts. The reverse is true of the CMS, the very architecture of which forbids such peer-to-peer information sharing. For example, the Resource Center for Cyberculture Studies (RCCS) has for many years maintained a descriptive listing of courses about digital culture across the disciplines. During the 1990s that number grew with each semester. But after the explosive growth of CMSes that number dwindled. Not the number of classes taught, but the number of classes which could even be described. According to the RCCS’ founder, David Silver (University of San Francisco), this is due to the isolation of class content in CMS containers.
  • unless we consider the CMS environment to be a sort of corporate intranet simulation, the CMS set of community skills is unusual, rarely applicable to post-graduation examples. In other words, while a CMS might help privacy concerns, it is at best a partial, not sufficient solution, and can even be inappropriate for already online students.
  • That experiential, teachable moment of selecting one’s copyright stance is eliminated by the CMS.
  • Another argument in favor of CMSes over Web 2.0 concerns the latter’s open nature. It is too open, goes the thought, constituting a “Wild West” experience of unfettered information flow and unpleasant forms of access. Campuses should run CMSes to create shielded environments, iPhone-style walled gardens that protect the learning process from the Lovecraftian chaos without.
  • social sifting, information literacy, using the wisdom of crowds, and others. Such strategies are widely discussed, easily accessed, and continually revised and honed.
  • at present, radio CMS is the Clear Channel of online learning.
  • For now, the CMS landsape is a multi-institutional dark Web, an invisible, unsearchable, un-mash-up-able archipelago of hidden learning content.
  • Can the practice of using a CMS prepare either teacher or student to think critically about this new shape for information literacy? Moreover, can we use the traditional CMS to share thoughts and practices about this topic?
  • The internet of things refers to a vastly more challenging concept, the association of digital information with the physical world. It covers such diverse instances as RFID chips attached to books or shipping pallets, connecting a product’s scanned UPC code to a Web-based database, assigning unique digital identifiers to physical locations, and the broader enterprise of augmented reality. It includes problems as varied as building search that covers both the World Wide Web and one’s mobile device, revising copyright to include digital content associated with private locations, and trying to salvage what’s left of privacy. How does this connect with our topic? Consider a recent article by Tim O’Reilly and John Battle, where they argue that the internet of things is actually growing knowledge about itself. The combination of people, networks, and objects is building descriptions about objects, largely in folksonomic form. That is, people are tagging the world, and sharing those tags. It’s worth quoting a passage in full: “It’s also possible to give structure to what appears to be unstructured data by teaching an application how to recognize the connection between the two. For example, You R Here, an iPhone app, neatly combines these two approaches. You use your iPhone camera to take a photo of a map that contains details not found on generic mapping applications such as Google maps – say a trailhead map in a park, or another hiking map. Use the phone’s GPS to set your current location on the map. Walk a distance away, and set a second point. Now your iPhone can track your position on that custom map image as easily as it can on Google maps.” (http://www.web2summit.com/web2009/public/schedule/detail/10194) What world is better placed to connect academia productively with such projects, the open social Web or the CMS?
  • imagine the CMS function of every class much like class email, a necessary feature, but not by any means the broadest technological element. Similarly the e-reserves function is of immense practical value. There may be no better way to share copyrighted academic materials with a class, at this point. These logistical functions could well play on.
1 - 20 of 22 Next ›
Showing 20 items per page