Skip to main content

Home/ Bucknell Digital Pedagogy & Scholarship/ Group items tagged big

Rss Feed Group items tagged

Leslie Harris

For Big-Data Scientists, 'Janitor Work' Is Key Hurdle to Insights - NYTimes.com - 0 views

  •  
    New York Times article about some of the challenges of "big data" analysis - particularly the data cleanup needed to make useful inferences.
Todd Suomela

Big data: are we making a big mistake? - 0 views

  •  
    Very good description of the problems that big data claims to solve, but may not actually solve.
jatolbert

The Differences between Digital History and Digital Humanities | Dr Stephen Robertson - 0 views

  •  
    A useful article that challenges the "big tent" model of digital humanities and has important implications for digital scholarship more generally.
Todd Suomela

This VR cycle is dead | TechCrunch - 0 views

  • Occasional training tool and/or the future of high end entertainment — just possibly. But, hey, no one likes to feel sick at the movies. Feeling like you want to barf is only a good use of your time and energy if you’ve eaten something that might otherwise kill you. The wider question in the gaming space, which excels at entertainment and escapism, remains how big VR might get when the hardware isn’t so pricey and clunky, and when investing time and energy in building compelling content can start to look like less of a sinkhole prospect for developers who can now be seen pulling their horns in. These same types of content makers may well be having their heads turned by the potential of AR gaming. Which can, as noted above, already point to the existence of essential content generating wild excitement among millions of consumers. Rather like the solar eclipse, AR has been shown turning masses of heads and even gathering huge crowds of like-minded folks together in public. For now there is simply no argument: AR > VR.
  • As TechCrunch’s Lucas Matney noted: “Over the past several months it’s become clear that the war is no longer HTC and Oculus trying to discover who is Betamax and who is VHS, now they’re just trying to ensure that high-end VR doesn’t turn out to be LaserDisc. Though few of the big players are keen to readily admit it, many investors and analysts have been less than thrilled with the pace of headset sales over the past year.” The bald fact that neither HTC nor Facebook/Oculus has released sales figures for their respective VR headsets speaks volumes. (Analyst estimates aren’t generous, suggesting <500k units apiece.) Sony did put out some sales figures for its Playstation VR headset in February which caused some initial excitement — with the company claiming 915,000 units sold since October 2016. But by June that figure had merely drifted past 1M.
jatolbert

The Digital-Humanities Bust - The Chronicle of Higher Education - 0 views

  • To ask about the field is really to ask how or what DH knows, and what it allows us to know. The answer, it turns out, is not much. Let’s begin with the tension between promise and product. Any neophyte to digital-humanities literature notices its extravagant rhetoric of exuberance. The field may be "transforming long-established disciplines like history or literary criticism," according to a Stanford Literary Lab email likely unread or disregarded by a majority in those disciplines. Laura Mandell, director of the Initiative for Digital Humanities, Media, and Culture at Texas A&M University, promises to break "the book format" without explaining why one might want to — even as books, against all predictions, doggedly persist, filling the airplane-hanger-sized warehouses of Amazon.com.
  • A similar shortfall is evident when digital humanists turn to straight literary criticism. "Distant reading," a method of studying novels without reading them, uses computer scanning to search for "units that are much smaller or much larger than the text" (in Franco Moretti’s words) — tropes, at one end, genres or systems, at the other. One of the most intelligent examples of the technique is Richard Jean So and Andrew Piper’s 2016 Atlantic article, "How Has the MFA Changed the American Novel?" (based on their research for articles published in academic journals). The authors set out to quantify "how similar authors were across a range of literary aspects, including diction, style, theme, setting." But they never cite exactly what the computers were asked to quantify. In the real world of novels, after all, style, theme, and character are often achieved relationally — that is, without leaving a trace in words or phrases recognizable as patterns by a program.
  • Perhaps toward that end, So, an assistant professor of English at the University of Chicago, wrote an elaborate article in Critical Inquiry with Hoyt Long (also of Chicago) on the uses of machine learning and "literary pattern recognition" in the study of modernist haiku poetry. Here they actually do specify what they instructed programmers to look for, and what computers actually counted. But the explanation introduces new problems that somehow escape the authors. By their own admission, some of their interpretations derive from what they knew "in advance"; hence the findings do not need the data and, as a result, are somewhat pointless. After 30 pages of highly technical discussion, the payoff is to tell us that haikus have formal features different from other short poems. We already knew that.
  • ...2 more annotations...
  • The outsized promises of big-data mining (which have been a fixture in big-figure grant proposals) seem curiously stuck at the level of confident assertion. In a 2011 New Left Review article, "Network Theory, Plot Analysis," Moretti gives us a promissory note that characterizes a lot of DH writing: "One day, after we add to these skeletons the layers of direction, weight and semantics, those richer images will perhaps make us see different genres — tragedies and comedies; picaresque, gothic, Bildungsroman … — as different shapes; ideally, they may even make visible the micro-patterns out of which these larger network shapes emerge." But what are the semantics of a shape when measured against the tragedy to which it corresponds? If "shape" is only a place-holder meant to allow for more-complex calculations of literary meaning (disburdened of their annoyingly human baggage), by what synesthetic principle do we reconvert it into its original, now reconfigured, genre-form? It is not simply that no answers are provided; it is that DH never asks the questions. And without them, how can Moretti’s "one day" ever arrive?
  • For all its resources, the digital humanities makes a rookie mistake: It confuses more information for more knowledge. DH doesn’t know why it thinks it knows what it does not know. And that is an odd place for a science to be.
karch10k

GPAs don't really show what students learned. Here's why. - 0 views

  •  
    An interesting look into the problem of GPA not being representative to what a student knows, learned, or can demonstrate in any given course. This certainly isn't a new issue and solutions vary - from competency based education (CBE) to utilizing metrics like GPAM. I think one of the big takeaways here is that utilizing any singular metric to measure student achievement is missing the forest for the trees.
Todd Suomela

The Internet as existential threat « Raph's Website - 1 views

  • Our medical systems have terrible Internet security… MRI machines you can connect to with USB that still have “admin:password” to gain root access. That’s horrifying, sure, but that’s not an attack at scale. More frightening: we’re busily uploading all our medical records to the cloud. Take down that cloud, and no patients can be treated, because nobody will know what they have, what meds they are on. Software swallows your insulin pumps and your pacemakers. To kill people, all you need is to hack that database, or simply erase it or block access to it. After all, we don’t tend to realize that in an Internet of Things, humans are just Things too. As this software monster has encroached on stuff like election systems, the common reaction has been to go back to paper. So let’s consider a less obvious example. We should be going back to paper for our libraries too! We’ve outsourced so much of our knowledge to digital that the amount of knowledge available in analog has dropped notably. There are less librarians in the fewer libraries with smaller collections than there used to be. If the net goes down, how much reference material is simply not accessible that was thirty years ago? Google Search is “critical cultural infrastructure.” How much redundancy do we actually have? Could a disconnected town actually educate its children? How critical is Google as a whole? If Google went down for a month, I am pretty sure we would see worldwide economic collapse. How much of the world economy passes through Google hosting? How much of it is in GMail? How much is dependent on Google Search, Google Images, Google Docs? The answer is a LOT. And because financial systems are now also JIT, ten thousand corporate blips where real estate agencies and local car washes and a huge pile of software companies and a gaggle of universities and so on are suddenly 100% unable to function digitally (no payroll! no insurance verification!) would absolutely have ripple effects into their suppliers and their customers, and thence to the worldwide economic market. Because interconnection without redundancy increases odds of cascades.
  • But just as critically, governments and state actors seem to be the source of so many of the problems precisely because the Internet is now too many forms of critical infrastructure, and therefore too juicy a target. If software eats everything, then the ability to kill software is the ability to kill anything. Net connectivity becomes the single point of failure for every system connected to it. Even if the Net itself is designed to route around damage, that doesn’t help if it is the single vector of attack that can take down any given target. It’s too juicy a target for the military, too juicy a target for terror, too juicy a target for criminal ransom. The old adage goes “when they came for this, I said nothing. When they came for that…” — we all know it. Consider that the more we hand gleefully over to the cloud because we want convenience, big data, personalization, and on, we’re creating a single thing that can be taken from us in an instant. We’ve decided to subscribe to everything, instead of owning it. When they came for your MP3s, your DVDs, fine,. not “critical infrastructure.” When they came for your resumes, OK, getting closer.
  • As we rush towards putting more and more things “in the cloud,” as we rush towards an Internet of Things with no governance beyond profit motive and anarchy, what we’re effectively doing is creating a massive single point of failure for every system we put in it.
Todd Suomela

DSHR's Blog: Ithaka's Perspective on Digital Preservation - 0 views

  • Second, there is very little coverage of Web archiving, which is clearly by far the largest and most important digital preservation initiative both for current and future readers. The Internet Archive rates only two mentions, in the middle of a list of activities and in a footnote. This is despite the fact that archive.org is currently the 211th most visited site in the US (272nd globally) with over 5.5M registered users, adding over 500 per day, and serving nearly 4M unique IPs per day. For comparison, the Library of Congress currently ranks 1439th in the US (5441st globally). The Internet Archive's Web collection alone probably dwarfs all other digital preservation efforts combined both in size and in usage. Not to mention their vast collections of software, digitized books, audio, video and TV news.. Rieger writes: There is a lack of understanding about how archived websites are discovered, used, and referenced. “Researchers prefer to cite the original live-web as it is easier and shorter,” pointed out one of the experts. “There is limited awareness of the existence of web archives and lack of community consensus on how to treat them in scholarly work. The problems are not about technology any more, it is about usability, awareness, and scholarly practices.” The interviewee referred to a recent CRL study based on an analysis of referrals to archived content from papers that concluded that the citations were mainly to articles about web archiving projects. It is surprising that the report doesn't point out that the responsibility for educating scholars in the use of resources lies with the "experts and thought leaders" from institutions such as the University of California, Michigan State, Cornell, MIT, NYU and Virginia Tech. That these "experts and thought leaders" don't consider the Internet Archive to be a resource worth mentioning might have something to do with the fact that their scholars don't know that they should be using it. A report whose first major section, entitled "What's Working Well", totally fails to acknowledge the single most important digital preservation effort of the last two decades clearly lacks credibility
  • Finally, there is no acknowledgement that the most serious challenge facing the field is economic. Except for a few corner cases, we know how to do digital preservation, we just don't want to pay enough to have it done. Thus the key challenge is to achieve some mixture of significant increase in funding for, and significant cost reduction in the processes of, digital preservation. Information technology processes naturally have very strong economies of scale, which result in winner-take-all markets (as W. Brian Arthur pointed out in 1985). It is notable that the report doesn't mention the winners we already have, in Web and source code archiving, and in emulation. All are at the point where a competitor is unlikely to be viable. To be affordable, digital preservation needs to be done at scale. The report's orientation is very much "let a thousand flowers bloom", which in IT markets only happens at a very early stage. This is likely the result of talking only to people nurturing a small-scale flower, not to people who have already dominated their market niche. It is certainly a risk that each area will have a single point of failure, but trying to fight against the inherent economics of IT pretty much guarantees ineffectiveness.
  • 1) The big successes in the field haven't come from consensus building around a roadmap, they have come from idiosyncratic individuals such as Brewster Kahle, Roberto di Cosmo and Jason Scott identifying a need and building a system to address it no matter what "the community" thinks. We have a couple of decades of experience showing that "the community" is incapable of coming to a coherent consensus that leads to action on a scale appropriate to the problem. In any case, describing road-mapping as "research" is a stretch. 2) Under severe funding pressure, almost all libraries have de-emphasized their custodial role of building collections in favor of responding to immediate client needs. Rieger writes: As one interviewee stated, library leaders have “shifted their attention from seeing preservation as a moral imperative to catering to the university’s immediate needs.” Regrettably, but inevitably given the economics of IT markets, this provides a market opportunity for outsourcing. Ithaka has exploited one such opportunity with Portico. This bullet does describe "research" in the sense of "market research".  Success is, however, much more likely to come from the success of an individual effort than from a consensus about what should be done among people who can't actually do it. 3) In the current climate, increased funding for libraries and archives simply isn't going to happen. These institutions have shown a marked reluctance to divert their shrinking funds from legacy to digital media. Thus the research topic with the greatest leverage in turning funds into preserved digital content is into increasing the cost-effectiveness of the tools, processes and infrastructure of digital preservation.
Todd Suomela

Beyond buttonology: Digital humanities, digital pedagogy, and the ACRL Framework | Russ... - 0 views

  • Here are a few specific examples you can apply to your instructional design process to help learners with metacognition: Model the metacognitive process during instruction (or in one-on-one consultations) to ask and reflect on big picture questions such as: “What questions can you answer with this tool?” “What can you not do with this tool?” Keep in mind some answers may be simple (e.g., this tool can only work with data in this way, so it is excluded automatically). Also, “Did I get the results I expected? What could I have done differently?” Start with inquiry and build conversations based on the learner’s answers. “Is it the data that does not work? Or is the research question fundamentally wrong to begin with?” Collaborate with faculty to teach together, modelling your practices while demonstrating a specific tool. This could include thinking aloud as you make decisions so learners can self-correct assumptions. Also, be aware of your own expert bias so you can demonstrate how to clear obstacles. Ask learners to specifically define what is difficult for them during the process of instruction. Digital humanities tools are complex and are based on complex methodologies and research questions. By constructing opportunities for learners to self-question as they move from one task to another, they learn to self-assess their progress and adjust accordingly. There are several instructional design activities that promote metacognition: think-pair-share, one minute paper (“share a key concept learned” or “what comes next?”), and case studies.
  • There are specific strategies we can implement to help learners escape the recursive spiral of the liminal state they experience while managing complex digital projects: One of the most challenging aspects of teaching digital tools is forgetting what it is like to be a novice learner. Sometimes being a near-novice oneself helps you better prepare for the basic problems and frustrations learners are facing. But recognizing liminality is a reminder to you as a teacher that the learning process is not smooth, and it requires anticipating common difficulties and regularly checking in with learners to make sure you are not leaving them behind. When meeting with learners one-on-one, make sure to use your in-depth reference interview skills to engage in methods discussions. When a learner is in the liminal state, they are not always able to “see the forest for the trees.” Your directed questions will illuminate the problems they are having and the solutions they had not seen. Pay close attention to the digital humanities work and discussions happening on your own campus, as well as across the academic community. Working through the liminal space may require helping learners make connections to others facing similar problems. Also follow online discussions in order to point your learners to a wide variety of group learning opportunities, such as the active digital humanities community on Slack.9 When designing instructional opportunities, such as workshops and hackathons, pay particular attention to outreach strategies that may bring like-minded learners together, as well as diverse voices. For example, invite the scholar whose project was completed last year to add a more experienced voice to the conversation. By encouraging the formation of learning communities on your campus, you are creating safe spaces to help learners navigate the liminal state with others who may be on the other side of struggling with specific digital project issues. In designing instructional activities, guide learners through visualization exercises that help to identify “stuck” places. Making graphic representations of one’s thoughts (e.g., concept maps) can highlight areas that require clarification.
1 - 11 of 11
Showing 20 items per page