Skip to main content

Home/ Data Working Group/ Group items tagged social

Rss Feed Group items tagged

Amy West

PLoS Computational Biology: Defrosting the Digital Library: Bibliographic Tools for the... - 0 views

  • Presently, the number of abstracts considerably exceeds the number of full-text papers,
  • full papers that are available electronically are likely to be much more widely read and cited
  • Since all of these libraries are available on the Web, increasing numbers of tools for managing digital libraries are also Web-based. They rely on Uniform Resource Identifiers (URIs [25] or “links”) to identify, name, and locate resources such as publications and their authors.
  • ...27 more annotations...
  • We often take URIs for granted, but these humble strings are fundamental to the way the Web works [58] and how libraries can exploit it, so they are a crucial part of the cyberinfrastructure [59] required for e-science on the Web.
  • link to data (the full-text of a given article),
  • To begin with, a user selects a paper, which will have come proximately from one of four sources: 1) searching some digital library, “SEARCH” in Figure 4; 2) browsing some digital library (“BROWSE”); 3) a personal recommendation, word-of-mouth from colleague, etc., (“RECOMMEND”); 4) referred to by reading another paper, and thus cited in its reference list (“READ”)
  • There is no universal method to retrieve a given paper, because there is no single way of identifying publications across all digital libraries on the Web
  • Publication metadata often gets “divorced” from the data it is about, and this forces users to manage each independently, a cumbersome and error-prone process.
  • There is no single way of representing metadata, and without adherence to common standards (which largely already exist, but in a plurality) there never will be.
  • Where DOIs exist, they are supposed to be the definitive URI. This kind of automated disambiguation, of publications and authors, is a common requirement for building better digital libraries
  • Publication metadata are essential for machines and humans in many tasks, not just the disambiguation described above. Despite their importance, metadata can be frustratingly difficult to obtain.
  • So, given an arbitrary URI, there are only two guaranteed options for getting any metadata associated with it. Using http [135], it is possible to for a human (or machine) to do the following.
  • This technique works, but is not particularly robust or scalable because every time the style of a particular Web site changes, the screen-scraper will probably break as well
  • This returns metadata only, not the whole resource. These metadata will not include the author, journal, title, date, etc., of
  • As it stands, it is not possible to perform mundane and seemingly simple tasks such as, “get me all publications that fulfill some criteria and for which I have licensed access as PDF” to save locally, or “get me a specific publication and all those it immediately references”.
  • Having all these different metadata standards would not be a problem if they could easily be converted to and from each other, a process known as “round-tripping”.
  • many of these mappings are non-trivial, e.g., XML to RDF and back again
  • more complex metadata such as the inbound and outbound citations, related articles, and “supplementary” information.
  • Personalization allows users to say this is my library, the sources I am interested in, my collection of references, as well as literature I have authored or co-authored. Socialization allows users to share their personal collections and see who else is reading the same publications, including added information such as related papers with the same keyword (or “tag”) and what notes other people have written about a given publication.
  • CiteULike normalizes bookmarks before adding them to its database, which means it calculates whether each URI bookmarked identifies an identical publication added by another user, with an equivalent URI. This is important for social tagging applications, because part of their value is the ability to see how many people (and who) have bookmarked a given publication. CiteULike also captures another important bibliometric, viz how many users have potentially read a publication, not just cited it.
  • Connotea uses MD5 hashes [157] to store URIs that users bookmark, and normalizes them after adding them to its database, rather than before.
  • he source code for Connotea [159] is available, and there is an API that allows software engineers to build extra functionality around Connnotea, for example the Entity Describer [160].
  • Personalization and socialization of information will increasingly blur the distinction between databases and journals [175], and this is especially true in computational biology where contributions are particularly of a digital nature.
  • This is usually because they are either too “small” or too “big” to fit into journals.
  • As we move in biology from a focus on hypothesis-driven to data-driven science [1],[181],[182], it is increasingly recognized that databases, software models, and instrumentation are the scientific output, rather than the conventional and more discursive descriptions of experiments and their results.
  • In the digital library, these size differences are becoming increasingly meaningless as data, information, and knowledge become more integrated, socialized, personalized, and accessible. Take Postgenomic [183], for example, which aggregates scientific blog posts from a wide variety of sources. These posts can contain commentary on peer-reviewed literature and links into primary database sources. Ultimately, this means that the boundaries between the different types of information and knowledge are continually blurring, and future tools seem likely to continue this trend.
  • he identity of people is a twofold problem because applications need to identify people as users in a system and as authors of publications.
  • Passing valuable data and metadata onto a third party requires that users trust the organization providing the service. For large publishers such as Nature Publishing Group, responsible for Connotea, this is not necessarily a problem.
  • business models may unilaterally change their data model, making the tools for accessing their data backwards incompatible, a common occurrence in bioinformatics.
  • Although the practice of sharing raw data immediately, as with Open Notebook Science [190], is gaining ground, many users are understandably cautious about sharing information online before peer-reviewed publication.
  •  
    Yes, but Alexandria was also a lot smaller; not totally persuaded by analogy here...
David Govoni

Science & Social Media | Tamara Zemlo, BioInformatics, LLC | SciVee - 1 views

  •  
    "On Jan. 6, 2009, in Arlington, Virginia, the National Science Foundation, The Ballston Science and Technology Alliance, and BioInformatics, LLC, hosted a Cafe Scientifique on Science and Social Media. In part 1 of this 4 part video, Dr. Tamara Zemlo from
David Govoni

Science & Social Media | Chris Condayan, ASM/MicrobeWorld | SciVee - 0 views

  •  
    "On Jan. 6, 2009, in Arlington, Virginia, the National Science Foundation, The Ballston Science and Technology Alliance, and BioInformatics, LLC, hosted a Cafe Scientifique on Science and Social Media. In part 2 of this 4 part video, Chris Condayan, Manag
David Govoni

Science & Social Media | Stephanie Stockman, NASA | SciVee - 0 views

  •  
    "On Jan. 6, 2009, in Arlington, Virginia, the National Science Foundation, The Ballston Science and Technology Alliance, and BioInformatics, LLC, hosted a Cafe Scientifique on Science and Social Media. In part 3 of this 4 part video, Stephanie Stockman, a
David Govoni

Science & Social Media | Nancy Shute, US News & World Report | SciVee - 0 views

  •  
    "On Jan. 6, 2009, in Arlington, Virginia, the National Science Foundation, The Ballston Science and Technology Alliance, and BioInformatics, LLC, hosted a Cafe Scientifique on Science and Social Media. In the final segment of this 4 part video, Nancy Shut
David Govoni

SciTechNet(sm): Science and Technology Social Networking Services - 0 views

  •  
    SciTechNet(sm) is devoted to describing and documenting online social networking services in the Sciences and Technology.
umgeoglib

DVN - Homepage - 0 views

  •  
    List of Social Sciences data available via IQSS Dataverse Network
Amy West

The Enduring Value of Social Science Research: The Use and Reuse of Primary Research Da... - 2 views

  •  
    Paper on data sharing from social sciences perspective; also some analysis of sharing so far.
David Govoni

Scratchpads | Biodiversity Online - 0 views

  •  
    "Scratchpads are an easy to use, social networking application that enable communities of researchers to manage, share and publish taxonomic data online. Sites are hosted at the Natural History Museum London, and offered free to any scientist that complet
Amy West

Developing the Capability and Skills to Support eResearch - 0 views

  • Of particular concern to this article is the need for improved levels of data stewardship to enable good data management for long-term sustainability, both at national and institutional levels.
  • researchers, particularly those engaged in data-intensive research; systems developers, data scientists and other technical staff; data managers of institutional repositories, data archives and discipline-based data centres and their support staff, with those who liaise between depositors and the repository as being of particular interest; and those who are engaged in high-level policy formulation, either in government or research institutions.
  • Interviews were conducted with twelve key established researchers in six Australian institutions, with a focus on academics engaged in data-intensive research. Interviews were conducted also with the manager of a large data centre, and a repository administrator. The institutions concerned were the Australian National University, the University of Melbourne, the University of Tasmania, the University of Queensland, the University of Sydney and one area of the Commonwealth Scientific and Industrial Research Organisation (CSIRO).
  • ...18 more annotations...
  • There was wide agreement that there are three types of skills required for practitioners of eResearch, their support staff and repository staff. Not surprisingly, there was a strong need for technical skills. Perhaps not as obvious was the identification of a wide range of non-technical skills. Less obvious again was mention of an assortment of personal qualities, which, while not skills in the formal sense of the term, were singled out as being important.
  • The surveys indicated that not everyone needs the same level of technical skills to conduct or support eResearch.
  • So you need a basic literacy level to look after your computers where you’re storing your data, and then in order to access, like a remote repository, you need to know something about how to connect to that remote repository, what the format of the data should be to go in it, how to convert your data to that required format
  • included skills related to high-performance computing (HPC) and the access grid, data (and database) management, data curation, information engineering, information modelling, software development, remote communications, distributed processing, informatics, portal design, computational fluid dynamics, database integration, visualisation and programming of all kinds.
  • Some of these skills are tightly connected to specific disciplines, especially informatics.
  • The need for technical skills is allied to the ability to understand end-to-end workflows, especially for repository managers and developers who need to be able to think like the researcher and to apply that understanding to developing the repository. By workflows, I mean the many software applications, processing operations and interactions required for research tasks to be carried through to completion.
  • The group of librarians at ‘The Researcher Librarian Nexus’ workshop identified a need for further development of their technical skills, mentioning in particular metadata, something which did not feature among any of the other responses, other than by implication.
  • These vary from skills in data analysis (including the use of statistical packages and other techniques such as data mining) through information seeking to a broader range of general skills. Project management, business analysis, communications, negotiation, intellectual property, team building and train the trainer were mentioned specifically. Another was generic problem solving, because, as one researcher aptly put it, the kinds of problems which arise when undertaking eResearch mean that ‘There’s never going to be someone who has done it before.’
  • The librarians involved with the Researcher/Librarian Nexus workshop also identified it as being of high priority for repository managers, along with marketing, advocacy, copyright, metadata, educational outreach and grant submission writing. They also singled out the intriguing skill of ‘researcher management’ while not specifying precisely what this might entail.
  • A good grasp of copyright and intellectual property issues was seen as essential,
  • These were listed as: open-mindedness, patience and an ‘ability to cooperate and collaborate rather than compete’
  • For example, one researcher, in the field of finance told me of his need for programmers who have a high level of expertise in economics, econometrics, statistics, maths and programming; ‘otherwise all the programming expertise doesn’t really help because then they make strange assumptions in their coding that just result in nonsense output.’
  • One solution to the need to bridge the disciplinary gap is to use graduate students to help with the technical aspects, where those students have an interest and aptitude for this kind of work. In some cases this might be done by providing scholarships, the students then graduating with a PhD on the basis their contribution to the research project has been of sufficient originality to warrant the degree.
  • The barrier to research most often mentioned was the difficulty in assembling all the skills required to conduct a project, particularly in relation to data management and stewardship. In some cases the gap is organisational, as happens for example when the researcher is either unaware of or unable to tap into the skills of a central IT unit. More often the gap was in a lack of understanding of what each group needs, what each has to offer and where responsibilities lie. Examples of this can be seen in comments like the following:
  • For instance if you’ve got data in say NetCDF file formats and the repository wants it in TIFF format, well you need to know something about the technicality of getting your data from NetCDF format into TIFF format.
  • The humanities and social sciences are notable areas where the take-up rate of eResearch has been slower than, for example, in the hard sciences, and where there have been calls for exemplars to be publicised. Many practitioners in the humanities and social sciences find it difficult to envisage where their work might fit into the concept of eResearch.
  • Few researchers are aware that there are such things as repositories, so it is important that the repository is seen as (and indeed is) ‘a good repository – that it’s good in the sense of its high quality but also good in that it adds value for [the researcher].’
  • If research institutions are to minimise the gap between the ideals and realities of eResearch, there is some way to go in providing both institutional capacity and appropriately qualified individuals. While eResearch is dependent on good ICT infrastructure, this is not sufficient in itself. The results of the survey outlined here show that capacity in information technology skills is important but must be accompanied by a range of non-technical skills in such areas as project management. Equally important is the creation of research environments which are covered by well-propagated and understood policies, which are appropriately organised into structures with clearly delineated roles and responsibilities and which minimise the current barriers experienced by many researchers.
Lisa Johnston

Chronopolis -- Digital Preservation Program -- Long-Term Mass-Scale Federated Digital P... - 0 views

  •  
    The Chronopolis Digital Preservation Demonstration Project, one of the Library of Congress' latest efforts to collect and preserve at-risk digital information, has been officially launched as a multi-member partnership to meet the archival needs of a wide range of cultural and social domains. Chronopolis is a digital preservation data grid framework being developed by the San Diego Supercomputer Center (SDSC) at UC San Diego , the UC San Diego Libraries (UCSDL) , and their partners at the National Center for Atmospheric Research (NCAR) in Colorado and the University of Maryland's Institute for Advanced Computer Studies (UMIACS) . A key goal of the Chronopolis project is to provide cross-domain collection sharing for long-term preservation. Using existing high-speed educational and research networks and mass-scale storage infrastructure investments, the partnership is designed to leverage the data storage capabilities at SDSC, NCAR, and UMIACS to provide a preservation data grid that emphasizes heterogeneous and highly redundant data storage systems.
Amy West

DataTrain project - Social Anthropology - 2 views

  •  
    We could definitely use some of this for our own outreach and teaching (the PPTs, reading lists & online resources lists)! To adapt it, just remove the UK and Social Anthro parts.
Amy West

ingentaconnect Citing data sources in the social sciences: do authors do it? - 1 views

Amy West

Liveblog: BRDI: Author Deposit Mandates for Federal Research Grantees : Gavin Baker - 0 views

  • DC Principles Coalition: We believe in free access to science, within the constraints of our business models.
  • The public doesn’t need access to the full articles
  • The problem is that consumers want everything for free.
  • ...6 more annotations...
  • Repositories can do all the functions of journals except quality control, and we don’t want government doing that.
  • Social sciences often left out of discussions about data curation, open access, etc.
  • We could argue that taxpayers paid for the research in general, not necessarily each publication.
  • But the Public Access Policy requires the peer-reviewed manuscript, not the one after which the publishers add value. The America COMPETES model, for un-peer-reviewed grant proposals, is almost useless to the public. In health, you want the refereed results, not the grantee’s report to the agency.
  • If journals can’t survive, from an economic perspective, that’s not harm — it’s just a failure to adapt.
  • Journal growth trends with funding for researchers. As universities want to be more prestigious, they aim to publish more. Trying to have access to everything requires too much money — you have to prioritize.
Amy West

In case you can't read…. | Prof-Like Substance - 1 views

  • When I am putting a talk together it would never occur to me not to include a health dose of unpublished data. The only times in my career that I have talked about mostly published data have been when I first started as a postdoc and in the early days of being a PI, when I didn't have enough new data to even make a coherent story, but that accounts for maybe three professional talks out of man
  • s it a fear of being scooped or a penchant for keeping one's ideas close to the chest that promotes the Summary Talk?
  • I think it's field dependent. Personally, I can rarely get enough information from a talk to know whether to believe a result or not. This means that unpublished data usually ends up with me thinking "maybe, maybe not".
  • ...10 more annotations...
  • (A good talk like this has enough of a citation on the slide that I can jot down where to go if I want to know details on any particular result.)
  • I'm in a highly competitive biomed field, and I was taught never to present something unless it was either submitted or ready to be submitted.
  • I don't really spend any time worrying about being scooped because I collect my own data.
  • Why look at a poster or talk of 100% published work, I've already seen the stuff in a journal to start with
  • Final year materials chemist = keeping cards close to my chest. Once bitten, never again.
  • In neuro, I'd say that at smaller conferences and less high-profile talks at big conferences (i.e. not keynotes or featured lectures), the bulk of what you're hearing is unpublished. ALL posters are unpublished--in fact, I think (?) it's a rule at SfN that the content of posters can't be published already.
  • In my field I'd guess that most talks include data that is in press or at some close to publication sta
  • A big name should be more generous, but then again they do have to save guard the career of the student/postdoc who generated the data. Also the star or keynote speaker is expected to address a wider audience, and make their talk relevant to the overall theme of the conference.
  • In my (experimental) social science, most conferences explicitly say that you cannot submit to present already published or even accepted work.
  • In my field (Astronomy), I'd say 95% of the talks are about unpublished data.
  •  
    A blog post & comments on what's preferred in conference presentations: published or unpublished data. Interesting.
Amy West

WHAT EXPLAINS THE GERMAN LABOR MARKET MIRACLE IN THE GREAT RECESSION? - 0 views

  •  
    This paper uses, among other sources, the US Bureau of Labor Statistics CPS data that covers 1960-2009 to analyze just 2 years of data. The authors do cite the whole CPS, but you have to read the paper to see which bits of that set matter to this paper. The bulk of the paper itself is their explanation of the various statistical methods they used to support their conclusions. The data is neither novel or unique to them. Their analysis however, may be novel and is certainly unique to them. They also provide some technical documentation, e.g. we did x with SPSS. So, ideally, it would be nice to have a citation to the paper, to the 2 year subset of data relevant to it and a citation to the entire BLS CPS data. This is not agricultural economics, but I think that pretty similar patterns will be found there too.
1 - 17 of 17
Showing 20 items per page