Skip to main content

Home/ OpenSciInfo/ Group items tagged data

Rss Feed Group items tagged

Mike Chelen

Protocol for Implementing Open Access Data - 0 views

  • information for the Internet community
  • distributing data or databases
  • “open” and “open access”
  • ...69 more annotations...
  • requirements for gaining and using the Science Commons Open Access Data Mark and metadata
  • interoperability of scientific data
  • terms and conditions around data make integration difficult to legally perform
  • single license
  • data with this license can be integrated with any other data under this license
  • too many databases under too many terms already
  • unlikely that any one license or suite of licenses will have the correct mix of terms
  • principles for open access data and a protocol for implementing those principles
  • Open Access Data Mark and metadata
  • databases and data
  • the foundation to legally integrate a database or data product
  • another database or data product
  • no mechanisms to manage transfer or negotiations of rights unrelated to integration
  • submitted to Science Commons for certification as a conforming implementation
  • Open Access Data trademarks (icons and phrases) and metadata on databases
  • protocol must promote legal predictability and certainty
  • easy to use and understand
  • lowest possible transaction costs on users
  • Science Commons’ experience in distributing a database licensing Frequently Asked Questions (FAQ) file
  • hard to apply the distinction between what is copyrightable and what is not copyrightable
  • lack of simplicity restricts usage
  • reducing or eliminating the need to make the distinction between copyrightable and non-copyrightable elements
  • satisfy the norms and expectations of the disciplines providing the database
  • norms for citation will differ
  • norms must be attached
  • Converge on the public domain by waiving all rights based on intellectual property
  • reconstruction of the public domain
  • scientific norms to express the wishes of the data provider
  • public domain
  • waiving the relevant rights on data and asserting that the provider makes no claims on the data
  • Requesting behavior, such as citation, through norms rather than as a legal requirement based on copyright or contracts, allows for different scientific disciplines to develop different norms for citation.
  • waive all rights necessary for data extraction and re-use
  • copyright
  • sui generis database rights
  • claims of unfair competition
  • implied contracts
  • and other legal rights
  • any obligations on the user of the data or database such as “copyleft” or “share alike”, or even the legal requirement to provide attribution
  • non-legally binding set of citation norms
  • waiving other statutory or intellectual property rights
  • there are other rights, in addition to copyright, that may apply
  • uncopyrightable databases may be protected in some countries
  • sui generis rights apply in the European Union
  • waivers of sui generis and other legal grounds for database protection
  • no contractual controls
  • using contract, rather than intellectual property or statutory rights, to apply terms to databases
  • affirmatively declare that contractual constraints do not apply to the database
  • interoperation with databases and data not available under the Science Commons Open Access Data Protocol through metadata
  • data that is not or cannot be made available under this protocol
  • owner provides metadata (as data) under this protocol so that the existence of the non-open access data is discoverable
  • digital identifiers and metadata describing non-open access data
  • “Licensing” a database typically means that the “copyrightable elements” of a database are made available under a copyright license
  • Database FAQ, in its first iteration, recommended this method
  • recommendation is now withdrawn
  • copyright begins in and ends in many databases
  • database divided into copyrightable and non copyrightable elements
  • user tends to assume that all is under copyright or none is under copyright
  • share-alike license on the copyrightable elements may be falsely assumed to operate on the factual contents of a database
  • copyright in situations where it is not necessary
  • query across tens of thousands of data records across the web might return a result which itself populates a new database
  • selective waiving of intellectual property rights fail to provide a high degree of legal certainty and ease of use
  • problem of false expectations
  • apply a “copyleft” term to the copyrightable elements of a database, in hopes that those elements result in additional open access database elements coming online
  • uncopyrightable factual content
  • republish those contents without observing the copyleft or share-alike terms
  • cascading attribution if attribution is required as part of a license approach
  • Would a scientist need to attribute 40,000 data depositors in the event of a query across 40,000 data sets?
  • conflict with accepted norms in some disciplines
  • imposes a significant transaction cost
Mike Chelen

Science in the open » A breakthrough on data licensing for public science? - 0 views

  • Peter Murray-Rust and others at the Unilever Centre for Molecular Informatics at Cambridge
  • conversation we had over lunch with Peter, Jim Downing, Nico Adams, Nick Day and Rufus Pollock
  • appropriate way to license published scientific data
  • ...27 more annotations...
  • value of share-alike or copyleft provisions of GPL and similar licenses
  • spreading the message and use of Open Content
  • prevent “freeloaders” from being able to use Open material and not contribute back to the open community
  • presumption in this view is that a license is a good, or at least acceptable, way of achieving both these goals
  • allow people the freedom to address their concerns through copyleft approaches
  • Rufus
  • concerned more centrally with enabling re-use and re-purposing of data as far as is possible
  • make it easy for researchers to deliver on their obligations
  • worried by the potential for licensing to make it harder to re-use and re-mix disparate sets of data and content into new digital objects
  • “license”, will have scientists running screaming in the opposite direction
  • we focused on what we could agree on
  • common position statement
  • area of best practice for the publication of data that arises from public science
  • there is a window of opportunity to influence funder positions
  • data sharing policies
  • “following best practice”
  • don’t tend to be concerned about freeloading
  • providing clear guidance and tools
  • if it is widely accepted by their research communities
  • “best practice is X”
  • enable re-use and re-purposing of that data
  • share-alike approaches as a community expectation
  • Explicit statements of the status of data are required and we need effective technical and legal infrastructure to make this easy for researchers.
  • “Where a decision has been taken to publish data deriving from public science research, best practice to enable the re-use and re-purposing of that data, is to place it explicitly in the public domain via {one of a small set of protocols e.g. cc0 or PDDL}.”
  • focuses purely on what should be done once a decision to publish has been made
  • data generated by public science
  • describing this as best practice it also allows deviations that may, for whatever reason, be justified by specific people in specific circumstances
Mike Chelen

BioMart - 0 views

shared by Mike Chelen on 11 Dec 08 - Cached
  •  
    BioMart is a query-oriented data management system developed jointly by the Ontario Institute for Cancer Research (OiCR) and the European Bioinformatics Institute (EBI). The system can be used with any type of data and is particularly suited for providing 'data mining' like searches of complex descriptive data. BioMart comes with an 'out of the box' website that can be installed, configured and customised according to user requirements. Further access is provided by graphical and text based applications or programmatically using web services or API written in Perl and Java. BioMart has built-in support for query optimisation and data federation and in addition can be configured to work as a DAS 1.5 Annotation server. The process of converting a data source into BioMart format is fully automated by the tools included in the package. Currently supported RDBMS platforms are MySQL, Oracle and Postgres. BioMart is completely Open Source, licensed under the LGPL, and freely available to anyone without restrictions.
Mike Chelen

Open Knowledge Foundation Blog » Blog Archive » Open Data: Openness and Licen... - 0 views

  • Why bother about openness and licensing for data
  • It’s crucial because open data is so much easier to break-up and recombine, to use and reuse.
  • want people to have incentives to make their data open and for open data to be easily usable and reusable
  • ...8 more annotations...
  • good definition of openness acts as a standard that ensures different open datasets are ‘interoperable’
  • Licensing is important because it reduces uncertainty. Without a license you don’t know where you, as a user, stand: when are you allowed to use this data? Are you allowed to give to others? To distribute your own changes, etc?
  • licensing and definitions are important even though they are only a small part of the overall picture
  • If we get them wrong they will keep on getting in the way of everything else.
  • Everyone agrees that requiring attribution is OK
    • Mike Chelen
       
      My opinion is that there should be no requirements, including attribution, and that standards should be community-based instead of legal.
  • Even if a basic license is used it can be argued that any ‘requirements’ for attribution or share-alike should not be in a license but in ‘community norms’.
    • Mike Chelen
       
      Licenses and community norms are not exclusive. It's recommended to adopt a Public Domain license, and encourage attribution through community standards.
  • A license is likely to elicit at least as much, and almost certainly more, conformity with its provisions than community norms.
    • Mike Chelen
       
      Ease of access and should be the goal, not conformity.
  • (even to a user it is easy to comply with the open license)
    • Mike Chelen
       
      It is important to specifically publish using a Public Domain dedication.
  •  
    Why bother about openness and licensing for data? After all they don't matter in themselves: what we really care about are things like the progress of human knowledge or the freedom to understand and share.
Mike Chelen

SWAN (Semantic Web Applications in Neuromedicine) Project - 0 views

  •  
    SWAN (Semantic Web Applications in Neuromedicine) is a Web-based collaborative program that aims to organize and annotate scientific knowledge about Alzheimer disease (AD) and other neurodegenerative disorders. Its goal is to facilitate the formation, development and testing of hypotheses about the disease. The ultimate goal of this project is to create tools and resources to manage the evolving universe of data and information about AD in such a way that researchers can easily comprehend their larger context ("what hypothesis does this support or contradict?"), compare and contrast hypotheses ("where do these two hypotheses agree and disagree?"), identify unanswered questions and synthesize concepts and data into ever more comprehensive and useful hypotheses and treatment targets for this disease. The SWAN project is designed to allow the community of AD researchers to author, curate and connect a diversity of data and ideas about AD via secure personal and public SWAN workspaces, using the emerging Semantic Web paradigm for deep interconnection of data, information and knowledge. We are initially focusing on developing a fully public Web resource deployed as part of the Alzheimer Research Forum web site (www.alzforum.org). After the public resource has been launched, we will also develop secure personal workspaces (MySWAN) and semi-private lab workspaces (LabSWAN). An essential component of this project is development of an initial, core knowledge base within SWAN, which will provide immediate value to researchers at the time of deployment. This is a critically important part of our strategy to ensure that the SWAN system gains wide adoption and active participation by the AD research community. As part of our development strategy, we are also recruiting a "beta test" community of AD researchers to enter their own hypotheses, add commentaries and citations, and provide feedback on the technology and content. SWAN is being developed by a collaborative team from
Mike Chelen

FTP Download - 0 views

  •  
    If required, entire databases can be downloaded from our FTP site in a variety of formats, from flat files to MySQL dumps. Please be aware that these files can run to many gigabytes of data. To facilitate storage and download all databases are GNU Zip (gzip, *.gz) compressed. Please note: Ensembl supports downloading of many correlation tables via the highly customisable BioMart data mining tool. You may find exploring this web-based data mining tool easier than extracting information from our database dumps.
Mike Chelen

Open Knowledge Foundation Blog » Blog Archive » Comments on the Science Commo... - 0 views

  • the protocol does not discuss any of the possible attractions of allowing such provisions
  • Protocol gives 3 basic reasons for preferring the ‘PD’ approach
  • Science Commons Protocol for Implementing Open Access Data
  • ...7 more annotations...
  • I am not really convinced by any of these points that attribution or share-alike provisions should not be included in open data licenses
  • application of obligations based on copyright in situations where it is not necessary
  • non-copyrightable elements extends to the entire database and inadvertently infringe
  • If intellectual property rights are involved
  • requirements carrying a stiff penalty for failure
  • selective waiving of intellectual property rights
  • interpretative problems
Mike Chelen

The National Center for Biomedical Ontology - 0 views

  •  
    The National Center for Biomedical Ontology is a consortium of leading biologists, clinicians, informaticians, and ontologists who develop innovative technology and methods allowing scientists to create, disseminate, and manage biomedical information and knowledge in machine-processable form. Our visionis that all biomedical knowledge and data are disseminated on the Internet using principled ontologies, such that they are semantically interoperable and useful for improving biomedical science and clinical care. Our resources include the Open Biomedical Ontologies (OBO) library, the Open Biomedical Data (OBD) repositories, and tools for accessing and using this information in research. The Center collaborates with biomedical researchers conducting Driving Biological Projects to enable outside research and stimulate technology development in the Center. The Center undertakes outreach and educational activities (Biomedical Informatics Program) to train future researchers to use biomedical ontologies and related tools with the goal of enhancing scientific discovery.
Mike Chelen

SourceForge.net: CloudBurst - cloudburst-bio - 0 views

  •  
    CloudBurst: Highly Sensitive Short Read Mapping with MapReduce Michael Schatz Center for Bioinformatics and Computational Biology, University of Maryland Next-generation DNA sequencing machines are generating an enormous amount of sequence data, placing unprecedented demands on traditional single-processor read mapping algorithms. CloudBurst is a new parallel read-mapping algorithm optimized for mapping next-generation sequence data to the human genome and other reference genomes, for use in a variety of biological analyses including SNP discovery, genotyping, and personal genomics. It is modeled after the short read mapping program RMAP, and reports either all alignments or the unambiguous best alignment for each read with any number of mismatches or differences. This level of sensitivity could be prohibitively time consuming, but CloudBurst uses the open-source Hadoop implementation of MapReduce to parallelize execution using multiple compute nodes. CloudBurst's running time scales linearly with the number of reads mapped, and with near linear speedup as the number of processors increases. In a 24-processor core configuration, CloudBurst is up to 30 times faster than RMAP executing on a single core, while computing an identical set of alignments. In a large remote compute clouds with 96 cores, CloudBurst reduces the running time from hours to mere minutes for typical jobs involving mapping of millions of short reads to the human genome. CloudBurst is available open-source as a model for parallelizing other bioinformatics algorithms with MapReduce.
Mike Chelen

SourceForge.net: Running CloudBurst on Amazon EC2 - cloudburst-bio - 0 views

  •  
    Hadoop comes bundled with launch scripts to simplify initializing an Amazon Elastic Compute Cloud (EC2) cloud for Hadoop. Once initialized, running CloudBurst is identical to running on a local cluster. If you use EC2 regularly with the same datasets (i.e. the human genome as a reference), you will probably want to copy the data once to Amazon Simple Storage Service (S3) so you can quickly copy the data from S3 to your cloud at low cost.
Mike Chelen

NIF - 0 views

  •  
    The Neuroscience Information Framework (NIF) is a dynamic inventory of web-based neurosciences data, resources, and tools that scientists and students can access via any computer connected to the Internet. An initiative of the NIH Blueprint for Neuroscience Research, the NIF will advance neuroscience research by enabling discovery and access to public research data and tools worldwide through an open source, networked environment.
Mike Chelen

A pitfall of wiki solution for biological database...[Brief Bioinform. 2008] - PubMed R... - 0 views

  •  
    Not a few biologists tend to consider wiki as a solution to manage and reorganize data by a community. However, in its basic functionality, wiki lacks a measure to check data consistency and is not suitable for a database. To circumvent this pitfall, installation of page dependency through in-line page searches is necessary. We also introduce two existing approaches that support in-line queries.
Mike Chelen

Public MySQL Server - 0 views

  • For large amounts of data and more detailed analysis, we recommend you use our publicly-accessible MySQL server, ensembldb.ensembl.org, which you can access as user 'anonymous'. A second server, martdb.ensembl.org provides public access to the BioMart databases.
  •  
    For large amounts of data and more detailed analysis, we recommend you use our publicly-accessible MySQL server, ensembldb.ensembl.org, which you can access as user 'anonymous'. A second server, martdb.ensembl.org provides public access to the BioMart databases.
Mike Chelen

Main Page - GenBioWiki - 0 views

  •  
    GenBioWiki is the student home page for the Genetics, Bioinformatics, and Computational Biology (GBCB) program at Virginia Tech. Bioinformatics and computational biology provide a research platform to acquire, manage, analyze, and display large amounts of data, which in turn catalyze a systems approach to understanding biological organisms, as well as making useful predictions about their behavior in response to environmental and other perturbations. Moreover, bioinformatics is the study of biological systems and large biological data sets using analytical methods borrowed from computer science, mathematics, statistics, and the physical sciences. This transdisciplinary approach to research requires graduates with extensive cross-cultural professional and technical training and provides ample employment opportunities for Ph.D. graduates. [1]
Mike Chelen

BioLit Project - 0 views

  •  
    The establishment of open access literature makes it possible for knowledge to be extracted from scholarly articles and included in other resources. BioLit aims to extract database identifiers and rich meta-data from open access articles in the life sciences and integrate that information with existing biological databases. We have begun prototyping this effort using a clone of the RCSB Protein Data Bank, a database of macromolecular structures.
Mike Chelen

YouTube - Hans Rosling: No more boring data: TEDTalks - 0 views

  •  
    With the drama and urgency of a sportscaster, statistics guru Hans Rosling uses an amazing new presentation tool, Gapminder, to debunk several myths about world development. Rosling is professor of international health at Sweden's Karolinska Institute, and founder of Gapminder, a nonprofit that brings vital global data to life. (Recorded February 2006 in Monterey, CA.)
Mike Chelen

de.bezier.mysql - 0 views

  •  
    Processing (BETA) library to communicate with MySQL (or any other SQL) databases. note that due to java security restrictions this will not work with applets "out of the box" and that many remote mysql-servers will only allow local access ("localhost") or connections from trusted hosts. (see notes. ) also note that you should have some experience with SQL to put, change and retrieve data from the database.
1 - 20 of 54 Next › Last »
Showing 20 items per page