Open linked data: Difference between revisions

From Opasnet
Jump to navigation Jump to search
(scope for open data policy)
 
(5 intermediate revisions by the same user not shown)
Line 1: Line 1:
[[op_fi:Avoin data]]
[[Category:Data]]
[[Category:Data]]
[[Category:Openness]]
[[Category:Openness]]
Line 4: Line 5:
'''Open linked data''' is an idea of a new way of organising data. Each piece of data has a [[:en:URI|URI]] which is used as an identifier. Pieces of data can be linked semantically together.
'''Open linked data''' is an idea of a new way of organising data. Each piece of data has a [[:en:URI|URI]] which is used as an identifier. Pieces of data can be linked semantically together.


==Scope==
==Question==


There is a need for a open data policy. What should it be like?
There is a need for a open data policy. What should it be like?
Line 14: Line 15:
* It must be based on [[open assessment]].
* It must be based on [[open assessment]].
* It must save public money.
* It must save public money.
==Answer==
Today, a typical experiment can simultaneously measure hundreds to thousands of individual features (e.g. genes) in dozens of biological conditions, resulting in gigabytes of data that need to be processed, analyzed and potentially reproduced.
Multiple ongoing open-source, community-based projects have emerged that enable researchers to share study protocols, experiment constructs, resulting data sets, as well as complete and fully reproducible data analysis pipelines.
Experimental and computational groups need to work together when developing novel assays, standards and analysis tools ensuring that all steps leading to the results of a study are optimized and reproducible.
The availability of open-acess, high-quality and reproducible data, will also lead to more powerful analyses (or meta-analyses) where multiple data sets are combined to generate new knowledge.
Funding agencies, publishers and researchers need to set strict C&R policies that would allow rapid revealation and correction of scientific errors instead of giving birth to new scientific projects and clinical trials based on erroneous results.
==Rationale==
Conclusions come from Huang and Gottardo <ref>Yunda Huang and Raphael Gottardo. Comparability and reproducibility of biomedical data. Brief Bioinform (2012) {{doi|10.1093/bib/bbs078}} First published online: November 27, 2012</ref>


==See also==
==See also==


* [[Open science]]
* [[The Ball (an information ecosystem)]]
* http://www.data.gov
* http://www.data.gov
* http://data.gov.uk
* http://data.gov.uk
Line 24: Line 43:
* http://test.opengov.fi/
* http://test.opengov.fi/
* [http://www.crazytown.fi/index.php/yritykset/ctjklyritykset/27-hila-open-oy HILA Open Oy]
* [http://www.crazytown.fi/index.php/yritykset/ctjklyritykset/27-hila-open-oy HILA Open Oy]
* [http://www.slideshare.net/apoikola/julkinen-data-oppaan-julkistamistilaisuus Julkinen data -opas: julkistamistilaisuus]
* [http://www.sitra.fi/fi/Ajankohtaista/Puheenvuorot/puheenvuoro_poikola_20100416.htm Antti Poikola: Julkishallinnon avoin data]
* [http://blogs.openaccesscentral.com/blogs/bmcblog/entry/labarchives_and_biomed_central_a LabArchives and BioMed Central: a new platform for publishing scientific data]


==References==
==References==

Latest revision as of 08:53, 9 December 2013


Open linked data is an idea of a new way of organising data. Each piece of data has a URI which is used as an identifier. Pieces of data can be linked semantically together.

Question

There is a need for a open data policy. What should it be like?

  • It must improve international competitiveness of the countries that apply it.
  • It must be global.
  • It must reward for sharing information.
  • It must be tempting to participate, not something that replaces something important you already have but should give up.
  • It must be easy to understand.
  • It must be based on open assessment.
  • It must save public money.

Answer

Today, a typical experiment can simultaneously measure hundreds to thousands of individual features (e.g. genes) in dozens of biological conditions, resulting in gigabytes of data that need to be processed, analyzed and potentially reproduced.

Multiple ongoing open-source, community-based projects have emerged that enable researchers to share study protocols, experiment constructs, resulting data sets, as well as complete and fully reproducible data analysis pipelines.

Experimental and computational groups need to work together when developing novel assays, standards and analysis tools ensuring that all steps leading to the results of a study are optimized and reproducible.

The availability of open-acess, high-quality and reproducible data, will also lead to more powerful analyses (or meta-analyses) where multiple data sets are combined to generate new knowledge.

Funding agencies, publishers and researchers need to set strict C&R policies that would allow rapid revealation and correction of scientific errors instead of giving birth to new scientific projects and clinical trials based on erroneous results.

Rationale

Conclusions come from Huang and Gottardo [1]

See also

References

  1. Yunda Huang and Raphael Gottardo. Comparability and reproducibility of biomedical data. Brief Bioinform (2012) doi:10.1093/bib/bbs078 First published online: November 27, 2012

Related files

<mfanonymousfilelist></mfanonymousfilelist>