PRONOM is an online technical registry providing impartial and definitive information about file formats, software products and other technical components required to support long-term access of electronic records.
Data obtained from the open survey developed by the LEARN project (http://www.learn-rdm.eu/) as a self-assessment tool to assist institutions discover how ready they are for managing research data. This dataset replaces the first one published at http://doi.org/10.5281/zenodo.61903. The survey is based on the issues posed to institutions by the LERU Roadmap for Research Data published at the end of 2013, and available at: http://www.learn-rdm.eu/material/leru_roadmap_for_research_data The survey has thirteen questions addressing the main elements to be taken into account in developing an institutional strategy for research data management. Each question has three possible answers representing green, yellow or red light. The more ‘green light’ responses recorded, the readier an institution probably is for managing its research data. The survey is available in English at http://learn-rdm.eu/en/rdm-readiness-survey/ and in Spanish at http://learn-rdm.eu/encuesta-rdm/
BioSchemas relies and extends from schema.org and aims to reuse existing standards and reach consensus among a wide number of life sciences organizations and communities.
This document describes RML, a generic mapping language, based on and extending [R2RML]. The RDF Mapping language (RML) is a mapping language defined to express customized mapping rules from heterogeneous data structures and serializations to the RDF [RDF-CONCEPTS] data model. RML is defined as a superset of the W3C-standardized mapping language [R2RML], aiming to extend its applicability and broaden its scope, adding support for data in other structured formats. [R2RML] is the W3C standard to express customized mappings from relational databases to RDF. RML follows exactly the same syntax as R2RML; therefore, RML mappings are themselves RDF graphs. The present document describes the RML language and its concepts through definitions and examples.
The Compliance Assessment Toolkit will support the EOSC PID policy with services to encode, record, and query compliance with the policy. To do so, a wide range of compliance requirements ( TRUST, FAIR, PID Policy, Reproducibility, GDPR, Licences) will be evaluated as use cases for definition of a conceptual model. At the same time, vocabularies, concepts, and designs are intended to be re-usable for other compliance needs: TRUST, FAIR, POSI, CARE, Data Commons.
The Cross-Domain Interoperability Framework (CDIF) is a set of guidelines and practice for using domain-agnostic standards to support the interoperability and reusability of FAIR data, especially across domain and institutional boundaries. It is being developed in response to the need for agreements on the use of standards in FAIR
The CEOS System Engineering Office (SEO) worked with the CEOS Working Group on Information Systems and Services (WGISS) to gather and organize key information on data policies, data access portals and interoperability protocols.
CEOS is currently operating and planning hundreds of Earth observation satellites. The information contained in this portal will improve the efficiency and effectiveness of gaining access to space-based Earth observation data to support many global intiatives with vast societal impact.
Calcyte is (will be) a toolkit for managing metadata for collections of content
via automatically generated spreadsheets and for creating static HTML repositories.
Calcyte targets the Draft DataCrate Packaging format v0.2.
At this stage Calcyte does not Bag content, it jsut creates Working DataCrates.
This document specifies a method of organising file-based data with associated metadata, known as DataCrate in both human and machine readable formats, based on the schema.org linked-data vocabularly, supplemented with terms from the SPAR ontologies and [PCDM] where schema.org does not have coverage. The motivation for this work comes from the research domain.
A DataCrate is a dataset a set of files contained in a single directory. There are two ways of organizing a DataCrate.
For working data or data that does not need to be distributed with checksums, a Working DataCrate is a plain-old directory containing payload data files, with two metadata files at the root; one for humans and one for machines.
For distribution, or archiving; where integrity is important, a Bagged DataCrate is a BagIt bag conforming to the DataCrate BagIt profile with the payload files in the /data directory. A Bagged DataCrate has a clear separation between metadata and payload, and can be integrity-checked using the checksums in the BagIt manifest.
This website is for information related to the CESAER Taskforce on Open Science, and in particular on its sub-group looking at how the Technical Universities in Europe deal with Engineering and Research Data Management. The group is working on two tasks Task 1 - FAIR Data in Engineering (2018-19) Read - Summary of First Findings on…
Status: Recognised & Endorsed The Metadata IG will concern itself with all aspects of metadata for research data. In particular it will attempt to coordinate the efforts of the WGs concerned with metadata to produce a coherent approach to metadata covering metadata modalities of description, restriction, navigation, provenance, preservation and the use of metadata for the purposes discovery, contextualisation, validation, analytical processing, simulation, visualisation and interoperation. It will also liaise with the other WGs especially Data Foundation and Terminology, PIDs, Standardisation of data categories and codes and Data Citation. This IG activity relates to data management policies and plans of research organisations and researchers, and to policies and standards of research funders and of research communities which may or may not be official standards.
The Library of Congress and its digital preservation partners from the federal, library, creative, publishing, technology, and copyright communities are working to develop a national strategy to collect, archive, and preserve digital content.