Raman spectroscopy.pptx M Pharm, M Sc, Advanced Spectral Analysis
Ten Habits of Highly Effective Data
1. Ten Habits of Highly Effective Data
Anita de Waard
VP Research Data Collaborations
a.dewaard@elsevier.com
http://researchdata.elsevier.com/
2. The Maslow Hierarchy for humans:
9. Usable (allow tools to run on it)
8. Citable (able to point & track citations)
7. Trusted (validated/checked by reviewers)
6. Reproducible (others can redo
experiments)
5. Discoverable (can be indexed by a system)
4. Comprehensible (others can understand
data & processes)
3. Accessible (can be accessed by others)
2. Archived (long-term & format-independent)
1. Preserved (existing in some form)
3. A Maslow Hierarchy for Data:
9. Usable (allow tools to run on it)
8. Citable (able to point & track citations)
7. Trusted (validated/checked by reviewers)
6. Reproducible (others can redo
experiments)
5. Discoverable (can be indexed by a system)
4. Comprehensible (others can understand
data & processes)
3. Accessible (can be accessed by others)
2. Archived (long-term & format-independent)
1. Preserved (existing in some form)
4. 1. Preserve: Data Rescue Challenge
• With IEDA/Lamont: award succesful data
rescue attempts
• Awarded at AGU 2013
• 23 submissions of data that was digitized,
preserved, made available
• Winner: NIMBUS Data Rescue:
– Recovery, reprocessing and digitization of the
infrared and visible observations along with their
navigation and formatting.
– Over 4000 7-track tapes of global infrared
satellite data were read and reprocessed.
– Nearly 200,000 visible light images were
scanned, rectified and navigated.
– All the resultant data was converted to HDF-5
(NetCDF) format and freely distributed to users
from NASA and NSIDC servers.
– This data was then used to calculate monthly sea
ice extents for both the Arctic d the Antarctic.
• Conclusion: we (collectively) need to do more
of this! How can we fund it?
9. Usable (allow tools to run on it)
8. Citable (able to point & track citations)
7. Trusted (validated/checked by reviewers)
6. Reproducible (others can redo
experiments)
5. Discoverable (can be indexed by a system)
4. Comprehensible (others can understand
data & processes)
3. Accessible (can be accessed by others)
2. Archived (long-term & format-independent)
1. Preserved (existing in some form)
5. 2. Archive: Olive Project
9. Usable (allow tools to run on it)
8. Citable (able to point & track citations)
7. Trusted (validated/checked by reviewers)
6. Reproducible (others can redo
experiments)
5. Discoverable (can be indexed by a system)
4. Comprehensible (others can understand
data & processes)
3. Accessible (can be accessed by others)
2. Archived (long-term & format-independent)
1. Preserved (existing in some form)
• CMU CS & Library: funded by a grant
from the IMLS, Elsevier is partner
• Goal: Preservation of executable content
- nowadays a large part of intellectual
output, and very fragile
• Identified a series of software packages
and prepared VM to preserve
• Does it work? Yes – see video (1:24)
6. 3. Access: Urban Legend
9. Usable (allow tools to run on it)
8. Citable (able to point & track citations)
7. Trusted (validated/checked by reviewers)
6. Reproducible (others can redo
experiments)
5. Discoverable (can be indexed by a system)
4. Comprehensible (others can understand
data & processes)
3. Accessible (can be accessed by others)
2. Archived (long-term & format-independent)
1. Preserved (existing in some form)
• Part 1: Metadata acquisition
• Step through experimental process in series of dropdown
menus in simple web UI
• Can be tailored to workflow of individual researcher
• Connected to shared ontologies through lookup table,
managed centrally in lab
• Connect to data input console (Igor Pro)
7. 4. Comprehend: Urban Legend
9. Usable (allow tools to run on it)
8. Citable (able to point & track citations)
7. Trusted (validated/checked by reviewers)
6. Reproducible (others can redo
experiments)
5. Discoverable (can be indexed by a system)
4. Comprehensible (others can understand
data & processes)
3. Accessible (can be accessed by others)
2. Archived (long-term & format-independent)
1. Preserved (existing in some form)
• Part 2: Data Dashboard
• Access, select and manipulate data (calculate
properties, sort and plot)
• Final goal: interactive figures linked to data
• Plan to expand to more labs, other data
8. 5. Discover: Data Discovery Index
9. Usable (allow tools to run on it)
8. Citable (able to point & track citations)
7. Trusted (validated/checked by reviewers)
6. Reproducible (others can redo
experiments)
5. Discoverable (can be indexed by a system)
4. Comprehensible (others can understand
data & processes)
1. Preserved (existing in some form)
• NIH interested in creating DDI consortium
• Three places where data is deposited:
1. Curated sources for a single data type (e.g.Protein
Data Bank, VentDB, Hubble Space Data)
2. Non- or semicurated sources for different data types
(e.g. DataDryad, Dataverse, Figshare)
3. Tables in papers:
• Ways to find this:
– Cross-domain query tools, i.e. NIF, DataOne, etc
– Search for papers -> link to data
– How to find data in papers??
• Propose to build prototypes across all of these
data sources:
– Needs NLP, models of data patterns? What else?
3. Accessible (can be accessed by others)
2. Archived (long-term & format-independent)
Papers
Non-curated DBs
Curated DBs
9. 6. Reproduce: Resource Identifier Initiative
9. Usable (allow tools to run on it)
8. Citable (able to point & track citations)
7. Trusted (validated/checked by reviewers)
6. Reproducible (others can redo
experiments)
5. Discoverable (can be indexed by a system)
4. Comprehensible (others can understand
data & processes)
1. Preserved (existing in some form)
Force11 Working Group to add data identifiers
to articles that is
– 1) Machine readable;
– 2) Free to generate and access;
– 3) Consistent across publishers and journals.
• Authors publishing in participating journals
will be asked to provide RRID's for their
resources; these are added to the keyword
field
• RRID's will be drawn from:
– The Antibody Registry
– Model Organism Databases
– NIF Resource Registry
• So far, Springer, Wiley, Biomednet, Elsevier
journals have signed up with 11 journals,
more to come
• Wide community adoption!
3. Accessible (can be accessed by others)
2. Archived (long-term & format-independent)
10. 9. Usable (allow tools to run on it)
7.Trust: Moonrocks
8. Citable (able to point & track citations)
7. Trusted (validated/checked by reviewers)
6. Reproducible (others can redo
experiments)
5. Discoverable (can be indexed by a system)
4. Comprehensible (others can understand
data & processes)
3. Accessible (can be accessed by others)
2. Archived (long-term & format-independent)
1. Preserved (existing in some form)
How can we scale up data curation?
Pilot project with IEDA:
• Lunar geochemistry database:
leapfrog & improve curation time
• 1-year pilot, funded by Elsevier
• If spreadsheet columns/headers
map to RDB schema, we can scale up
curation process and move from
tables > curated databases!
11. 8. Cite: Force11 Data Citation Principles
9. Usable (allow tools to run on it)
8. Citable (able to point & track citations)
7. Trusted (validated/checked by reviewers)
6. Reproducible (others can redo
experiments)
5. Discoverable (can be indexed by a system)
4. Comprehensible (others can understand
data & processes)
1. Preserved (existing in some form)
• Another Force11 Working group
• Defined 8 principles:
• Now seeking endorsement/working on
implementation
3. Accessible (can be accessed by others)
2. Archived (long-term & format-independent)
1. Importance: Data should be considered legitimate, citable products of
research. Data citations should be accorded the same importance in
the scholarly record as citations of other research objects, such as
publications.
2. Credit and attribution: Data citations should facilitate giving scholarly
credit and normative and legal attribution to all contributors to the
data, recognizing that a single style or mechanism of attribution may
not be applicable to all data.
3. Evidence: Where a specific claim rests upon data, the corresponding
data citation should be provided.
4. Unique Identification: A data citation should include a persistent
method for identification that is machine actionable, globally unique,
and widely used by a community.
5. Access: Data citations should facilitate access to the data themselves
and to such associated metadata, documentation, and other materials,
as are necessary for both humans and machines to make informed use
of the referenced data.
6. Persistence: Metadata describing the data, and unique identifiers
should persist, even beyond the lifespan of the data they describe.
7. Versioning and granularity: Data citations should facilitate
identification and access to different versions and/or subsets of data.
Citations should include sufficient detail to verifiably link the citing
work to the portion and version of data cited.
8. Interoperability and flexibility: Data citation methods should be
sufficiently flexible to accommodate the variant practices among
communities but should not differ so much that they compromise
interoperability of data citation practices across communities.
12. 9. Use: Executable Papers
9. Usable (allow tools to run on it)
8. Citable (able to point & track citations)
7. Trusted (validated/checked by reviewers)
6. Reproducible (others can redo
experiments)
5. Discoverable (can be indexed by a system)
4. Comprehensible (others can understand
data & processes)
1. Preserved (existing in some form)
• Result of a challenge to come up with
cyberinfrastructure components to
enable executable papers
• Pilot in Computer Science journals
– See all code in the paper
– Save it, export it
– Change it and rerun on data set:
3. Accessible (can be accessed by others)
2. Archived (long-term & format-independent)
13. 10: Let’s allow our data to be happy!
Experimental Metadata:
Objects, Procedures, Properties
9. Usable (allow tools to run on it)
8. Citable (able to point & track citations)
7. Trusted (validated/checked by reviewers)
6. Reproducible (others can redo
experiments)
5. Discoverable (can be indexed by a system)
4. Comprehensible (others can understand
data & processes)
3. Accessible (can be accessed by others)
2. Archived (long-term & format-independent)
1. Preserved (existing in some form)
Execute: Direct settings on equipment,
circumstances of measurement
Raw Data
Analyze: Mathematical/computational
Processed processes and analytics
Data
Record Metadata:
DOI, Date, Author, Institute, etc.
Prepare: Reagents, species/specimen/cell
type, preparation details
Entity IDs
Validation Metadata:
Reproduction, Curation; Selection, Citation,
Usage, Metrics
14. Minimize your metadata footprint!
Reuse:
• ‘The good thing about standards is that there are
so many to choose from’
• Haendel et al looking at 54 (!!) data standards:
many have only been used once/for one group
• Employ a common element set + modular
additions over whole new schema
Recycle:
• Make sure you design upstream metadata
with downstream processes in mind
• Useful exercise: ‘buy a tag’ where
users/systems that will store/query/cite data
say what they need to do their job
• Learn from genetics: one datum can play
several different roles!
Reduce:
• Every tag needs to be added and read by
someone/thing: this adds cost and waste
• Consider ‘return on investment’ per metadata item
• TBL: what if “http://” was “h/”?