OSFair2017 Training | FAIR metrics - Starring your data setsOpen Science Fair
Peter Doorn, Marjan Grootveld & Elly Dijk talk about FAIR data principles and present the assessment tool that DANS is developing for data repositories | OSFair2017 Workshop
Workshop title: FAIR metrics - Starring your data sets
Workshop overview:
Do you want to join our effort to put the FAIR data principles into practice? Come and explore the assessment tool that DANS, Data Archiving and Networked Services in the Netherlands, is developing for data repositories.
The aim of our work is to implement the FAIR principles into a data assessment tool so that every dataset which is deposited or reused from any digital repository can be assessed in terms of a score on the principles Findable, Accessible, Interoperable, and Reusable, using a ‘FAIRness’ scale from 1 to 5 stars. In this interactive session participants can explore the pilot version of FAIRdat: the FAIR data assessment tool. The organisers would like to inform you about the project, and look forward to all feedback to improve the tool, or to improve the metrics that are used.
DAY 3 - PARALLEL SESSION 7
FAIR Data in Trustworthy Data Repositories Webinar - 12-13 December 2016| www...EUDAT
| www.eudat.eu | This webinar was co-organised by DANS, EUDAT and OpenAIRE and was held on 12th and 13th December 2016.
Everybody wants to play FAIR, but how do we put the principles into practice?
There is a growing demand for quality criteria for research datasets. In this webinar we will argue that the DSA (Data Seal of Approval for data repositories) and FAIR principles get as close as possible to giving quality criteria for research data. They do not do this by trying to make value judgements about the content of datasets, but rather by qualifying the fitness for data reuse in an impartial and measurable way. By bringing the ideas of the DSA and FAIR together, we will be able to offer an operationalization that can be implemented in any certified Trustworthy Digital Repository.
In 2014 the FAIR Guiding Principles (Findable, Accessible, Interoperable and Reusable) were formulated. The well-chosen FAIR acronym is highly attractive: it is one of these ideas that almost automatically get stuck in your mind once you have heard it. In a relatively short term, the FAIR data principles have been adopted by many stakeholder groups, including research funders.
The FAIR principles are remarkably similar to the underlying principles of DSA (2005): the data can be found on the Internet, are accessible (clear rights and licenses), in a usable format, reliable and are identified in a unique and persistent way so that they can be referred to. Essentially, the DSA presents quality criteria for digital repositories, whereas the FAIR principles target individual datasets.
In this webinar the two sets of principles will be discussed and compared and a tangible operationalization will be presented.
An introduction to the FAIR principles and a discussion of key issues that must be addressed to ensure data is findable, accessible, interoperable and re-usable. The session explored the role of the CDISC and DDI standards for addressing these issues.
Presented by Gareth Knight at the ADMIT Network conference, organised by the Association for Data Management in the Tropics, in Antwerp, Belgium on December 1st 2015.
OSFair2017 Training | FAIR metrics - Starring your data setsOpen Science Fair
Peter Doorn, Marjan Grootveld & Elly Dijk talk about FAIR data principles and present the assessment tool that DANS is developing for data repositories | OSFair2017 Workshop
Workshop title: FAIR metrics - Starring your data sets
Workshop overview:
Do you want to join our effort to put the FAIR data principles into practice? Come and explore the assessment tool that DANS, Data Archiving and Networked Services in the Netherlands, is developing for data repositories.
The aim of our work is to implement the FAIR principles into a data assessment tool so that every dataset which is deposited or reused from any digital repository can be assessed in terms of a score on the principles Findable, Accessible, Interoperable, and Reusable, using a ‘FAIRness’ scale from 1 to 5 stars. In this interactive session participants can explore the pilot version of FAIRdat: the FAIR data assessment tool. The organisers would like to inform you about the project, and look forward to all feedback to improve the tool, or to improve the metrics that are used.
DAY 3 - PARALLEL SESSION 7
FAIR Data in Trustworthy Data Repositories Webinar - 12-13 December 2016| www...EUDAT
| www.eudat.eu | This webinar was co-organised by DANS, EUDAT and OpenAIRE and was held on 12th and 13th December 2016.
Everybody wants to play FAIR, but how do we put the principles into practice?
There is a growing demand for quality criteria for research datasets. In this webinar we will argue that the DSA (Data Seal of Approval for data repositories) and FAIR principles get as close as possible to giving quality criteria for research data. They do not do this by trying to make value judgements about the content of datasets, but rather by qualifying the fitness for data reuse in an impartial and measurable way. By bringing the ideas of the DSA and FAIR together, we will be able to offer an operationalization that can be implemented in any certified Trustworthy Digital Repository.
In 2014 the FAIR Guiding Principles (Findable, Accessible, Interoperable and Reusable) were formulated. The well-chosen FAIR acronym is highly attractive: it is one of these ideas that almost automatically get stuck in your mind once you have heard it. In a relatively short term, the FAIR data principles have been adopted by many stakeholder groups, including research funders.
The FAIR principles are remarkably similar to the underlying principles of DSA (2005): the data can be found on the Internet, are accessible (clear rights and licenses), in a usable format, reliable and are identified in a unique and persistent way so that they can be referred to. Essentially, the DSA presents quality criteria for digital repositories, whereas the FAIR principles target individual datasets.
In this webinar the two sets of principles will be discussed and compared and a tangible operationalization will be presented.
An introduction to the FAIR principles and a discussion of key issues that must be addressed to ensure data is findable, accessible, interoperable and re-usable. The session explored the role of the CDISC and DDI standards for addressing these issues.
Presented by Gareth Knight at the ADMIT Network conference, organised by the Association for Data Management in the Tropics, in Antwerp, Belgium on December 1st 2015.
Core Trust Seal for Trustworthy Data Repositories, 2018-04-19Ciarán Quinn
The CONUL Research Group sponsored a workshop on the certification of digital repositories, focusing on the CoreTrustSeal framework. The workshop was presented by Dr John B Howard of University College Dublin, a founding member of the CoreTrustSeal Board of Directors.
The workshop reviewed the concept of "trust" in the context of data management and Open Science, identifying stakeholders and the issues that identify trustworthiness as a significant issue for managers of digital repository services. An overview of initiatives and services that provide a basis for the certification of digital repositories will be provided, including the European Framework for Audit & Certification of Digital Repositories.
Focus was made on the CoreTrustSeal, a framework that represents a merger of the previously separate Data Seal of Approval and ICSU/World Data Systems assessment and certification approaches. Attendees were introduced to the assessment process and requirements, including a walk through the 17 assessment categories in the CTS questionnaire.
This presentation was provided by Marilyn White, Katelynd Bucher, and Briget Wynne, all of NIST, during the NISO webinar, Engineering Access Under the Hood, Part Two, held on November 15, 2017.
How the Core Trust Seal (CTS) Enables FAIR Datadri_ireland
Presentation by Natalie Harrower, Director of the The Digital Repository of Ireland, on how the Core Trust Seal requirements and implementation process help prepare a digital repository for supporting FAIR data.
Presentation at the 'Services to Support FAIR data' workshop in Vienna on 24th April 2019. Workshop series supported by OpenAire, the Research Data Alliance, FAIRsFAIR and the EOSChub
How core trust seal enables FAIR data - Natalie HarrowerOpenAIRE
How core trust seal enables FAIR data presented Natalie Harrower during the OpenAIRE workshop Services to support FAIR data, Vienna: https://www.openaire.eu/openaire-workshop-making-services-fair-vienna-april-24th-2019
This presentation was provided by Adam Rusbridge of EDINA during a NISO webinar on the topic of Providing Access: Ensuring What Libraries Have Licensed is What Users Can Reach on Feb 8, 2017
Research Data Management: An Introductory Webinar from OpenAIRE and EUDATTony Ross-Hellauer
OpenAIRE and EUDAT co-present this webinar which aims to introduce researchers and others to the concept of research data management (RDM). As well as presenting the benefits of taking an active approach to research data management – including increased speed and ease of access, efficiency (fund once, reuse many times), and improved quality and transparency of research – the webinar will advise on strategies for successful RDM, resources to help manage data effectively, choosing where to store and deposit data, the EC H2020 Open Data Pilot and the basics of data management, stewardship and archiving.
Webinar recording available: http://www.instantpresenter.com/eifl/EB57D6888147
EUDAT Research Data Management | www.eudat.eu | EUDAT
| www.eudat.eu | The presentation gives an introduction to Research Data Management, explaining why it is important to manage and share data.
November 2016
The academic research data lifecycle. Session 1.4 of the RDMRose v3 materials.
The JISC funded RDMRose project (June 2012-May 2013) was a collaboration between the libraries of the University of Leeds, Sheffield and York, with the Information School at Sheffield to provide an Open Educational Resource for information professionals on Research Data Management. The materials were revised between November 2014 and February 2015 for the consortium of North West Academic Libraries (NoWAL).
http://www.sheffield.ac.uk/is/research/projects/rdmrose
Persistent Identifiers in EUDAT services| www.eudat.eu | EUDAT
| www.eudat.eu | The EUDAT data domain handles registered data. Each digital object should have a persistent identifier. This persistent identifier is used for: Replica identification; Identification of the repository of record (in the case of replication); Querying of additional information; Checksum (time stamped)...
Data Publishing Models by Sünje Dallmeier-Tiessendatascienceiqss
Data Publishing is becoming an integral part of scholarly communication today. Thus, it is indispensable to understand how data publishing works across disciplines. Are there best practices others can learn from or even data publishing standards? How do they impact interoperability in the Open Science landscape? The presentation will look at a range of examples, and the main building blocks of data publishing today. The work has been conducted as part of the RDA Data Publishing Workflows group.
Leverage DSpace for an enterprise, mission critical platformAndrea Bollini
Conference: Open Repository, Indianapolis, 8-12 June 2015
Presenters: Andrea Bollini, Michele Mennielli
Cineca, Italy
We would like to share with the DSpace Community some useful tips, starting from how to embed DSpace into a larger IT ecosystem that can provide additional value to the information managed. We will then show how publication data in DSpace - enriched with a proper use of the authority framework - can be combined with information coming from the HR system. Thanks to this, the system can provide rich and detailed reports and analysis through a business intelligence solution based on the Pentaho’s Mondrian OLAP open source data integration tools.
We will also present other use cases related to the management of publication information for reporting purpose: publication record has an extended lifecycle compared to the one in a basic IR; system load is much bigger, especially in writing, since the researchers need to be able to make changes to enrich data when new requirements come from the government or the university researcher office; data quality requires the ability to make distributed changes to the publication also after the conclusion of a validation workflow.
Finally we intend to present our direct experience and the challenges we faced to make DSpace easily and rapidly deployable to more than 60 sites.
A presentation on FAIR, FAIRsharing and the FAIR ecosystem for the ENVRI-FAIR community on the 13th December 2019. This presentation covers the basics of what FAIR is, how FAIRsharing can help 'FAIRify' standards, repositories, knowledgebases and data policies, and then the connections FAIRsharing has with other initiatives, such as the FAIR Evaluator, Data Stewardship Wizard, our RDA WG, GO-FAIR and EOSC-Life.
Core Trust Seal for Trustworthy Data Repositories, 2018-04-19Ciarán Quinn
The CONUL Research Group sponsored a workshop on the certification of digital repositories, focusing on the CoreTrustSeal framework. The workshop was presented by Dr John B Howard of University College Dublin, a founding member of the CoreTrustSeal Board of Directors.
The workshop reviewed the concept of "trust" in the context of data management and Open Science, identifying stakeholders and the issues that identify trustworthiness as a significant issue for managers of digital repository services. An overview of initiatives and services that provide a basis for the certification of digital repositories will be provided, including the European Framework for Audit & Certification of Digital Repositories.
Focus was made on the CoreTrustSeal, a framework that represents a merger of the previously separate Data Seal of Approval and ICSU/World Data Systems assessment and certification approaches. Attendees were introduced to the assessment process and requirements, including a walk through the 17 assessment categories in the CTS questionnaire.
This presentation was provided by Marilyn White, Katelynd Bucher, and Briget Wynne, all of NIST, during the NISO webinar, Engineering Access Under the Hood, Part Two, held on November 15, 2017.
How the Core Trust Seal (CTS) Enables FAIR Datadri_ireland
Presentation by Natalie Harrower, Director of the The Digital Repository of Ireland, on how the Core Trust Seal requirements and implementation process help prepare a digital repository for supporting FAIR data.
Presentation at the 'Services to Support FAIR data' workshop in Vienna on 24th April 2019. Workshop series supported by OpenAire, the Research Data Alliance, FAIRsFAIR and the EOSChub
How core trust seal enables FAIR data - Natalie HarrowerOpenAIRE
How core trust seal enables FAIR data presented Natalie Harrower during the OpenAIRE workshop Services to support FAIR data, Vienna: https://www.openaire.eu/openaire-workshop-making-services-fair-vienna-april-24th-2019
This presentation was provided by Adam Rusbridge of EDINA during a NISO webinar on the topic of Providing Access: Ensuring What Libraries Have Licensed is What Users Can Reach on Feb 8, 2017
Research Data Management: An Introductory Webinar from OpenAIRE and EUDATTony Ross-Hellauer
OpenAIRE and EUDAT co-present this webinar which aims to introduce researchers and others to the concept of research data management (RDM). As well as presenting the benefits of taking an active approach to research data management – including increased speed and ease of access, efficiency (fund once, reuse many times), and improved quality and transparency of research – the webinar will advise on strategies for successful RDM, resources to help manage data effectively, choosing where to store and deposit data, the EC H2020 Open Data Pilot and the basics of data management, stewardship and archiving.
Webinar recording available: http://www.instantpresenter.com/eifl/EB57D6888147
EUDAT Research Data Management | www.eudat.eu | EUDAT
| www.eudat.eu | The presentation gives an introduction to Research Data Management, explaining why it is important to manage and share data.
November 2016
The academic research data lifecycle. Session 1.4 of the RDMRose v3 materials.
The JISC funded RDMRose project (June 2012-May 2013) was a collaboration between the libraries of the University of Leeds, Sheffield and York, with the Information School at Sheffield to provide an Open Educational Resource for information professionals on Research Data Management. The materials were revised between November 2014 and February 2015 for the consortium of North West Academic Libraries (NoWAL).
http://www.sheffield.ac.uk/is/research/projects/rdmrose
Persistent Identifiers in EUDAT services| www.eudat.eu | EUDAT
| www.eudat.eu | The EUDAT data domain handles registered data. Each digital object should have a persistent identifier. This persistent identifier is used for: Replica identification; Identification of the repository of record (in the case of replication); Querying of additional information; Checksum (time stamped)...
Data Publishing Models by Sünje Dallmeier-Tiessendatascienceiqss
Data Publishing is becoming an integral part of scholarly communication today. Thus, it is indispensable to understand how data publishing works across disciplines. Are there best practices others can learn from or even data publishing standards? How do they impact interoperability in the Open Science landscape? The presentation will look at a range of examples, and the main building blocks of data publishing today. The work has been conducted as part of the RDA Data Publishing Workflows group.
Leverage DSpace for an enterprise, mission critical platformAndrea Bollini
Conference: Open Repository, Indianapolis, 8-12 June 2015
Presenters: Andrea Bollini, Michele Mennielli
Cineca, Italy
We would like to share with the DSpace Community some useful tips, starting from how to embed DSpace into a larger IT ecosystem that can provide additional value to the information managed. We will then show how publication data in DSpace - enriched with a proper use of the authority framework - can be combined with information coming from the HR system. Thanks to this, the system can provide rich and detailed reports and analysis through a business intelligence solution based on the Pentaho’s Mondrian OLAP open source data integration tools.
We will also present other use cases related to the management of publication information for reporting purpose: publication record has an extended lifecycle compared to the one in a basic IR; system load is much bigger, especially in writing, since the researchers need to be able to make changes to enrich data when new requirements come from the government or the university researcher office; data quality requires the ability to make distributed changes to the publication also after the conclusion of a validation workflow.
Finally we intend to present our direct experience and the challenges we faced to make DSpace easily and rapidly deployable to more than 60 sites.
A presentation on FAIR, FAIRsharing and the FAIR ecosystem for the ENVRI-FAIR community on the 13th December 2019. This presentation covers the basics of what FAIR is, how FAIRsharing can help 'FAIRify' standards, repositories, knowledgebases and data policies, and then the connections FAIRsharing has with other initiatives, such as the FAIR Evaluator, Data Stewardship Wizard, our RDA WG, GO-FAIR and EOSC-Life.
Turning FAIR into Reality - Role for Libraries dri_ireland
Presentation by Dr. Natalie Harrower, Director Digital Repository of Ireland and European Commission FAIR data expert group member, on what role librarians can play in the FAIR ecosystem. "Applying the FAIR data principles in day-to-day library practice" session by the Research Data Management Working Group, LIBER Steering Committee Research Infrastructures, LIBER2019, Dublin, 26 June 2019
PARTHENOS Common Policies and Implementation StrategiesParthenos
Presentation by Hella Hollander for the PARTHENOS workshop "Introducing PARTHENOS - Integrating the Digital Humanities" on 14 December 2016 in Prato, Italy.
Researchers require infrastructures that ensure a maximum of accessibility, stability and reliability to facilitate working with and sharing of research data. Such infrastructures are being increasingly summarised under the term Research Data Repositories (RDR). The project re3data.org – Registry of Research Data Repositories – began to index research data repositories in 2012 and offers researchers, funding organisations, libraries and publishers an overview of the heterogeneous research data repository landscape. In December 2014 re3data.org listed more than 1,030 research data repositories, which are described in detail using the re3data.org schema (http://dx.doi.org/10.2312/re3.003). Information icons help researchers to identify easily an adequate repository for the storage and reuse of their data. This talk describes the heterogeneous RDR landscape and presents a typology of institutional, disciplinary, multidisciplinary and project-specific RDR. Further, it outlines the features of re3data. org and it shows current developments for integration into data management planning tools and other services.
By the end of 2015 re3data.org and Databib (Purdue University, USA) will merge their services, which will then be managed under the auspices of DataCite. The aim of this merger is to reduce duplication of effort and to serve the research community better with a single, sustainable registry of research data repositories. The talk will present this organisational development as a best practice example for the development of international research information services.
Research Data Management: An Introductory Webinar from OpenAIRE and EUDATOpenAIRE
OpenAIRE and EUDAT co-present this webinar which aims to introduce researchers and others to the concept of research data management (RDM). As well as presenting the benefits of taking an active approach to research data management – including increased speed and ease of access, efficiency (fund once, reuse many times), and improved quality and transparency of research – the webinar will advise on strategies for successful RDM, resources to help manage data effectively, choosing where to store and deposit data, the EC H2020 Open Data Pilot and the basics of data management, stewardship and archiving.
Webinar recording available: http://www.instantpresenter.com/eifl/EB57D6888147
This slideshow was used in an Introduction to Research Data Management course taught for the Mathematical, Physical and Life Sciences Division, University of Oxford, on 2017-02-15. It provides an overview of some key issues, looking at both day-to-day data management, and longer term issues, including sharing, and curation.
Presentation by Hugo Leroux and Liming Zhu, CSIRO, to the 'Unlocking value from publicly funded Clinical Research Data' workshop, cohosted by ARDC and CSIRO at ANU on 6 March 2019.
PIDs, Data and Software: How Libraries Can Support Researchers in an Evolving...Sarah Anna Stewart
Presentation given at the M25 Consortium of Academic Libraries, CPD25 Event on 'The Role of the Library in Supporting Research'. Provides an introduction to data, software and PIDs and a brief look at how libraries can enable researchers to gain impact and credit for their research data and software.
Application of recently developed FAIR metrics to the ELIXIR Core Data ResourcesPistoia Alliance
The FAIR (Findable, Accessible, Interoperable and Reusable) principles aim to maximize the discovery and reuse of digital resources. Using recently developed software and metrics to assess FAIRness and supported through an ELIXIR Implementation Study, Michel worked with a subset of ELIXIR Core Data Resources to apply these technologies. In this webinar, he will discuss their approach, findings, and lessons learned towards the understanding and promotion of the FAIR principles.
FAIRsharing - Mapping the Landscape of Databases, Repositories, Standards and...Peter McQuilton
A 15 minute slide set presented at two workshops at #biocuration2019; the first on ontologies and FAIRification, the second to map the landscape of biocuration.
This presentation introduced participants to the DC 101 course and was given at the Digital Curation and Preservation Outreach and Capacity Building Workshop in Belfast on September 14-15 2009.
http://www.dcc.ac.uk/events/workshops/digital-curation-and-preservation-outreach-and-capacity-building-workshop
OU Library Research Support webinar: Working with research dataIzzyChad
Slides from a webinar delivered on 31st January 2018 for OU research staff and students. Covers practical strategies for managing research data, including policies, file naming, information security, metadata and working with sensitive data.
Similar to 2010 CLARA Nijmegen - Data Seal of Approval tutorial (20)
Text-Fabric: how to do text research in a FAIR way.
Text is one of the simplest and most common data types in computer science.
But there is a lot in text that does not meet the eye, and so people have been annotating texts, century-by-century.
When you research texts, you consume and produce such annotations.
Suddenly you find yourself in the midst of a big fabric of thoughts, contributed by many authors.
Text-Fabric is a tool that helps you to follow the threads that came before you and to weave a few of your own and add them to the scholarly record.
I'll show you how that looks for clay tablets of the Uruk period (the oldest writing on earth), the much more recent Hebrew Bible, and the ultramodern General Missives of the VOC time.
Towards TextPy, a module for processing text.
If we define annotated text as a graph with additional structure, we can make text processing more efficient, in the same way that Pandas makes processing dataframes more efficient.
We demonstrate how Text-Fabric can handle the display of text and annotations, even when chunks of text are not properly embedded in each other. This demo contains examples from the Hebrew Bible and the Old Babylonian Letters (cuneiform clay tablets).
Researchers in ancient text corpora can take control over their data. We show a way to do so by means of Text-Fabric.
Co-production of Cody Kingham and Dirk Roorda
Biblia Hebraica Stuttgartensia Amstelodamensis. Coding the Hebrew Bible with an Open Science ethos: Text-Fabric.
Text-Fabric is several things: (1) a browser for ancient text corpora; (2) a Python3 package for processing ancient corpora
A corpus of ancient texts and linguistic annotations represents a large body of knowledge. Text-Fabric makes that knowledge accessible to non-programmers by means of built-in a search interface that runs in your browser.
From there the step to program your own analytics is not so big anymore. Because you can call the Text-Fabric API from your Python programs, and it works really well in Jupyter notebooks.
Developing a tool for handling text with linguistic annotations. Text-Fabric is meant to support researchers that wnat to contribute portions of the data, and weaves the contributions in into a meaningful whole. Currently, it is primarily meant for working with the Hebrew Bible, based on the ETCBC (Amsterdam) linguistic database.
Conference presentation for 2016 annual meeting of the Society of Biblical Literature, San Antonio. (https://www.sbl-site.org).
Authors: Janet Dyk (linguistic ideas) and Dirk Roorda (computational implementation).
A verb organizes the elements in a sentence. Different patterns of constituents affect the meaning of a verb in a given context. The potential of a verb to combine with patterns of elements is known as its valence. A single set of questions, organized as a flow chart, selects the relevant building blocks within the context of a verb. The resulting pattern provides a particular significance for the verb in question. Because all contexts are submitted to the same flow chart, similarities and differences between verbs come to light. For example, verbs of movement in their causative formation manifest the same patterns as transitive verbs with an object that gets moved. We apply this approach to the whole Hebrew Bible, using the database of the Eep Talstra Centre for Bible and Computer (ETCBC), which contains the relevant linguistic annotations. This allows us to have a complete listing of all patterns for all verbs. It provides the basis for consistent proposals for the significance of specific patterns occurring with a particular verb. The valence results are made available in SHEBANQ, an online research tool based on the ETCBC database. It presents the basic data, text and linguistic features, together with annotations by researchers. The valence results consist of a set of algorithmically generated annotations which show up between the lines of the text. The algorithm itself and its documentation can be found at https://shebanq.ancient-data.org/tools?goto=valence. By using SHEBANQ we achieve several goals with respect to the scholarly workflow: (1) all our results are openly accessible online, and other researchers may comment on them; (2) all resources needed to reproduce this research are available online and can be downloaded (Open Access).
Text as Data: processing the Hebrew BibleDirk Roorda
The merits of stand-off markup (LAF) versus inline markup (TEI) for processing text as data. Ideas applied to work with the Hebrew Bible, resulting in tools for researchers and end-users.
Datamanagement for Research: A Case StudyDirk Roorda
How practices of data sharing can help researchers to produce more science.
Session in the data management course organized by RDNL (Research Data in the Netherlands)
Hebrew Bible as Data: Laboratory, Sharing, LessonsDirk Roorda
Recently, the Hebrew Bible has been published online as a database. We show what you can do with it, and how to share your results with others. Work by the Amsterdam scholars of the Eep Talstra Centre for Bible and Computer, supported by CLARIN-NL.
LAF-Fabric: a tool to process the ETCBC Hebrew Text Database in Linguistic Annotation Framework.
How researchers in theology and linguistics can create workflows to analyse the text of the Hebrew Bible and extract data for visualization. Those workflows can be written in Python, and run conveniently in the IPython Notebook.
Joint work with Martijn Naaijer (VU University).
With the Hebrew Bible encoded in Linguistic Annotation Framework (LAF-ISO), and with a new LAF processing tool, we demonstrate how you can do practical data analysis. The tool, LAF-Fabric, integrates with the ipython notebook approach. Our example here is lexeme cooccurrence analysis of bible books. For now, the road from data to visualization is more important than the exact visualization.
This is a presentation by Dada Robert in a Your Skill Boost masterclass organised by the Excellence Foundation for South Sudan (EFSS) on Saturday, the 25th and Sunday, the 26th of May 2024.
He discussed the concept of quality improvement, emphasizing its applicability to various aspects of life, including personal, project, and program improvements. He defined quality as doing the right thing at the right time in the right way to achieve the best possible results and discussed the concept of the "gap" between what we know and what we do, and how this gap represents the areas we need to improve. He explained the scientific approach to quality improvement, which involves systematic performance analysis, testing and learning, and implementing change ideas. He also highlighted the importance of client focus and a team approach to quality improvement.
The French Revolution, which began in 1789, was a period of radical social and political upheaval in France. It marked the decline of absolute monarchies, the rise of secular and democratic republics, and the eventual rise of Napoleon Bonaparte. This revolutionary period is crucial in understanding the transition from feudalism to modernity in Europe.
For more information, visit-www.vavaclasses.com
Synthetic Fiber Construction in lab .pptxPavel ( NSTU)
Synthetic fiber production is a fascinating and complex field that blends chemistry, engineering, and environmental science. By understanding these aspects, students can gain a comprehensive view of synthetic fiber production, its impact on society and the environment, and the potential for future innovations. Synthetic fibers play a crucial role in modern society, impacting various aspects of daily life, industry, and the environment. ynthetic fibers are integral to modern life, offering a range of benefits from cost-effectiveness and versatility to innovative applications and performance characteristics. While they pose environmental challenges, ongoing research and development aim to create more sustainable and eco-friendly alternatives. Understanding the importance of synthetic fibers helps in appreciating their role in the economy, industry, and daily life, while also emphasizing the need for sustainable practices and innovation.
Palestine last event orientationfvgnh .pptxRaedMohamed3
An EFL lesson about the current events in Palestine. It is intended to be for intermediate students who wish to increase their listening skills through a short lesson in power point.
How to Create Map Views in the Odoo 17 ERPCeline George
The map views are useful for providing a geographical representation of data. They allow users to visualize and analyze the data in a more intuitive manner.
Operation “Blue Star” is the only event in the history of Independent India where the state went into war with its own people. Even after about 40 years it is not clear if it was culmination of states anger over people of the region, a political game of power or start of dictatorial chapter in the democratic setup.
The people of Punjab felt alienated from main stream due to denial of their just demands during a long democratic struggle since independence. As it happen all over the word, it led to militant struggle with great loss of lives of military, police and civilian personnel. Killing of Indira Gandhi and massacre of innocent Sikhs in Delhi and other India cities was also associated with this movement.
We all have good and bad thoughts from time to time and situation to situation. We are bombarded daily with spiraling thoughts(both negative and positive) creating all-consuming feel , making us difficult to manage with associated suffering. Good thoughts are like our Mob Signal (Positive thought) amidst noise(negative thought) in the atmosphere. Negative thoughts like noise outweigh positive thoughts. These thoughts often create unwanted confusion, trouble, stress and frustration in our mind as well as chaos in our physical world. Negative thoughts are also known as “distorted thinking”.
Model Attribute Check Company Auto PropertyCeline George
In Odoo, the multi-company feature allows you to manage multiple companies within a single Odoo database instance. Each company can have its own configurations while still sharing common resources such as products, customers, and suppliers.
Instructions for Submissions thorugh G- Classroom.pptxJheel Barad
This presentation provides a briefing on how to upload submissions and documents in Google Classroom. It was prepared as part of an orientation for new Sainik School in-service teacher trainees. As a training officer, my goal is to ensure that you are comfortable and proficient with this essential tool for managing assignments and fostering student engagement.
Read| The latest issue of The Challenger is here! We are thrilled to announce that our school paper has qualified for the NATIONAL SCHOOLS PRESS CONFERENCE (NSPC) 2024. Thank you for your unwavering support and trust. Dive into the stories that made us stand out!
The Roman Empire A Historical Colossus.pdfkaushalkr1407
The Roman Empire, a vast and enduring power, stands as one of history's most remarkable civilizations, leaving an indelible imprint on the world. It emerged from the Roman Republic, transitioning into an imperial powerhouse under the leadership of Augustus Caesar in 27 BCE. This transformation marked the beginning of an era defined by unprecedented territorial expansion, architectural marvels, and profound cultural influence.
The empire's roots lie in the city of Rome, founded, according to legend, by Romulus in 753 BCE. Over centuries, Rome evolved from a small settlement to a formidable republic, characterized by a complex political system with elected officials and checks on power. However, internal strife, class conflicts, and military ambitions paved the way for the end of the Republic. Julius Caesar’s dictatorship and subsequent assassination in 44 BCE created a power vacuum, leading to a civil war. Octavian, later Augustus, emerged victorious, heralding the Roman Empire’s birth.
Under Augustus, the empire experienced the Pax Romana, a 200-year period of relative peace and stability. Augustus reformed the military, established efficient administrative systems, and initiated grand construction projects. The empire's borders expanded, encompassing territories from Britain to Egypt and from Spain to the Euphrates. Roman legions, renowned for their discipline and engineering prowess, secured and maintained these vast territories, building roads, fortifications, and cities that facilitated control and integration.
The Roman Empire’s society was hierarchical, with a rigid class system. At the top were the patricians, wealthy elites who held significant political power. Below them were the plebeians, free citizens with limited political influence, and the vast numbers of slaves who formed the backbone of the economy. The family unit was central, governed by the paterfamilias, the male head who held absolute authority.
Culturally, the Romans were eclectic, absorbing and adapting elements from the civilizations they encountered, particularly the Greeks. Roman art, literature, and philosophy reflected this synthesis, creating a rich cultural tapestry. Latin, the Roman language, became the lingua franca of the Western world, influencing numerous modern languages.
Roman architecture and engineering achievements were monumental. They perfected the arch, vault, and dome, constructing enduring structures like the Colosseum, Pantheon, and aqueducts. These engineering marvels not only showcased Roman ingenuity but also served practical purposes, from public entertainment to water supply.
3. Overview
• Introduction and Theory
• qualities
• trust, simplicity
• guidelines
• Process and Demo
• assessment and review
• Discussion and Application
• CLARIN centers
• language resources
8. Quality control
• by the stakeholders
• data producers
• data custodians
• date consumers
• custodians = repositories
• substantial role for repositories
• guidelines for producers
• agreements for consumers
9. Quality issues
• metadata standards
• CMDI and www.isocat.org
• preferred formats
• TEI, XML
• referencing systems
• persistent identifiers
• long term preservation
• after the live-environment has died off
• interoperability
• OAI-PMH
10. Quality issues
• search engines
• CLARIN search and develop
• access rights
• comply with privacy law, copyright law
• respect people from which data is obtained
• accountability
• for all repository operations
11. Quality and Trust
• imperfection lurks everywhere
• trust works where certainty blocks
• trust is a process
• to greater quality
• to better relationships
• to more certainty
12. Quality and Simplicity
reduce organize
time learn differences
context
emotion trust
failure
focus:
subtract what is obvious
add what is meaningful
http://lawsofsimplicity.com/
13. Guidelines: producers
http://www.datasealofapproval.org/
1.The data producer deposits the research data in
a data repository with sufficient information for
others to assess the scientific and scholarly quality
of the research data and compliance with
disciplinary and ethical norms.
2. The data producer provides the research data in
formats recommended by the data repository
3. The data producer provides the research data
together with the metadata requested by the data
repository
14. Guidelines: consumers
http://www.datasealofapproval.org/
14. The data consumer complies with access
regulations set by the data repository
15. The data consumer conforms to and agrees
with any codes of conduct that are generally
accepted in higher education and research for the
exchange and proper use of knowledge and
information
16. The data consumer respects the applicable
licenses of the data repository regarding the use of
the research data
15. Guidelines: repositories
http://www.datasealofapproval.org/
4. The data repository has an explicit mission in the area
of digital archiving and promulgates it
5. The data repository uses due diligence to ensure
compliance with legal regulations and contracts
including, when applicable, regulations governing the
protection of human subjects.
6. The data repository applies documented processes
and procedures for managing data storage
7. The data repository has a plan for long-term
preservation of its digital assets
16. Guidelines: repositories
http://www.datasealofapproval.org/
8. Archiving takes place according to explicit workflows across the
data life cycle
9. The data repository assumes responsibility from the data
producers for access and availability of the digital objects
10. The data repository enables the users to utilize the research
data and refer to them
11. The data repository ensures the integrity of the digital objects
and the metadata
12. The data repository ensures the authenticity of the digital
objects and the metadata
13. The technical infrastructure explicitly supports the tasks and
functions described in internationally accepted archival standards
like OAIS
17. Guidelines: outsourcing
http://www.datasealofapproval.org/
repositories may outsource digital preservation
to specialist repositories
• implement all except 4,6,7,8 and 13
• store a copy of the data in another (TDR) that
• has acquired the DSA logo
• by implementing each of the sixteen guidelines
• (including 4, 6, 7, 8 and 13).
18. Seal of Approvement
• a repository shows it on its webpage
• if conditions are fulfilled
• as testified by
• a self-assessment
• with reviews
• on a yearly basis
• the exact level of compliance is
• transparently published under the seal
19. Assessment and review
minimum requirements
threshold will go up
as time proceeds
score actions taken comments issues
* nothing done give a reason
** theoretical concept point to initiation doc describe main issues
*** implementation phase point to definition doc describe main issues
**** fully implemented point to definition doc
N/A not applicable give a reason
20. Organisation
• repositories represented by a board
• tools to facilitate the procedure
• modifiaction record
• the DSA website links to compliant
repositories
21.
22.
23. CLARIN centres
• A = provide infrastructure
• managing the federation
• B = provide services
• data and webservices
• C = provide metadata
• harvestable metadata
• R = respected = recognised
• offer LRT resources in whatever form
• E = external
• offer non-LRT resources or services
• identity federations
• national libraries
24. Group assignment
• P(roducers)
• invent p-guidelines for B/C centers
• R(epositories)
• invent r-guidelines for A/B centers
• C(onsumers)
• invent c-guidelines for B/C/R centers
Suggestions for
• assessment
• review
• modification record
25. Wrap-up: P-Group
metadata about background
information about researchers
who, why, publications
DAI
In IMDI it is difficult to update information, affiliation updates,
use unique identifiers for participants in building a corpus, store records of people,
and link from the metadata of resources to the records of people
using formats depending on formats
formats maybe standardised, but not usable to researchers, I do not want to wrap
my data in dead formats: the repositories should support innovation in this respect,
when it is driven by researchers
26. Wrap up: C-group
goal is: finding info in a repository
we need:
overview of access rights
proper web-connection to the repository
user-friendly interface
low threshold for feedback for new features
we should be part of the chain in the design of the access tools
GUIDELINES
WE WANT ALL CENTERS IN THE CHAIN THAT PROVIDE US WITH THE
INFORMATION WE NEED TO OFFER US TRANSPARENCY AND VERIFIABILITY
ON HOW THEIR DATA IS OBTAINED, PROCESSED AND
CONTROLLED/MANAGED
WE WANT TOOLS WITH CLEAR COPYRIGHT PERMISSIONS THAT HAVE A
27. Wrap-up: R-group
we provide infrastructure and management for data
we want to standardize our stuff
we need knowledge, the right metadata of the stuff that is coming to us
we want the materials in the right format, allowing for some flexibility
retro-archiving: we offer tools for converting legacy data, so that producers may submit
raw materials
management of data concerning legal access
protect the providers, so that the providers can trust the consumers: licensing forms
share knowledge about services we provide with
potential users: people working in the field
other repositories
we want a forum as an instrument for developing trust between producers and
consumers: the community becomes more transparent
1reduce (restrict to the most important issues, a few guidelines will do)2organize (group the guidelines in sections for producers, custodians, consumers)3time (save time by a smooth assessment process)4learn (use expertise in preservation)5differences (reintroduce complexity in a controlled way, because sometimes it is needed)6context (exploit knowledge of the community, requirements of the users)7emotion (do not make it purely bureaucratical, keep the feeling of value, enjoy good relationships with stakeholders)8trust (by default trust, but know where your undo button is, even against the ones you trust)9failure (learn from failures, improve the guidelines, the assessment procedures)10focus: subtract what is obvious, add what is meaningful (this is not about the data in bank accounts, nor highly sensitive medical data, nor company archives, but about research data: the scientific value is protected by the guidelines)