This document discusses the importance of research data management. It defines research data management as organizing data from collection through dissemination and archiving to ensure results can be verified and built upon. Good research data management practices include developing data management plans, making data FAIR (findable, accessible, interoperable, and reusable) according to common standards, and storing data in trusted repositories with support from data stewards. Following best practices helps ensure research integrity and allows others to build upon existing work.
The document summarizes Hendrik Drachsler's presentation at an NSF expert meeting on big data and privacy in human subjects research. Some key points from Drachsler's presentation include:
- He discussed issues around learning analytics research and how privacy concerns often stop innovation;
- He questioned if big data should be considered the "new truth" and highlighted examples where big data provided inaccurate insights;
- Drachsler advocated for transparency, data security, informed consent and data anonymization to prevent issues like what happened with the inBloom student database project in the US.
In scientific communication, we observe a complex interaction of several stakeholder groups, each of which have distinct interests, strategies and approaches for Open Access and Open Data. The German government initiated a “Commission for the Future of the Information Infrastructure” (KII) in Germany. In this commission, most of the stakeholders are working together in order to design a future scenario for the supply of scientific information. The KII’s evaluation and recommendations for Open Access as well as research data will be particularly highly recognized and will significantly influence Open Access and Open Data developments in Germany.
I will outline the current situation in Germany – players and their interactions in terms of Open Access and Open Data – and present two initiatives and their work in detail. One of them, the KII process, will show the official site, the other one will show the grassroots site of the story.
TourPack: Packaging and Disseminating Touristic Services with Linked Data and...Anna Fensel
The document summarizes the TourPack project, which aims to build a linked data system for packaging and disseminating touristic services. Key goals are to integrate information from multiple sources, automatically generate optimized travel packages, and efficiently publish and book packages through multiple channels like social media and mobile apps. The technical approach involves semantic annotation of services using schemas like Schema.org to enable automatic clustering, packaging and publishing of offers across different platforms. Expected outcomes include scalable multi-channel communication solutions and methods for online interactions and booking of tourism services.
Online Marketing with Schema.org and Multi-channel CommunicationAnna Fensel
This document discusses using semantic annotations and a multi-channel communication tool to increase the online visibility and marketing success of hotels. It describes how the Kaysers Hotel semantically annotated its website pages in multiple languages to improve search engine optimization. A multi-channel communication tool was also used to suggest and publish social media posts for the hotel across different channels. An evaluation found the hotel's website traffic and traffic from social media increased while time spent on social media marketing decreased. Future work could involve extending semantic standards for the hotel domain and applying semantic technologies more broadly to online communication.
This document discusses a project called MEM0R1ES that aims to automatically organize a person's digital information from various devices and online services to generate useful digital memories. The project develops techniques for entity search, typing, clustering, and elicitation to extract, integrate and expose personal information from heterogeneous graphs. It has produced several open-source software components and published results in top conferences. The document outlines current research directions and concludes that the project addresses important societal issues through stimulating collaboration between institutions.
Austrian Experience in Building Data Value ChainAnna Fensel
- The document discusses open government data developments in Austria, including data.gv.at, which provides a central portal for Austrian open government data, and the Open Data Vienna Challenge contest, which resulted in around 80 apps being developed using open government data.
- It describes Linked Open Data (LOD) as a global data integration platform and some of the techniques used for data integration, including normalizing vocabularies and resolving entity identifiers.
- The Tourist Map Austria project is presented as a case study that combines open data and services through LOD to provide an integrated tourist information app and booking platform.
The document discusses curating and profiling linked data for educational applications. It describes the LinkedUp project, which aims to advance the use of open data and linked data technologies in education. The LinkedUp approach involves collecting and exposing open educational datasets, profiling the datasets to generate metadata, and linking datasets to create an "educational data graph." The profiling process extracts topic information from datasets by identifying entities, normalizing categories, and computing relevance scores to generate structured dataset profiles. This facilitates browsing, exploring, and querying across educational linked datasets.
The document summarizes Hendrik Drachsler's presentation at an NSF expert meeting on big data and privacy in human subjects research. Some key points from Drachsler's presentation include:
- He discussed issues around learning analytics research and how privacy concerns often stop innovation;
- He questioned if big data should be considered the "new truth" and highlighted examples where big data provided inaccurate insights;
- Drachsler advocated for transparency, data security, informed consent and data anonymization to prevent issues like what happened with the inBloom student database project in the US.
In scientific communication, we observe a complex interaction of several stakeholder groups, each of which have distinct interests, strategies and approaches for Open Access and Open Data. The German government initiated a “Commission for the Future of the Information Infrastructure” (KII) in Germany. In this commission, most of the stakeholders are working together in order to design a future scenario for the supply of scientific information. The KII’s evaluation and recommendations for Open Access as well as research data will be particularly highly recognized and will significantly influence Open Access and Open Data developments in Germany.
I will outline the current situation in Germany – players and their interactions in terms of Open Access and Open Data – and present two initiatives and their work in detail. One of them, the KII process, will show the official site, the other one will show the grassroots site of the story.
TourPack: Packaging and Disseminating Touristic Services with Linked Data and...Anna Fensel
The document summarizes the TourPack project, which aims to build a linked data system for packaging and disseminating touristic services. Key goals are to integrate information from multiple sources, automatically generate optimized travel packages, and efficiently publish and book packages through multiple channels like social media and mobile apps. The technical approach involves semantic annotation of services using schemas like Schema.org to enable automatic clustering, packaging and publishing of offers across different platforms. Expected outcomes include scalable multi-channel communication solutions and methods for online interactions and booking of tourism services.
Online Marketing with Schema.org and Multi-channel CommunicationAnna Fensel
This document discusses using semantic annotations and a multi-channel communication tool to increase the online visibility and marketing success of hotels. It describes how the Kaysers Hotel semantically annotated its website pages in multiple languages to improve search engine optimization. A multi-channel communication tool was also used to suggest and publish social media posts for the hotel across different channels. An evaluation found the hotel's website traffic and traffic from social media increased while time spent on social media marketing decreased. Future work could involve extending semantic standards for the hotel domain and applying semantic technologies more broadly to online communication.
This document discusses a project called MEM0R1ES that aims to automatically organize a person's digital information from various devices and online services to generate useful digital memories. The project develops techniques for entity search, typing, clustering, and elicitation to extract, integrate and expose personal information from heterogeneous graphs. It has produced several open-source software components and published results in top conferences. The document outlines current research directions and concludes that the project addresses important societal issues through stimulating collaboration between institutions.
Austrian Experience in Building Data Value ChainAnna Fensel
- The document discusses open government data developments in Austria, including data.gv.at, which provides a central portal for Austrian open government data, and the Open Data Vienna Challenge contest, which resulted in around 80 apps being developed using open government data.
- It describes Linked Open Data (LOD) as a global data integration platform and some of the techniques used for data integration, including normalizing vocabularies and resolving entity identifiers.
- The Tourist Map Austria project is presented as a case study that combines open data and services through LOD to provide an integrated tourist information app and booking platform.
The document discusses curating and profiling linked data for educational applications. It describes the LinkedUp project, which aims to advance the use of open data and linked data technologies in education. The LinkedUp approach involves collecting and exposing open educational datasets, profiling the datasets to generate metadata, and linking datasets to create an "educational data graph." The profiling process extracts topic information from datasets by identifying entities, normalizing categories, and computing relevance scores to generate structured dataset profiles. This facilitates browsing, exploring, and querying across educational linked datasets.
Big data is arising from multiple sources at high volume, velocity, and variety. Pull technology relies on users requesting information, as seen on early search engines and websites, while push technology proactively sends updates without requests. As big data grows, push technology faces challenges in speed and completeness of real-time updates for users, while pull relies on users to initiate longer downloads of large files. Both will continue evolving to optimize access and delivery of massive datasets.
Ontology Building vs Data Harvesting and Cleaning for Smart-city ServicesPaolo Nesi
Presently, a very large number of public and private data sets are available around the local governments. In most cases, they are not semantically interoperable and a huge human effort is needed to create integrated ontologies and knowledge base for smart city. Smart City ontology is not yet standardized, and a lot of research work is needed to identify models that can easily support the data reconciliation, the management of the complexity and reasoning. In this paper, a system for data ingestion and reconciliation of smart cities related aspects as road graph, services available on the roads, traffic sensors etc., is proposed. The system allows managing a big volume of data coming from a variety of sources considering both static and dynamic data. These data are mapped to smart-city ontology and stored into an RDF-Store where they are available for applications via SPARQL queries to provide new services to the users. The paper presents the process adopted to produce the ontology and the knowledge base and the mechanisms adopted for the verification, reconciliation and validation. Some examples about the possible usage of the coherent knowledge base produced are also offered and are accessible from the RDF-Store and related services. The article also presented the work performed about reconciliation algorithms and their comparative assessment and selection. Keywords Smart city, knowledge base construction, reconciliation, validation and verification of knowledge base, smart city ontology, linked open graph.
This document summarizes a presentation on data visualization. It introduces data visualization and its uses for exploring data, explaining results, and distant reading. It discusses the building blocks of visualization like charts, networks, and visualizing different data types. It explores some scholarly visualizations and exercises critiquing them. It also covers extracting data from text, images and video using computational methods, and preparing messy humanities data for visualization, including dealing with uncertainty. The presentation emphasizes choosing visualizations based on purpose, data, audience and structure. It recommends tools for creating simple visualizations like Viewshare that don't require programming.
Introduction to information visualisation for humanities PhDsMia
Training workshop for the CHASE Arts and Humanities in the Digital Age programme. (
This session will give you an overview of a variety of techniques and tools available for data visualisation and analysis in the humanities. You will learn about common types of visualisations and the role of exploratory and explanatory visualisations, explore examples of scholarly visualisations, try some visualisation tools, and know where to find further information about analysing and building data visualisations.
The HathiTrust Research Center: Enabling New Knowledge Through Shared Infrastructure
Robert McDonald - HathiTrust Research Center Executive committee member; Associate Dean for Library Technologies, Indiana University
This document summarizes a paper presented at the IST-Africa 2011 conference that introduces an approach called the "4th way to SDI building" and the concept of "Geoportal4everybody". The 4th way combines standardization efforts with commercial initiatives and support from voluntary communities. It aims to make spatial data infrastructure more accessible to people. Geoportal4everybody is a solution based on open source software that integrates spatial and non-spatial information using standards, and enables communication through social networks. The paper discusses background on previous related concepts of Geohosting and Uniform Resource Management, and proposes a "spider net" paradigm rather than a pyramid paradigm for building a global SDI.
Towards Semantic APIs for Research Data Services (Invited Talk)Anna Fensel
Rapid development of Internet and Web technology is changing the state of the art in communication of knowledge, or results of research activities. Particularly, Semantic technology, linked and open data become key enablers for successful and efficient progress in research. At first, I define the research data service (RDS) and discuss typical current and possible future usage scenarios involving RDS. Further, I discuss the state of the art in the areas of semantic service and data annotation and API construction, as well as infrastructural solutions, applicable for RDS realisation. At last, innovative methods of online dissemination, promotion and efficient communication of research are discussed.
The data!
Services from Data via Smart City API
IOT Applications and IOT
Personal Data vs Open
Big Data Analytics
App as data collection and User Engagement
Social Media Analysis
Visual Analytics and Dashboards
The Living Lab Approach
How can a cultural institution provide and spread information about itself and its assets? Just a web site is not enough. Semantic Technologies to representing cultural heritage data.
Graph Databases Lifecycle Methodology and Tool to Support Index/Store Versio...Paolo Nesi
Abstract— Graph databases are taking place in many different applications: smart city, smart cloud, smart education, etc. In most cases, the applications imply the creation of ontologies and the integration of a large set of knowledge to build a knowledge base as an RDF KB store, with ontologies, static data, historical data and real time data. Most of the RDF stores are endowed of inferential engines that materialize some knowledge as triples during indexing or querying. In these cases, deleting concepts may imply the removal and change of many triples, especially if the triples are those modeling the ontological part of the knowledge base, or are referred by many other concepts. For these solutions, the graph database versioning feature is not provided at level of the RDF stores tool, and it is quite complex and time consuming to be addressed as black box approach. In most cases the indexing is a time consuming process, and the rebuilding of the KB may imply manually edited long scripts that are error prone. Therefore, in order to solve these kinds of problems, this paper proposes a lifecycle methodology and a tool supporting versioning of indexes for RDF KB store. The solution proposed has been developed on the basis of a number of knowledge oriented projects as Sii-Mobility (smart city), RESOLUTE (smart city risk assessment), ICARO (smart cloud). Results are reported in terms of time saving and reliability.
Keywords — RDF Knowledge base versioning, graph stores versioning, RDF store management, knowledge base life cycle.
Tut mathematics and hypermedia research seminar 2011 11-11Yleisradio
The document discusses visualization and analysis of social media networks. It begins by defining information visualization and social network analysis. It then explains how social media data can be gathered from systems through crawling or backend collection. Tools for visualizing the data include Gephi and Gource. Use cases shown include visualizing collaboration networks in academic courses and events like data journalism workshops. The document concludes that visualizations can reveal hidden patterns and recommends more dynamic, user-oriented visualizations.
B2: Open Up: Open Data in the Public SectorMarieke Guy
Parallel session [B2: Open Up: Open Data in the Public Sector] run at the Institutional Web Management Workshop 2013 (IWMW 2013) event, University of Bath on 26 - 28th June 2013.
DISIT Lab overview: smart city, big data, semantic computing, cloudPaolo Nesi
Smart City
• Projects: http://www.disit.org/5501
– Sii-Mobility, http://www.sii-mobility.org
– Service Map: http://servicemap.disit.org
– Social Innovation: Coll@bora http://www.disit.org/5479
– Navigation Indoor/outdoor: Mobile Emergency http://www.disit.org/5404
– Mobility and Transport: TRACE-IT, RAISSS, TESYSRAIL
• Tools: http://www.disit.org/5489
– Data gathering, data mining and reconciliation
– Data reasoning, deduction, prediction
– Smart city ontology and reasoning tools
– Service analysis and recommendations
– Autonomous train operator, train signaling
– Risk analysis, decision support systems
– Mobile Applications
Data Analytics - Big data
• Projects: http://www.disit.org/5501
– Linked Open Graph: http://LOG.disit.org
– Sii-Mobility, http://www.sii-mobility.org
– Service on a number of projects
• Tools: http://www.disit.org/5489
– Open data and Linked Open Data
– LOG LOD service and tools
– Data mining and reconciliation
– Data reasoning, deduction, prediction, decision support
– SN Analysis and recommendations
– User behavior monitoring and analysis
Smart Cloud - Computing
• Projects: http://www.disit.org/5501
– ICARO: http://www.disit.org/5482
– Cloud ontology: http://www.disit.org/5604
– Cloud simulator:
– Smart Cloud: http://www.disit.org/6544
• Tools: http://www.disit.org/5489
– Cloud Monitoring
– Smart Cloud Engine and reasoner,
– Service Level Analyzer and control
– Configuration analysis and checker
– Cloud Simulation
Text and Web Mining
• Projects: http://www.disit.org/5501
– OSIM: http://www.disit.org/5482
– SACVAR: http://www.disit.org/5604
– Blog/Twitter Vigilance
• Tools: http://www.disit.org/5489
– Text and web mining, Natural Language Processing
– Service localization
– Web Crawling
– Competence analysis
– Blog Vigiliance, sentiment analysis
Social Media and e-Learning
• Projects: http://www.disit.org/5501
– ECLAP, http://www.eclap.eu
– ApreToscana: http://www.apretoscana.org
– Others: AXMEDIS, VARIAZIONI, SMNET, etc.
– Samsung Smart TV: http://www.disit.org/6534
• Tools: http://www.disit.org/5489
– XLMS, Cross Media Learning System
– IPR and content protection and distribution
– Mobile and SmartTv Applications
– Suggestions and recommendations
– Matchmaking solutions
– Media Tools for cross media content
Mobile Computing
• Projects:
– ECLAP: http://www.eclap.eu
– Mobile Medicine: http://mobmed.axmedis.org
– Mobile Emergency: http://www.disit.org/5500
– Smart City, FODD 2015: http://www.disit.org/6593
– Resolute: Mobiles as sensors
• Tools and support:
– Content distribution: e-learning
– Integrated Indoor/outdoor navigation
– User networking and collaboration
– Service localization
– Smart city and services
– OS: iOS, Android, Windows Phone, etc.
– Tech: IOT, iBeacoms, NFC, QR, ….
Hybrid Publishing Lab - Scholarly Communication in the Digital AgeChristian Heise
The Hybrid Publishing Lab analyzes and implements new concepts and technologies for digital scholarly publishing. It develops open source software and business models through projects like the Hybrid Publishing Consortium, which aims to provide affordable publishing solutions for small cultural and academic publishers. The lab also explores new forms of scholarly communication and knowledge dissemination through projects like HyperImage, which allows researchers to annotate and link details within and across images. The goal is to make academic knowledge more openly accessible and support innovative approaches to publishing and research communication in the digital age.
This document provides an introduction to the Semantic Web and Linked Open Data. It discusses how standards like RDF, XML, and OWL allow machines to better understand the meaning of data on the web. It describes how ontologies provide a vocabulary to define relationships between resources. The document outlines the benefits of publishing data as Linked Open Data using these standards, including making data more interoperable and accessible to both humans and machines. Examples are given of biomedical research projects that use Semantic Web technologies to integrate and link different types of data.
HybridDocs - A Digital Learning Environment based on FlashCardsChristian Heise
This document discusses the development of HybridDocs, a software framework and tool for converting raw learning materials like lecture manuscripts and books into structured datasets and digital flashcards. It aims to address the impact of digital learning environments on education and determine how analog materials can be converted to hybrid formats. The project involved workshops with experts, developing a theoretical framework based on flashcard pedagogy, and building an open source prototype. Going forward, the plan is to finalize HybridDocs as a web platform for working with open educational resources and materials, launch it as an open service, and get feedback through workshops with schools and students.
Presentation at the Open Knowledge Festival: Open Research and Education Stream, 20 September 2012, Helsinki; also
Presentation at the DINI-Jahrestagung - Bausteine für Open Science, 24 September 2012, Karlsruhe;
also Belgian Open Access Week: Open Access to Excellence in Research, 22 October 2012, Brussels.
From DARPA to Shakespeare: All the Data we Can Handle Kimberly Hoffman
This document discusses the opportunities and challenges of big data for libraries, researchers, and digital humanities. It notes that big data is growing exponentially from sensors, internet data, and scientific instruments. Libraries and librarians have new roles to play in data management, curation, and research data services. Researchers need help with data literacy, data management plans, and archiving research data. Digital humanities can use big data and visualization to gain new insights. Standards like TEI and services like data repositories are important to enable access and reuse of data.
Digital humanities involves the intersection of digital technologies and humanities research. It can include building digital collections and tools for authoring, analyzing, and managing research. Digital humanities centers typically offer resources like databases, tools for analysis, and training. They serve as hubs for innovation and experimentation in applying new technologies to answer humanities questions. Debates include whether digital humanities should apply technologies or critically examine their impact, and whether databases can support narrative scholarship. Visualizations are increasingly important in digital humanities for exploring subjects like ancient cities in new ways.
The document provides an overview of Thorhildur Jetzek's background and career. It summarizes her educational qualifications including a Ph.D. in Information Technology Management from CBS in 2015. It also lists some of her past roles working as an economist, IT consultant, and in various positions at CBS where she is currently a postdoctoral fellow. The document then discusses CBS' ranking and focus on industry collaboration through projects like industrial Ph.D. programs and crowdsourcing competitions for students.
Social Media Metrics for the Cultural Heritage sectorHU-Crossmedialab
1. The document discusses the development of a prototype social media monitor to provide Dutch museums better insight into the effects of their social media usage.
2. The monitor collects publicly available data from Facebook, Twitter, and Flickr for Dutch museums registered in the Netherlands Museum Register.
3. Developing their own custom monitor allows the researchers to experiment and customize the tool to better understand social media metrics for the cultural heritage sector, though it is acknowledged the monitor is only a prototype.
Big data is arising from multiple sources at high volume, velocity, and variety. Pull technology relies on users requesting information, as seen on early search engines and websites, while push technology proactively sends updates without requests. As big data grows, push technology faces challenges in speed and completeness of real-time updates for users, while pull relies on users to initiate longer downloads of large files. Both will continue evolving to optimize access and delivery of massive datasets.
Ontology Building vs Data Harvesting and Cleaning for Smart-city ServicesPaolo Nesi
Presently, a very large number of public and private data sets are available around the local governments. In most cases, they are not semantically interoperable and a huge human effort is needed to create integrated ontologies and knowledge base for smart city. Smart City ontology is not yet standardized, and a lot of research work is needed to identify models that can easily support the data reconciliation, the management of the complexity and reasoning. In this paper, a system for data ingestion and reconciliation of smart cities related aspects as road graph, services available on the roads, traffic sensors etc., is proposed. The system allows managing a big volume of data coming from a variety of sources considering both static and dynamic data. These data are mapped to smart-city ontology and stored into an RDF-Store where they are available for applications via SPARQL queries to provide new services to the users. The paper presents the process adopted to produce the ontology and the knowledge base and the mechanisms adopted for the verification, reconciliation and validation. Some examples about the possible usage of the coherent knowledge base produced are also offered and are accessible from the RDF-Store and related services. The article also presented the work performed about reconciliation algorithms and their comparative assessment and selection. Keywords Smart city, knowledge base construction, reconciliation, validation and verification of knowledge base, smart city ontology, linked open graph.
This document summarizes a presentation on data visualization. It introduces data visualization and its uses for exploring data, explaining results, and distant reading. It discusses the building blocks of visualization like charts, networks, and visualizing different data types. It explores some scholarly visualizations and exercises critiquing them. It also covers extracting data from text, images and video using computational methods, and preparing messy humanities data for visualization, including dealing with uncertainty. The presentation emphasizes choosing visualizations based on purpose, data, audience and structure. It recommends tools for creating simple visualizations like Viewshare that don't require programming.
Introduction to information visualisation for humanities PhDsMia
Training workshop for the CHASE Arts and Humanities in the Digital Age programme. (
This session will give you an overview of a variety of techniques and tools available for data visualisation and analysis in the humanities. You will learn about common types of visualisations and the role of exploratory and explanatory visualisations, explore examples of scholarly visualisations, try some visualisation tools, and know where to find further information about analysing and building data visualisations.
The HathiTrust Research Center: Enabling New Knowledge Through Shared Infrastructure
Robert McDonald - HathiTrust Research Center Executive committee member; Associate Dean for Library Technologies, Indiana University
This document summarizes a paper presented at the IST-Africa 2011 conference that introduces an approach called the "4th way to SDI building" and the concept of "Geoportal4everybody". The 4th way combines standardization efforts with commercial initiatives and support from voluntary communities. It aims to make spatial data infrastructure more accessible to people. Geoportal4everybody is a solution based on open source software that integrates spatial and non-spatial information using standards, and enables communication through social networks. The paper discusses background on previous related concepts of Geohosting and Uniform Resource Management, and proposes a "spider net" paradigm rather than a pyramid paradigm for building a global SDI.
Towards Semantic APIs for Research Data Services (Invited Talk)Anna Fensel
Rapid development of Internet and Web technology is changing the state of the art in communication of knowledge, or results of research activities. Particularly, Semantic technology, linked and open data become key enablers for successful and efficient progress in research. At first, I define the research data service (RDS) and discuss typical current and possible future usage scenarios involving RDS. Further, I discuss the state of the art in the areas of semantic service and data annotation and API construction, as well as infrastructural solutions, applicable for RDS realisation. At last, innovative methods of online dissemination, promotion and efficient communication of research are discussed.
The data!
Services from Data via Smart City API
IOT Applications and IOT
Personal Data vs Open
Big Data Analytics
App as data collection and User Engagement
Social Media Analysis
Visual Analytics and Dashboards
The Living Lab Approach
How can a cultural institution provide and spread information about itself and its assets? Just a web site is not enough. Semantic Technologies to representing cultural heritage data.
Graph Databases Lifecycle Methodology and Tool to Support Index/Store Versio...Paolo Nesi
Abstract— Graph databases are taking place in many different applications: smart city, smart cloud, smart education, etc. In most cases, the applications imply the creation of ontologies and the integration of a large set of knowledge to build a knowledge base as an RDF KB store, with ontologies, static data, historical data and real time data. Most of the RDF stores are endowed of inferential engines that materialize some knowledge as triples during indexing or querying. In these cases, deleting concepts may imply the removal and change of many triples, especially if the triples are those modeling the ontological part of the knowledge base, or are referred by many other concepts. For these solutions, the graph database versioning feature is not provided at level of the RDF stores tool, and it is quite complex and time consuming to be addressed as black box approach. In most cases the indexing is a time consuming process, and the rebuilding of the KB may imply manually edited long scripts that are error prone. Therefore, in order to solve these kinds of problems, this paper proposes a lifecycle methodology and a tool supporting versioning of indexes for RDF KB store. The solution proposed has been developed on the basis of a number of knowledge oriented projects as Sii-Mobility (smart city), RESOLUTE (smart city risk assessment), ICARO (smart cloud). Results are reported in terms of time saving and reliability.
Keywords — RDF Knowledge base versioning, graph stores versioning, RDF store management, knowledge base life cycle.
Tut mathematics and hypermedia research seminar 2011 11-11Yleisradio
The document discusses visualization and analysis of social media networks. It begins by defining information visualization and social network analysis. It then explains how social media data can be gathered from systems through crawling or backend collection. Tools for visualizing the data include Gephi and Gource. Use cases shown include visualizing collaboration networks in academic courses and events like data journalism workshops. The document concludes that visualizations can reveal hidden patterns and recommends more dynamic, user-oriented visualizations.
B2: Open Up: Open Data in the Public SectorMarieke Guy
Parallel session [B2: Open Up: Open Data in the Public Sector] run at the Institutional Web Management Workshop 2013 (IWMW 2013) event, University of Bath on 26 - 28th June 2013.
DISIT Lab overview: smart city, big data, semantic computing, cloudPaolo Nesi
Smart City
• Projects: http://www.disit.org/5501
– Sii-Mobility, http://www.sii-mobility.org
– Service Map: http://servicemap.disit.org
– Social Innovation: Coll@bora http://www.disit.org/5479
– Navigation Indoor/outdoor: Mobile Emergency http://www.disit.org/5404
– Mobility and Transport: TRACE-IT, RAISSS, TESYSRAIL
• Tools: http://www.disit.org/5489
– Data gathering, data mining and reconciliation
– Data reasoning, deduction, prediction
– Smart city ontology and reasoning tools
– Service analysis and recommendations
– Autonomous train operator, train signaling
– Risk analysis, decision support systems
– Mobile Applications
Data Analytics - Big data
• Projects: http://www.disit.org/5501
– Linked Open Graph: http://LOG.disit.org
– Sii-Mobility, http://www.sii-mobility.org
– Service on a number of projects
• Tools: http://www.disit.org/5489
– Open data and Linked Open Data
– LOG LOD service and tools
– Data mining and reconciliation
– Data reasoning, deduction, prediction, decision support
– SN Analysis and recommendations
– User behavior monitoring and analysis
Smart Cloud - Computing
• Projects: http://www.disit.org/5501
– ICARO: http://www.disit.org/5482
– Cloud ontology: http://www.disit.org/5604
– Cloud simulator:
– Smart Cloud: http://www.disit.org/6544
• Tools: http://www.disit.org/5489
– Cloud Monitoring
– Smart Cloud Engine and reasoner,
– Service Level Analyzer and control
– Configuration analysis and checker
– Cloud Simulation
Text and Web Mining
• Projects: http://www.disit.org/5501
– OSIM: http://www.disit.org/5482
– SACVAR: http://www.disit.org/5604
– Blog/Twitter Vigilance
• Tools: http://www.disit.org/5489
– Text and web mining, Natural Language Processing
– Service localization
– Web Crawling
– Competence analysis
– Blog Vigiliance, sentiment analysis
Social Media and e-Learning
• Projects: http://www.disit.org/5501
– ECLAP, http://www.eclap.eu
– ApreToscana: http://www.apretoscana.org
– Others: AXMEDIS, VARIAZIONI, SMNET, etc.
– Samsung Smart TV: http://www.disit.org/6534
• Tools: http://www.disit.org/5489
– XLMS, Cross Media Learning System
– IPR and content protection and distribution
– Mobile and SmartTv Applications
– Suggestions and recommendations
– Matchmaking solutions
– Media Tools for cross media content
Mobile Computing
• Projects:
– ECLAP: http://www.eclap.eu
– Mobile Medicine: http://mobmed.axmedis.org
– Mobile Emergency: http://www.disit.org/5500
– Smart City, FODD 2015: http://www.disit.org/6593
– Resolute: Mobiles as sensors
• Tools and support:
– Content distribution: e-learning
– Integrated Indoor/outdoor navigation
– User networking and collaboration
– Service localization
– Smart city and services
– OS: iOS, Android, Windows Phone, etc.
– Tech: IOT, iBeacoms, NFC, QR, ….
Hybrid Publishing Lab - Scholarly Communication in the Digital AgeChristian Heise
The Hybrid Publishing Lab analyzes and implements new concepts and technologies for digital scholarly publishing. It develops open source software and business models through projects like the Hybrid Publishing Consortium, which aims to provide affordable publishing solutions for small cultural and academic publishers. The lab also explores new forms of scholarly communication and knowledge dissemination through projects like HyperImage, which allows researchers to annotate and link details within and across images. The goal is to make academic knowledge more openly accessible and support innovative approaches to publishing and research communication in the digital age.
This document provides an introduction to the Semantic Web and Linked Open Data. It discusses how standards like RDF, XML, and OWL allow machines to better understand the meaning of data on the web. It describes how ontologies provide a vocabulary to define relationships between resources. The document outlines the benefits of publishing data as Linked Open Data using these standards, including making data more interoperable and accessible to both humans and machines. Examples are given of biomedical research projects that use Semantic Web technologies to integrate and link different types of data.
HybridDocs - A Digital Learning Environment based on FlashCardsChristian Heise
This document discusses the development of HybridDocs, a software framework and tool for converting raw learning materials like lecture manuscripts and books into structured datasets and digital flashcards. It aims to address the impact of digital learning environments on education and determine how analog materials can be converted to hybrid formats. The project involved workshops with experts, developing a theoretical framework based on flashcard pedagogy, and building an open source prototype. Going forward, the plan is to finalize HybridDocs as a web platform for working with open educational resources and materials, launch it as an open service, and get feedback through workshops with schools and students.
Presentation at the Open Knowledge Festival: Open Research and Education Stream, 20 September 2012, Helsinki; also
Presentation at the DINI-Jahrestagung - Bausteine für Open Science, 24 September 2012, Karlsruhe;
also Belgian Open Access Week: Open Access to Excellence in Research, 22 October 2012, Brussels.
From DARPA to Shakespeare: All the Data we Can Handle Kimberly Hoffman
This document discusses the opportunities and challenges of big data for libraries, researchers, and digital humanities. It notes that big data is growing exponentially from sensors, internet data, and scientific instruments. Libraries and librarians have new roles to play in data management, curation, and research data services. Researchers need help with data literacy, data management plans, and archiving research data. Digital humanities can use big data and visualization to gain new insights. Standards like TEI and services like data repositories are important to enable access and reuse of data.
Digital humanities involves the intersection of digital technologies and humanities research. It can include building digital collections and tools for authoring, analyzing, and managing research. Digital humanities centers typically offer resources like databases, tools for analysis, and training. They serve as hubs for innovation and experimentation in applying new technologies to answer humanities questions. Debates include whether digital humanities should apply technologies or critically examine their impact, and whether databases can support narrative scholarship. Visualizations are increasingly important in digital humanities for exploring subjects like ancient cities in new ways.
The document provides an overview of Thorhildur Jetzek's background and career. It summarizes her educational qualifications including a Ph.D. in Information Technology Management from CBS in 2015. It also lists some of her past roles working as an economist, IT consultant, and in various positions at CBS where she is currently a postdoctoral fellow. The document then discusses CBS' ranking and focus on industry collaboration through projects like industrial Ph.D. programs and crowdsourcing competitions for students.
Social Media Metrics for the Cultural Heritage sectorHU-Crossmedialab
1. The document discusses the development of a prototype social media monitor to provide Dutch museums better insight into the effects of their social media usage.
2. The monitor collects publicly available data from Facebook, Twitter, and Flickr for Dutch museums registered in the Netherlands Museum Register.
3. Developing their own custom monitor allows the researchers to experiment and customize the tool to better understand social media metrics for the cultural heritage sector, though it is acknowledged the monitor is only a prototype.
Open Science in Research Libraries: Research, Research Integrity and Legal As...Marlon Domingus
Session on RDM and legal aspects at the Erasmus+ Staff Mobility - Knowledge Sharing Open Science in Research Libraries. June 12-16 2017. TU Delft and Erasmus University Rotterdam.
Towards long-term preservation of linked data - the PRELIDA projectPRELIDA Project
This document summarizes a presentation about preserving linked data over the long term. It introduces the PRELIDA project, which aims to bridge the digital preservation and linked data communities. The presentation discusses what digital preservation can provide for linked data, such as file format standards, archival storage services, and documentation practices. It also outlines challenges for preserving linked data, like its dynamic and distributed nature. The PRELIDA project seeks to address these challenges through research and bringing the communities together.
Einstein published his ideas and became a pivotal element in shifting the way we think about physics - from the Newtonian model to the Quantum - in turn this changed the way we think about the world and allowed us to develop new ways of engaging with the world.
We are at a similar juncture. The development of computational technologies allows us to think about astronomical volumes of data and to make meaning of that data.
The mindshift that occurs is that “the machine is our friend”. The computer, like all machines, extends our capabilities. As a consequence the types of thinking now required in industry are those that get away from thinking like a computer and shift towards creative engagement with possibilities. Logical thinking is still necessary but it starts to be driven by imagination.
Computational thinking and data science change the way we think about defining and solving problems.
The age of creativity - which increasingly extends its impact from arts applications to business, scientific, technological, entrepreneurship, political, and other contexts.
The document discusses the European Commission's role in supporting open digital science through funding research and building infrastructures. It outlines several key Commission initiatives related to open access publications and data, and highlights upcoming work programs that aim to develop e-infrastructures, manage and compute on big research data, and support open access. Citizen science and alternative metrics for evaluating research are also mentioned as important aspects of advancing open digital science.
Talk at the World Science Festival at Columbia, June 2, 2017: session on Big Data and Physics: http://www.worldsciencefestival.com/programs/big-data-future-physics/
Drowning in information – the need of macroscopes for research fundingAndrea Scharnhorst
Andrea Scharnhorst (2015) Drowning in information – the need of macroscopes for research funding. Presentation at the international conference: PLANNING, PREDICTION, SCENARIOS - Using Simulations and Maps - 2015 Annual EA Conference - 11–12 May 2015 Bonn
The document discusses the European Commission's role and initiatives related to open, digital science. The EC acts as a policy maker, funding agency, and infrastructure builder to promote open access to publications and data from publicly funded research. Major EC actions include the Horizon 2020 program, which allocates €80 billion to digital science and open access goals from 2014-2020. The document also outlines initiatives to improve research tools and infrastructures, data sharing, citizen science, and metrics for evaluating open data and science.
Slides for presentation given at the first Digital Humanities Congress held in Sheffield from 6 – 8 September 2012 with the support of the Network of Expert Centres and Centernet.
URL http://www.shef.ac.uk/hri/dhc2012
Presentation given to EC project officers as part of workshops run by the FOSTER (foster open science) project. The presentation covers the Horizon 2020 open data pilot.
This document provides an overview of the MOVING project, which aims to build an innovative training platform to improve information literacy. The platform will provide access to data sources, search and visualization methods, and individually tailored training programs to enable open leadership innovation. A consortium led by CERTH-ITI is developing the platform over 3 years with H2020 funding. Two use cases of the platform involve assisting public administrators with business research and helping researchers manage and mine research information. The platform will include modules for data acquisition, processing, video analysis, lecture linking, and demonstrations.
This document discusses managing research data for open science based on the UK experience. It outlines key aspects of open science such as making research more open, global, collaborative and closer to society. The document discusses mandates for open research data from funding bodies in the UK and EU, including stipulations in Horizon 2020 and requirements from EPSRC. It defines what constitutes research data and examines challenges around research data management, including technology issues, people issues, policy issues and resources. The importance of data skills training for researchers and data professionals is also covered.
Efficient Use of Internet and Social Media Tools in Innovation ProcessesMikko Ahonen
1. The document discusses efficient use of internet and social media tools in innovation processes, including concepts like open innovation, crowdsourcing, and social network analysis.
2. It provides examples of innovation intermediaries and how communities and social media like blogs can be used to capture ideas and get user feedback.
3. Social network analysis is discussed as a way to map knowledge flows and connections between people in an organization to enhance innovative thinking.
Similar to Research Data management - Importance, Good Practices, Guidance (20)
Presentation by Rebecca Sachs and Joshua Varcie, analysts in CBO’s Health Analysis Division, at the 13th Annual Conference of the American Society of Health Economists.
The Power of Community Newsletters: A Case Study from Wolverton and Greenleys...Scribe
YOU WILL DISCOVER:
The engaging history and evolution of Wolverton and Greenleys Town Council's newsletter
Strategies for producing a successful community newsletter and generating income through advertising
The decision-making process behind moving newsletter design from in-house to outsourcing and its impacts
Dive into the success story of Wolverton and Greenleys Town Council's newsletter in this insightful webinar. Hear from Mandy Shipp and Jemma English about the newsletter's journey from its inception to becoming a vital part of their community's communication, including its history, production process, and revenue generation through advertising. Discover the reasons behind outsourcing its design and the benefits this brought. Ideal for anyone involved in community engagement or interested in starting their own newsletter.
Jennifer Schaus and Associates hosts a complimentary webinar series on The FAR in 2024. Join the webinars on Wednesdays and Fridays at noon, eastern.
Recordings are on YouTube and the company website.
https://www.youtube.com/@jenniferschaus/videos
Causes Supporting Charity for Elderly PeopleSERUDS INDIA
Around 52% of the elder populations in India are living in poverty and poor health problems. In this technological world, they became very backward without having any knowledge about technology. So they’re dependent on working hard for their daily earnings, they’re physically very weak. Thus charity organizations are made to help and raise them and also to give them hope to live.
Donate Us:
https://serudsindia.org/supporting-charity-for-elderly-people-india/
#oldagehome, #donateforeldersinkurnool, #donateforelders, #donationforelders, #donateforoldpeople, #donationforoldpeople, #sponsorforelders, #sponsorforoldpeople, #donationforcharity, #charity, #seruds, #kurnool, #donateforoldagehome, #oldagehomedonation
Research Data management - Importance, Good Practices, Guidance
1. Research Data Management
Graphic: University of Portsmouth Research Data web page
url: https://researchandinnovationportsmouth.com/2018/05/18/research-data-management-resources-uop/
Importance, Good practices, Guidance
By:
Frank Uiterwaal
f.uiterwaal@niod.knaw.nl
Twitter: @FrankUiterwaal
2. Introduction Frank Uiterwaal
NIOD’s area of work covers the 20th and 21st century, with a focus on research into the effects of wars,
the Holocaust and other genocides on individuals and society. NIOD:
• …collects, manages, opens up and makes accessible archives and collections about the Second
World War;
• …conducts academic research and publishes about it;
• …gives information to government bodies and individuals;
• …stimulates and organises debates and activities about war violence and processes that are at the
basis of war violence;
• …and coordinates the European Holocaust Research Infrastructure (EHRI).
Digital technologies have led to the creation of large digital archives, and to a variety of innovative
research methodologies applicable to these archives. PARTHENOS is eager to integrate these archives
and new methods to support digital research.
By working together, PARTHENOS…:
• …develops common standards to ease exploitation;
• …coordinates joint activities among research projects;
• …harmonises policy definition and implementation;
• …pools methods and services;
• …and shares solutions to the same problems.
3. Research Data Managament as a shared challenge
The PARTHENOS Training Suite as a shared solution
Graphic: Header PARTHENOS Training Suite
url: https://training.parthenos-project.eu/
4.
5. Graphic: PARTHENOS Training Suite: Manage, Improve And Open Up Your Research Data – Open Data, Open Access and Open Science
url: https://training.parthenos-project.eu/sample-page/manage-improve-and-open-up-your-research-and-data/open-data-open-access-and-open-science/
6.
7. Disclaimer 1
Research Data Management ≠ Digital Humanities
Data are everywhere!
In the ‘analogue’ Humanities…
primary sources, secondary sources, theoretical texts, notes, annotations,
references.
…and in the Digital Humanities (additional to the above)
Digital tools, other forms of code, interpretative data on top of primary sources
(semantically enriched text, statistics derived from data mining or natural language
processing, machine learning data (e.g. to enhance text or speech recognition), GIS
data, data visualisations…).
8. What is Research Data Management?
A working definition
“Research Data Management (RDM) concerns the organisation of data, from the start
when data are collected through to the dissemination and archiving of valuable results. It
aims to ensure reliable verification of results, and permits new and innovative research
built on existing information.
Research data management is part of the research process, aims to make the research
process as efficient as possible and meet expectations and requirements of the university
(…), research funders, and legislation.”
- TU Delft
Source: https://www.tudelft.nl/en/library/current-topics/research-data-
management/research-data-management/why-data-management/
9. Disclaimer 2
Research Data Management is not necessarily fun…
…but it is necessary!
Yes…:
…it takes time away from actual research you want to do;
…instead, it is necessary to take a step back and reflect on the things you are
producing.
In short: boring chore of scholarly housekeeping and sometimes it might feel like it’s
slowing you down.
10. Why bother?
The benefits of good Research Data Management
• It guarantees research integrity and replication;
• It ensures that research data are authentic, complete, and reliable;
• It minimises the risk of losing your data;
• It increase research efficiency;
• It prevents duplication of effort by enabling others to use your data;
• It helps you to meet funding agency requirements.
- TU Delft
Source: https://www.tudelft.nl/en/library/current-topics/research-data-
management/research-data-management/why-data-management/
12. Graphic 2: The Guardian – “Why did the Cologne city archive collapse?”
url: https://www.theguardian.com/world/2009/mar/27/germany-cologne
Graphic 1: ITV – “Major fire at the university of Nottingham”
url: https://www.itv.com/news/central/story/2014-09-12/major-fire-at-university-of-nottingham//
“Oh… I’ll get to that later…”
“It’s safely stored on my pc / on my desk / in my head”
13. Graphic: Suzette Lohmeyer - “Link rot: What happens when the internet isn’t forever”
url: https://gcn.com/articles/2016/07/27/link-rot.aspx
Or more plausible…
14. Data Management Planning
Required by big funding bodies:
- European Commission;
- NWO - Netherlands Organisation for Scientific Research.
“Data Section (E.C.)” / “Data Paragraph (NWO)” in proposal
Data management plan early in execution phase
15. Data Paragraph in Horizon 2020 proposal
“(…) all project proposals must include a section on research data management
which is evaluated under the criterion 'Impact'. Applicants must provide a short,
general outline of their policy for data management in which they answer the
following questions:
• What types of data will the project generate/collect?
• What standards will be used?
• How will this data be exploited and/or shared/made accessible for verification and
re-use? If data cannot be made available, explain why.
• How will this data be curated and preserved?”
TU Delft Library - “Guidelines for a data paragraph in a H2020 project proposal”
Source: https://www.tudelft.nl/en/library/current-topics/research-data-management/research-data-
management/setting-up-research/data-paragraphs-and-data-management-plans/
16. Data Paragraph in NWO proposal
4 central questions:
1. Will data be collected or generated that are suitable for reuse?
2. Where will the data be stored during the research?
3. After the project has been completed, how will the data be stored for the long-term and
made available for the use by third parties? To whom will the data be accessible?
4. Which facilities (ICT, (secure) archive, refrigerators or legal expertise) do you expect will
be needed for the storage of data during the research and after the research? Are these
available?
…and think about funding that!
TU Delft Library - “Guidelines for a data paragraph in a NWO project proposal”
Source: https://www.tudelft.nl/en/library/current-topics/research-data-management/research-data-
management/setting-up-research/data-paragraphs-and-data-management-plans/
17. From importance to good practices
The FAIR principles
FAIR Data Principles are drafted by a wide collaboration (pan-
disciplinary organisation, not one specific to arts and
humanities) including academia, industry, funding agencies,
and scholarly publishers - FORCE11.
Data needs to be:
• Findable;
• Accessible;
• Interoperable;
• Re-Usable.
Graphic: St. Lawrence Global Observatory – “FAIR data”
url: https://ogsl.ca/en/fair-principles
18. Advice per letter (F, A, I and R)…
…and for additional help, check the
PARTHENOS Guidelines to FAIRify data
Management and make data reusable!
- Gathered over 100 data mgmt policies…
- …by 50 PARTHENOS project members.
Graphic: PARTHENOS Guidelines to FAIRify data management and make data reusable
url: http://www.parthenos-project.eu/portal/policies_guidelines
From importance to good practices
“So… how can I make my data FAIR?”
19. Store them in a place where they can be found..
…but more on that later.
Graphic: Texas Digital Library
url: https://www.tdl.org/2016/04/unt-libraries-trac-selfaudit/
“How can I make my data FINDABLE?”
Ways to make sure your work gets discovered
20. “What happens when the internet isn’t forever?”
It isn’t…
…but a Persistent Identifier helps!
(stable referrent)
“How can I make my data ACCESSIBLE?”
Ways to make sure that people can access your data
Graphic: Suzette Lohmeyer - “Link rot: What happens when the internet isn’t forever”
url: https://gcn.com/articles/2016/07/27/link-rot.aspx
21.
22. “How can I make my data ACCESSIBLE?”
Open Access
Video: SHB Online “What is Open Access?”
url: https://www.youtube.com/watch?v=gzRgknylTEM
23.
24. But.. sharing of data within limits
“as open as possible, as closed as necessary”
“But shouldn’t we strive for open access?”
Sometimes restrictions are necessary:
- Personal data (GDPR) consider anonimisation;
- Portrait right;
Or you can make a conscious decision not to share some data:
- Embargo;
- Not relevant for re-use.
Graphic: “GDPR: All You Need to Know to Be Compliant!”
url: https://codeburst.io/gdpr-all-you-need-to-know-to-be-compliant-5f377dbff68a
25. “How can I make my data INTEROPERABLE?”
Making sure that your data works elsewhere
Graphic: Vector - Different type power socket set, electric outlet illustration for different country plugs. Vector illustration world standards icons set
url: https://www.123rf.com/photo_77571970_stock-vector-different-type-power-socket-set-electric-outlet-illustration-for-different-country-plugs-vector-illu.html
26. Example:
Preferred file formats
Preference for:
Future-proof, open source,
platform independent
software.
“How can I make my data INTEROPERABLE?”
Making sure that your data works elsewhere
Graphic: University Libraries – University of Washington – Preferred File Formats
url: https://www.lib.washington.edu/preservation/preservation_services/digitization-and-digital-preservation/preferred-file-formats
28. Please keep your data clean!
Carefully think about:
- A naming convention;
- A logical folder structure;
And preferably start early.
Graphic: “Data Cleansing and Decision making Quality”
url: https://www.promptcloud.com/blog/data-cleansing-decision-making-quality/
“How can I make my data RE-USABLE?”
Ways to make sure your work makes sense to others
29. Graphic: Cursus Leren Preserveren - Archiefordening
url: https://lerenpreserveren.nl/topic/archiefordening/
30. Graphic: Cursus Leren Preserveren - Archiefordening
url: https://lerenpreserveren.nl/topic/archiefordening/
31. Befriend the data stewards in your organisation, because:
- Not everyone has the time to learn about best practices in data management;
- Not everyone hosts a trusted repository in their garden shed.
The NIOD also has a partner for
its research data:
DANS – Data Archiving and
Networked Services.
Graphic: DANS homepage
url: https://dans.knaw.nl/en/front-page?set_language=en
…and back to FINDABLE!
Support is needed
32.
33.
34.
35. Thank you for your attention!
Graphic: University of Portsmouth Research Data web page
url: https://researchandinnovationportsmouth.com/2018/05/18/research-data-management-resources-uop/
By:
Frank Uiterwaal
f.uiterwaal@niod.knaw.nl
@FrankUiterwaal
Editor's Notes
Leon asked me if I was interested in giving a talk about Research Data Management. Something I’m always happy to do.
So here it is, my presentation which will focus on the importance of solid research data management, good practices that can be used and guidance you can ask for. As a researcher, you might sometimes feel like you’re dragging this pile of data around, while not being entirely sure how and where exactly you should deliver them.
Regarding the latter, guidance, Leon told me that there were a lot of questions from you about how you could best manage your data.
In this presentation we will learn that Research Data Management is a subject, very specific to institutions.
Also, because it is a subject very specific to institutions, I tried to keep the advice both focused on researchers as well as very widely applicable. That means that I took out a lot of the advice we usually give to data curators and repositories, which is more technical or policy-oriented. We scheduled around 45 minutes for this presentation, but because I wanted to keep it researcher focussed it might be somewhat shorter. But we will see!
So, who is Frank Uiterwaal? I work for the NIOD Institute for War, Holocaust and Genocide studies and – within the NIOD – for a European project with the name PARTHENOS.
So what is the NIOD, what does it do? Primarily… three things. It’s an archive, research center, public platform, mainly around war and conflict in the 20th and the 21st century.
The NIOD also has a European project it coordinates itself with the name EHRI (the European Holocaust Project). The aim of which is to organise Holocaust research on a European scale. Strong focus on innovative research methods and, therewith, Digital Humanities. Through EHRI project, it became part of the larger dialogue in Europe about Digital Humanities.
A result of that dialogue is the cluster project PARTHENOS, which includes – apart from EHRI – many more European projects in the (digital) Humanities. At the NIOD I’m primarily involved in this European project called PARTHENOS, through which I know Leon. PARTHENOS gathers and exchanges good practices in Digital Humanities research. Also focus on Research Data Management as a ‘shared challenge’, for which it wants to develop shared solutions for the field
As an example of a shared challenge and a shared solution… Research Data Management
All kinds of recommendations, lessons, suggestions for further reading, multimedia content – like this video about Open Access and Open Data. Created our own material…
…but also have famous movies from the past, like this one: Digital Preservation and Nuclear Disaster.. Featuring Digiman! Nerd-alert.. For Research Data Management enthusiasts…
However, when Leon mentioned that the course was about Digital Humanities, the idea to talk about Research Data Management was something I felt I had to think carefully about, because a misconseption is luring…….
Research Data Management is not the same as Digital Humanities.
But doesn’t Digital Humanities have anything to do with data..? Obviously, yes. In Digital Humanities research, all sorts of data are used, created and curated.
But, data are used, created and curated everywhere. Digital and analogue, machine created or human created, highly structured or utterly unstructured.
The ‘analogue’ Humanities also use and create data. Examples on the slide.
BUT, it is fair to say that Digital Humanities are more data intensive (especially on the data creation side). Additional to the above.. You wll find… [examples on the slide]
Will get to Disclaimer 2 in a little bit, but frst it’s time for a shared definition of what Research Data (and Research Data Management) are.
So what is Research Data Management. TU Delft has the name of being quite far ahead in the field and their definition is – I think – both very applicable as well as all-encompassing.
I always think the: “do it so research funders are happy is a bit circular”
To some of you, it may sound like fun, to MANY of you, it may not.
RDM is not necessarily fun BUT very necessary.
So you might feel like it takes time. You might feel like it sucks you out of the actual research..
…it might feel like a boring chore of scholarly housekeeping..
So why bother?
There are actually quite some reasons to bother…
Cringy motivational picture time. Isaac Newton refers to wat he learned from Descartes and that he built on the way Descartes saw the world, which enabled him to reach great heights. Cliché’s like research is done together and your world is not an island are very true.
(Apologies for the motivational picture.) However, what I hope to have demonstrated by now, is that RDM is very important. Opening up our data could open up many opportunities for using and reusing it, for collaborating, informing and increasing the impact of our work.
And you might think.. Sure… I’ll get to that later. All the information is on my pc / on my desk / in my head…But that mindset is not without it’s risks.
A more plausible risk.. What we call “link rot”. Picture from an article called: “what happens when the internet isn’t forever” In some sense the internet is not forever.
So, Data Management is something you need to plan and structure. Big funding bodies urge you to do so. The European Commission and NWO demand a Data section or Data paragraph in your research proposal which, not much later, you have to detail in a data management plan.
What exactly do they prescribe?
FAIR is a mantra… at the same time
FAIR is a mantra… at the same time
So… how do you get going? Luckily there are frameworks who presents sets of guidlines to propose recommendations.. Now moving from “why is it important?” to “what can you do?” section of the presentation
Strong focus of what you can do as a researcher!
So.. How can you make data findable? This is for a large part determined by where you store your data
So… how do you get going? Luckily there are frameworks who presents sets of guidlines
But the most logical answer to the question “how do you make your data accessible?” if – of course – open access. Your output not behind a paywall, but freely available. Short movie explaining the most important aspects.
So.. A license… where do I get that?
FAIR is a mantra… at the same time: “open as possible, closed as necessary”
Why not open access?
What is interoperability?
Any examples?
Power plug from Type C does not work in the UK
File types – ebooks: Can I use it on all my devices? Or is it proprietary: of, relating to, or characteristic of an owner or title holder proprietary rights
What is interoperability?
Preferably readable by sustainable, open source software
Experience with a file you couldn’t open?
No one downloads movies these days?
An sheet that only works on Macs (numbers)?
Why….
Xlsx medium confidence.. Xls. Lowest confidence?
Excel is well extablished software, everybody uses it.. Why not highest confidence?
So… how do you get going? Luckily there are frameworks who presents sets of guidlines
Quick exercise…
In Dutch… what could be improved?
Guilty of this one… as they say in the Netherlands “At the home of the plumber, the tap is leaking”
Strong focus of what you can do as a researcher!
And plenty of universities (if not all) offer Research Data Management Support.
I hope you know have a reasonable idea of why Research Data Management is important, what you can do yourself as a researcher and what kind of help is available if you need any. You will not see me anytime soon and unfortunately, my time is to limited to find solutions to all problems, but I’m sure the data steward team in your university knows best!