OpenAIRE and EUDAT co-present this webinar which aims to introduce researchers and others to the concept of research data management (RDM). As well as presenting the benefits of taking an active approach to research data management – including increased speed and ease of access, efficiency (fund once, reuse many times), and improved quality and transparency of research – the webinar will advise on strategies for successful RDM, resources to help manage data effectively, choosing where to store and deposit data, the EC H2020 Open Data Pilot and the basics of data management, stewardship and archiving.
Webinar recording available: http://www.instantpresenter.com/eifl/EB57D6888147
EUDAT & OpenAIRE Webinar: How to write a Data Management Plan - July 7, 2016|...EUDAT
| www.eudat.eu | 1st Session: July 7, 2016.
In this webinar, Sarah Jones (DCC) and Marjan Grootveld (DANS) talked through the aspects that Horizon 2020 requires from a DMP. They discussed examples from real DMPs and also touched upon the Software Management Plan, which for some projects can be a sensible addition
The presentation gives an overview of what metadata is and why it is important. It also addresses the benefits that metadata can bring and offers advice and tips on how to produce good quality metadata and, to close, how EUDAT uses metadata in the B2FIND service.
November 2016
Data Lakehouse Symposium | Day 1 | Part 1Databricks
The world of data architecture began with applications. Next came data warehouses. Then text was organized into a data warehouse.
Then one day the world discovered a whole new kind of data that was being generated by organizations. The world found that machines generated data that could be transformed into valuable insights. This was the origin of what is today called the data lakehouse. The evolution of data architecture continues today.
Come listen to industry experts describe this transformation of ordinary data into a data architecture that is invaluable to business. Simply put, organizations that take data architecture seriously are going to be at the forefront of business tomorrow.
This is an educational event.
Several of the authors of the book Building the Data Lakehouse will be presenting at this symposium.
Software's now-a-days became the life line of modern day organizations. Libraries also need software if they want to create a parallel digital library with features which we may not find in a traditional library.
This presentation was given at the FOSE 2012 Records Management (RIM) conference held the week of April 2 and identifies several of the important standards to consider when wanting to preserve digital content. With the advent of 'Big Data', relying on standards will help to ensure our data/information is properly preservered.
Presentation slides from a lecture given at the University of the West of England (UWE) as part of the MSc in Library and Library Management, University of the West of England, Frenchay Campus, Bristol, March 24, 2009
EUDAT & OpenAIRE Webinar: How to write a Data Management Plan - July 7, 2016|...EUDAT
| www.eudat.eu | 1st Session: July 7, 2016.
In this webinar, Sarah Jones (DCC) and Marjan Grootveld (DANS) talked through the aspects that Horizon 2020 requires from a DMP. They discussed examples from real DMPs and also touched upon the Software Management Plan, which for some projects can be a sensible addition
The presentation gives an overview of what metadata is and why it is important. It also addresses the benefits that metadata can bring and offers advice and tips on how to produce good quality metadata and, to close, how EUDAT uses metadata in the B2FIND service.
November 2016
Data Lakehouse Symposium | Day 1 | Part 1Databricks
The world of data architecture began with applications. Next came data warehouses. Then text was organized into a data warehouse.
Then one day the world discovered a whole new kind of data that was being generated by organizations. The world found that machines generated data that could be transformed into valuable insights. This was the origin of what is today called the data lakehouse. The evolution of data architecture continues today.
Come listen to industry experts describe this transformation of ordinary data into a data architecture that is invaluable to business. Simply put, organizations that take data architecture seriously are going to be at the forefront of business tomorrow.
This is an educational event.
Several of the authors of the book Building the Data Lakehouse will be presenting at this symposium.
Software's now-a-days became the life line of modern day organizations. Libraries also need software if they want to create a parallel digital library with features which we may not find in a traditional library.
This presentation was given at the FOSE 2012 Records Management (RIM) conference held the week of April 2 and identifies several of the important standards to consider when wanting to preserve digital content. With the advent of 'Big Data', relying on standards will help to ensure our data/information is properly preservered.
Presentation slides from a lecture given at the University of the West of England (UWE) as part of the MSc in Library and Library Management, University of the West of England, Frenchay Campus, Bristol, March 24, 2009
Best Practices for Building and Deploying Data Pipelines in Apache SparkDatabricks
Many data pipelines share common characteristics and are often built in similar but bespoke ways, even within a single organisation. In this talk, we will outline the key considerations which need to be applied when building data pipelines, such as performance, idempotency, reproducibility, and tackling the small file problem. We’ll work towards describing a common Data Engineering toolkit which separates these concerns from business logic code, allowing non-Data-Engineers (e.g. Business Analysts and Data Scientists) to define data pipelines without worrying about the nitty-gritty production considerations.
We’ll then introduce an implementation of such a toolkit in the form of Waimak, our open-source library for Apache Spark (https://github.com/CoxAutomotiveDataSolutions/waimak), which has massively shortened our route from prototype to production. Finally, we’ll define new approaches and best practices about what we believe is the most overlooked aspect of Data Engineering: deploying data pipelines.
Data Marketplace and the Role of Data VirtualizationDenodo
Watch full webinar here: https://bit.ly/3IS9sQS
A data marketplace is like an online shopping interface specializing in data. Ideally, it should work just like an online store, with minimal latency and maximum responsiveness. However, this does not mean that all of the data in the data marketplace needs to be stored in the same central repository.
In this session, Shadab Hussain, Americas Sales Head, Data Analytics at Wipro, a partner company with Denodo and a co-sponsor of DataFest 2021, talks about the role of data virtualization in enabling full-featured data marketplaces. Such data marketplaces provide real-time, curated access to data, even when the data is stored across many different sources throughout the organization.
You will learn:
- The main features of a data marketplace
- Why organizations need data marketplaces
- Why data marketplaces sometimes fail
- How data virtualization enables the most effective data marketplaces
- How one of Europe’s premiere public healthcare system organizations leveraged a data marketplace to improve data consumption and ease of access
How a Semantic Layer Makes Data Mesh Work at ScaleDATAVERSITY
Data Mesh is a trending approach to building a decentralized data architecture by leveraging a domain-oriented, self-service design. However, the pure definition of Data Mesh lacks a center of excellence or central data team and doesn’t address the need for a common approach for sharing data products across teams. The semantic layer is emerging as a key component to supporting a Hub and Spoke style of organizing data teams by introducing data model sharing, collaboration, and distributed ownership controls.
This session will explain how data teams can define common models and definitions with a semantic layer to decentralize analytics product creation using a Hub and Spoke architecture.
Attend this session to learn about:
- The role of a Data Mesh in the modern cloud architecture.
- How a semantic layer can serve as the binding agent to support decentralization.
- How to drive self service with consistency and control.
Libraries and their Role in Open Access: Challenges and OpportunitiesFrançois Renaville
The open access movement gains momentum with an increasing number of institutions and funders adopting open access mandates for their funded research. Consequently, an increasing amount of material becomes freely available, either from institutional repositories or from traditional or newly established journals. Libraries can play a dual role in supporting this movement: Firstly, they can provide services supporting the deposit of research output in their institutional repositories, including support for making it widely discoverable via indexes such as Google Scholar and library discovery systems. Secondly, libraries can make open access materials discoverable by their patrons through such indexes, thus expanding their collection to include materials that they would not necessarily license.
This session will describe the experience of the University Libraries of Liège in Belgium and Harvard. University of Liège chose a top-down approach and made it compulsory for researchers to deposit their output in the institutional repository—ORBi. To support this mandate, the library offers services that help researchers deposit and disseminate their publications. Both libraries—Liège and Harvard—enable their students and faculty to discover open access content beyond their library’s acquired collection via their library discovery system.
The session will also address challenges that arise from indexing open access publications and how index providers and libraries can deal with such publications, especially with articles that are deposited in different institutional repositories or published in so-called hybrid journals that contain a mix of open access and subscription articles.
Finally, we will discuss with the audience how they see libraries’ role evolving in this area, what challenges they are currently facing, and the solutions and opportunities they have found.
Geek Sync | Data Architecture and Data Governance: A Powerful Data Management...IDERA Software
You can watch the replay for this Geek Sync webcast, Data Architecture and Data Governance: A Powerful Data Management Duo, on the IDERA Resource Center, http://ow.ly/95yL50A4rZg.
Batman and Robin. Han Solo and Chewbacca. Mario and Luigi. Just like these famous pairings, so it is for data architecture and data governance — they’re aligned to support each other in a variety of ways. Like data governance, data architecture as a practice can be leveraged to identify and enforce standards within the systems landscape to support business objectives. And data architecture certainly benefits from sound business oversight and stakeholder influences inherent in a successful data governance program.
Join Kelle O’Neal to learn about the critical aspects of aligning data architecture and data governance, with a specific focus on:
-Why aligning data architecture and data governance is important
-The key intersections of people, processes and technology between data architecture and data governance
-How data architecture and data governance work together to enforce standards
-The capabilities that data governance can apply to data architecture without interfering
-How your project and development methodologies can help drive alignment
EUDAT & OpenAIRE Webinar: How to write a Data Management Plan - July 14, 2016...EUDAT
| www.eudat.eu | 2nd Session: July 14, 2016.
In this webinar, Sarah Jones (DCC) and Marjan Grootveld (DANS) talked through the aspects that Horizon 2020 requires from a DMP. They discussed examples from real DMPs and also touched upon the Software Management Plan, which for some projects can be a sensible addition
Best Practices for Building and Deploying Data Pipelines in Apache SparkDatabricks
Many data pipelines share common characteristics and are often built in similar but bespoke ways, even within a single organisation. In this talk, we will outline the key considerations which need to be applied when building data pipelines, such as performance, idempotency, reproducibility, and tackling the small file problem. We’ll work towards describing a common Data Engineering toolkit which separates these concerns from business logic code, allowing non-Data-Engineers (e.g. Business Analysts and Data Scientists) to define data pipelines without worrying about the nitty-gritty production considerations.
We’ll then introduce an implementation of such a toolkit in the form of Waimak, our open-source library for Apache Spark (https://github.com/CoxAutomotiveDataSolutions/waimak), which has massively shortened our route from prototype to production. Finally, we’ll define new approaches and best practices about what we believe is the most overlooked aspect of Data Engineering: deploying data pipelines.
Data Marketplace and the Role of Data VirtualizationDenodo
Watch full webinar here: https://bit.ly/3IS9sQS
A data marketplace is like an online shopping interface specializing in data. Ideally, it should work just like an online store, with minimal latency and maximum responsiveness. However, this does not mean that all of the data in the data marketplace needs to be stored in the same central repository.
In this session, Shadab Hussain, Americas Sales Head, Data Analytics at Wipro, a partner company with Denodo and a co-sponsor of DataFest 2021, talks about the role of data virtualization in enabling full-featured data marketplaces. Such data marketplaces provide real-time, curated access to data, even when the data is stored across many different sources throughout the organization.
You will learn:
- The main features of a data marketplace
- Why organizations need data marketplaces
- Why data marketplaces sometimes fail
- How data virtualization enables the most effective data marketplaces
- How one of Europe’s premiere public healthcare system organizations leveraged a data marketplace to improve data consumption and ease of access
How a Semantic Layer Makes Data Mesh Work at ScaleDATAVERSITY
Data Mesh is a trending approach to building a decentralized data architecture by leveraging a domain-oriented, self-service design. However, the pure definition of Data Mesh lacks a center of excellence or central data team and doesn’t address the need for a common approach for sharing data products across teams. The semantic layer is emerging as a key component to supporting a Hub and Spoke style of organizing data teams by introducing data model sharing, collaboration, and distributed ownership controls.
This session will explain how data teams can define common models and definitions with a semantic layer to decentralize analytics product creation using a Hub and Spoke architecture.
Attend this session to learn about:
- The role of a Data Mesh in the modern cloud architecture.
- How a semantic layer can serve as the binding agent to support decentralization.
- How to drive self service with consistency and control.
Libraries and their Role in Open Access: Challenges and OpportunitiesFrançois Renaville
The open access movement gains momentum with an increasing number of institutions and funders adopting open access mandates for their funded research. Consequently, an increasing amount of material becomes freely available, either from institutional repositories or from traditional or newly established journals. Libraries can play a dual role in supporting this movement: Firstly, they can provide services supporting the deposit of research output in their institutional repositories, including support for making it widely discoverable via indexes such as Google Scholar and library discovery systems. Secondly, libraries can make open access materials discoverable by their patrons through such indexes, thus expanding their collection to include materials that they would not necessarily license.
This session will describe the experience of the University Libraries of Liège in Belgium and Harvard. University of Liège chose a top-down approach and made it compulsory for researchers to deposit their output in the institutional repository—ORBi. To support this mandate, the library offers services that help researchers deposit and disseminate their publications. Both libraries—Liège and Harvard—enable their students and faculty to discover open access content beyond their library’s acquired collection via their library discovery system.
The session will also address challenges that arise from indexing open access publications and how index providers and libraries can deal with such publications, especially with articles that are deposited in different institutional repositories or published in so-called hybrid journals that contain a mix of open access and subscription articles.
Finally, we will discuss with the audience how they see libraries’ role evolving in this area, what challenges they are currently facing, and the solutions and opportunities they have found.
Geek Sync | Data Architecture and Data Governance: A Powerful Data Management...IDERA Software
You can watch the replay for this Geek Sync webcast, Data Architecture and Data Governance: A Powerful Data Management Duo, on the IDERA Resource Center, http://ow.ly/95yL50A4rZg.
Batman and Robin. Han Solo and Chewbacca. Mario and Luigi. Just like these famous pairings, so it is for data architecture and data governance — they’re aligned to support each other in a variety of ways. Like data governance, data architecture as a practice can be leveraged to identify and enforce standards within the systems landscape to support business objectives. And data architecture certainly benefits from sound business oversight and stakeholder influences inherent in a successful data governance program.
Join Kelle O’Neal to learn about the critical aspects of aligning data architecture and data governance, with a specific focus on:
-Why aligning data architecture and data governance is important
-The key intersections of people, processes and technology between data architecture and data governance
-How data architecture and data governance work together to enforce standards
-The capabilities that data governance can apply to data architecture without interfering
-How your project and development methodologies can help drive alignment
EUDAT & OpenAIRE Webinar: How to write a Data Management Plan - July 14, 2016...EUDAT
| www.eudat.eu | 2nd Session: July 14, 2016.
In this webinar, Sarah Jones (DCC) and Marjan Grootveld (DANS) talked through the aspects that Horizon 2020 requires from a DMP. They discussed examples from real DMPs and also touched upon the Software Management Plan, which for some projects can be a sensible addition
Presentación de Joy Davidson, Digital Curation Centre (UK) en FOSTER event: Data Management Plan and Social Impact of Research. Universitat Jaume I, 27 mayo 2016
Stuart Macdonald steps through the process of creating a robust data management plan for researchers. Presented at the European Association for Health Information and Libraries (EAHIL) 2015 workshop, Edinburgh, 11 June 2015.
Presentation given at the Consorcio Madrono conference on Data Management Plans in Horizon 2020 http://www.consorciomadrono.es/info/web/blogs/formacion/217.php
EUDAT Research Data Management | www.eudat.eu | EUDAT
| www.eudat.eu | The presentation gives an introduction to Research Data Management, explaining why it is important to manage and share data.
November 2016
A talk outlining the virtues and processes of Research Data Management for PhD students in the geosciences. Given by Stuart Macdonald at the Introduction to RDM Workshop, School of Geosciences, University of Edinburgh, on 2 November 2015
Presentation from a University of York Library workshop on research data management. The workshop provides an introduction to research data management, covering best practice for the successful organisation, storage, documentation, archiving, and sharing of research data.
Similar to Research Data Management: An Introductory Webinar from OpenAIRE and EUDAT (20)
OpenAIRE Content Providers Community Call, November 4th, 2020
This call was focused on the PROVIDE future developments, functionalities wishlist and PROVIDE service in EOSC.
Was also an opportunity to share the most recent updates and novelties in the OpenAIRE Content Provider Dashboard, and to get feedback from community.
Recordings: https://youtu.be/wY4fOS767Us
Follow the Community activities at https://www.openaire.eu/provide-community-calls
OpenAIRE in the European Open Science Cloud (EOSC)OpenAIRE
Openness is the success factor for EOSC. OpenAIRE has been working in delivering an open access scholarly communication in Europe for the past 10 years and we now present how our work fits into the EOSC core developments
OpenAIRE Content Providers Community Call, October 7th, 2020
This call was focused on the OpenAIRE Broker Service, specifying how the service works to deploy the enrichment events to the Content Providers managers.
Was also an opportunity to share the most recent updates and novelties in the OpenAIRE Content Provider Dashboard, and to get feedback from community.
Recording: https://youtu.be/3sF4B58EGcs
Follow the Community activities at https://www.openaire.eu/provide-community-calls
OpenAIRE Content Providers Community Call, July 1st, 2020
This call was focused on Data Repositories namely the OpenAIRE Research Graph and Data Repositories, the OpenAIRE Content Acquisition Policy, and the Guidelines for Data Archive Managers.
Was also an opportunity to share the most recent updates and novelties in the OpenAIRE Content Provider Dashboard, and to get feedback from community.
Follow the Community activities at https://www.openaire.eu/provide-community-calls
OpenAIRE Content Providers Community Call. May 6th, 2020.
This Call focused the presentation of the new User Interface of Provide Dashboard and the presentation of 4 use cases using the Provide service.
Was also an opportunity to share the most recent updates and novelties in the OpenAIRE Content Provider Dashboard, and to get feedback from community.
Recording available here: https://youtu.be/J4m_ryRxtnY
20200504_OpenAIRE Legal Policy Webinar: GDPR and Sharing DataOpenAIRE
Presentation by Jacques Flores Dourojeanni (Research Data Management Consultant Utrecht University Library), as delivered during the OpenAIRE Legal Policy Webinar series on May 4th 2020.
More information and recordings: https://www.openaire.eu/item/openaire-legal-policy-webinars
20200504_Research Data & the GDPR: How Open is Open?OpenAIRE
Presentation by Prodromos Tsiavos (Senior Legal Advisor - ARC/ Director - Onassis Group) as delivered during the OpenAIRE Legal Policy Webinar series on May 4th 2020.
More information and recordings: https://www.openaire.eu/item/openaire-legal-policy-webinars
20200504_Data, Data Ownership and Open ScienceOpenAIRE
Presentation by Thomas Margoni (Senior Lecturer in Intellectual Property and Internet Law, Co-director, CREATe, University of Glasgow) as delivered during the OpenAIRE Legal Policy Webinar series on May 4th 2020.
More information and recordings: https://www.openaire.eu/item/openaire-legal-policy-webinars
20200429_Research Data & the GDPR: How Open is Open? (updated version)OpenAIRE
Presentation by Prodromos Tsiavos (Senior Legal Advisor - ARC/ Director - Onassis Group) as delivered during the OpenAIRE Legal Policy Webinar series on April 29th 2020.
More information and recordings: https://www.openaire.eu/item/openaire-legal-policy-webinars
20200429_Data, Data Ownership and Open ScienceOpenAIRE
Presentation by Thomas Margoni (Senior Lecturer in Intellectual Property and Internet Law, Co-director, CREATe, University of Glasgow) as delivered during the OpenAIRE Legal Policy Webinar series on April 29th 2020.
More information and recordings: https://www.openaire.eu/item/openaire-legal-policy-webinars
20200429_OpenAIRE Legal Policy Webinar: GDPR and Sharing DataOpenAIRE
Presentation by Jacques Flores Dourojeanni (Research Data Management Consultant Utrecht University Library), as delivered during the OpenAIRE Legal Policy Webinar series on April 29th 2020.
More information and recordings: https://www.openaire.eu/item/openaire-legal-policy-webinars
COVID-19: Activities, tools, best practice and contact points in GreeceOpenAIRE
Presentation from the webinar organized by the Greek OpenAIRE and RDA Nodes (Athena RC) and Elixir-GR to inform participants of EU and national efforts, in collaboration with the following research organizations: Flemming, CERTH, HEAL-Link, Demokritos, Univ. of Athens (Medical School).
Presentation of the 2nd Content Providers Community Call, targeting the following topics: 1) OpenAIRE Content provider dashboard updates; Main topic: DSpace-CRIS for OpenAIRE: implementation of the CRIS guidelines and beyond; 3) Community questions & comments.
Presentation of the 2nd Content Providers Community Call, targeting the following topics: 1) OpenAIRE Content provider dashboard updates;
2) OpenAIRE aggregation and enrichment processes: specifications and good practices;
3) Community questions & comments.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdfPeter Spielvogel
Building better applications for business users with SAP Fiori.
• What is SAP Fiori and why it matters to you
• How a better user experience drives measurable business benefits
• How to get started with SAP Fiori today
• How SAP Fiori elements accelerates application development
• How SAP Build Code includes SAP Fiori tools and other generative artificial intelligence capabilities
• How SAP Fiori paves the way for using AI in SAP apps
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Communications Mining Series - Zero to Hero - Session 1
Research Data Management: An Introductory Webinar from OpenAIRE and EUDAT
1. Research Data Management
- an introductory webinar
Tony Ross-Hellauer, OpenAIRE
Sarah Jones, EUDAT
This work is licensed under the Creative
Commons CC-BY 4.0 licence
2. Open Access Infrastructure
for Research in Europe
www.openaire.eu
Who we are
Research Data Services, Expertise &
Technology https://www.eudat.eu
3. • Why manage data?
• RDM in Horizon 2020 (+ recent changes)
• How to manage and share research data?
• EUDAT and OpenAIRE services
Overview
5. Data explosion
• More and more data is
being created
• Issue is not creating data,
but being able to navigate
and use it
• Data management is
critical to make sure data
are well-organised,
understandable and
reusable
6. Digital data are fragile and susceptible to loss for a wide variety of reasons
• Natural disaster
• Facilities infrastructure failure
• Storage failure
• Server hardware/software failure
• Application software failure
• Format obsolescence
• Legal encumbrance
• Human error
• Malicious attack
• Loss of staffing competencies
• Loss of institutional commitment
• Loss of financial stability
• Changes in user expectations
Data loss
Image CC BY-NC-SA 2.0 by Dave Hill https://www.flickr.com/photos/dmh650/4031607067
8. Why manage data?
• Make your research easier
• Stop yourself drowning in irrelevant stuff
• Save data for later
• Avoid accusations of fraud or bad science
• Share your data for re-use
• Get credit for it
• Meet funder/institution requirements
Because well-managed data opens up
opportunities for re-use, sharing and
makes for better science!
9. RDM IN HORIZON 2020
Image “Open Data” CC BY 2.0 by http://www.descrier.co.uk
10. EC Open Research Data Pilot,
Jan 2015 -
• A limited, voluntary pilot (initially 8 programme areas) with opt-out and
safeguards
• Participating projects must:
• Keep a data management plan, to be updated at regular intervals
• Deposit in an open access repository:
1. the data, including associated metadata, needed to validate the
results presented in scientific publications as soon as possible;
2. other data, including associated metadata, as specified and within the
deadlines laid down in the data management plan
11. EC Open Research Data Pilot
Opt-out Reasons
https://open-data.europa.eu/data/dataset/open-research-data-the-uptake-of-
the-pilot-in-the-first-calls-of-horizon-2020
14. CREATING
DATA
PROCESSING
DATA
ANALYSING
DATA
PRESERVING
DATA
GIVING
ACCESS TO
DATA
RE-USING
DATA
Research data lifecycle
CREATING DATA: designing research,
DMPs, planning consent, locate existing
data, data collection and management,
capturing and creating metadata
RE-USING DATA: follow-
up research, new
research, undertake
research reviews,
scrutinising findings,
teaching & learning
ACCESS TO DATA:
distributing data,
sharing data,
controlling access,
establishing copyright,
promoting data PRESERVING DATA: data storage, back-
up & archiving, migrating to best format
& medium, creating metadata and
documentation
ANALYSING DATA:
interpreting, & deriving
data, producing outputs,
authoring publications,
preparing for sharing
PROCESSING DATA:
entering, transcribing,
checking, validating and
cleaning data, anonymising
data, describing data,
manage and store data
Ref: UK Data Archive: http://www.data-archive.ac.uk/create-manage/life-cycle
15. • Findable
– assign persistent IDs, provide rich metadata, register in a
searchable resource...
• Accessible
– Retrievable by their ID using a standard protocol, metadata remain
accessible even if data aren’t...
• Interoperable
– Use formal, broadly applicable languages, use standard
vocabularies, qualified references...
• Reusable
– Rich, accurate metadata, clear licences, provenance, use of
community standards...
www.force11.org/group/fairgroup/fairprinciples
FAIR data
16. A DMP is a brief plan to define:
• how the data will be created?
• how it will be documented?
• who will access it?
• where it will be stored?
• who will back it up?
• whether (and how) it will be shared & preserved?
DMPs are often submitted as part of grant applications, but
are useful whenever researchers are creating data.
Data Management Plans
17. DMPonline
A web-based tool to help researchers write DMPs
Includes a template for Horizon 2020
Guidance from EUDAT and OpenAIRE being added
https://dmponline.dcc.ac.uk
18. • Metadata and documentation is needed to locate and
understand research data
• Think about what others would need in order to find,
evaluate, understand, and reuse your data.
• Get others to check the metadata to improve quality
• Use standards to enable interoperability
Metadata & documentation
20. Where to store data?
• Your own drive (PC, server, flash drive, etc.)
– And if you lose it? Or it breaks?
• Somebody else’s drive / departmental drive
• “Cloud” drive
– Do they care as much about your data as you do?
• Large scale infrastructure services like EUDAT
21. How to backup?
• 3... 2... 1... backup!
– at least 3 copies of a file
– on at least 2 different media
– with at least 1 offsite
• Use managed services where possible e.g. University
filestores or infrastructure services like EUDAT rather
than local or external hard drives
• Ask IT teams for advice
22. Backup and preservation
– not the same thing!
• Backups
– Used to take periodic snapshots of data in case the current version
is destroyed or lost
– Backups are copies of files stored for short or near-long-term
– Often performed on a somewhat frequent schedule
• Archiving
– Used to preserve data for historical reference or potentially during
disasters
– Archives are usually the final version, stored for long-term, and
generally not copied over
– Often performed at the end of a project or during major milestones
24. A mistake in a spreadsheet led
to dramatically different results
from those published.
These results were cited by
the International Monetary
Fund and the UK Treasury to
justify austerity programmes.
Had the data been shared, this
could have been picked up
earlier.
The importance of sharing data
25. Concerns about data sharing
Concern Solution
inappropriate use due to
misunderstanding of research
purpose or parameters
security and confidentiality of
sensitive data
lack of acknowledgement / credit
loss of advantage when competing
for research funding
26. Concerns about data sharing
Concern Solution
inappropriate use due to
misunderstanding of research
purpose or parameters
security and confidentiality of
sensitive data
lack of acknowledgement / credit
loss of advantage when competing
for research funding
metadata
metadata
metadata
metadata
27. Concerns about data sharing
Concern Solution
inappropriate use due to
misunderstanding of research
purpose or parameters
provide rich Abstract, Purpose,
Constraints and Supplemental
Information where needed
security and confidentiality of
sensitive data
• the metadata does NOT
contain the data
• Use Constraints specify who
may access the data and how
lack of acknowledgement / credit
specify a required data citation
within the Use Constraints
loss of data insight and
competitive advantage when vying
for research funding
create second, public version with
generalised Data Processing
Description
28. Make data shareable
• Create robust metadata that has been checked
• Include reference information in metadata e.g. unique
IDs & properly formatted data citations
• Publish your metadata so it’s discoverable. Use portals,
clearing houses, online resources…
• Package up the data and associated metadata to deposit
in repositories
• License the data clearly
29. www.dcc.ac.uk/resources/how-guides/license-research-data
Licensing research data
This DCC guide outlines the pros and
cons of each approach and gives
practical advice on how to implement
your licence
CREATIVE COMMONS LIMITATIONS
NC Non-Commercial
What counts as commercial?
ND No Derivatives
Severely restricts use
These clauses are not open licenses
Horizon 2020 Open Access
guidelines point to:
or
30. EUDAT licensing tool
Answer questions to determine which licence(s) are
appropriate to use
http://ufal.github.io/public-license-selector
31. What to preserve & share
It’s not possible to keep everything. Select based on:
– What has to be kept e.g. data underlying publications
– What can’t be recreated e.g. environmental recordings
– What is potentially useful to others
– What has scientific, cultural or historical value
– What legally must be destroyed
How to select and appraise research data:
www.dcc.ac.uk/resources/how-guides/appraise-select-research-data
32. EUDAT & OPENAIRE SERVICES
Image CC-BY-NC ‘Data centre’ by Bob Mical www.flickr.com/photos/small_realm/15995555571
33. EUDAT services
EUDAT offers a pan-European solution, providing a
generic set of services to ensure minimum level of
interoperability
Building common
data services in
close collaboration
with 25+
communities
34. EUDAT B2 service suite
Covering both access and
deposit, from informal data
sharing to long-term
archiving, and addressing
identification,
discoverability and
computability of both long-
tail and big data, EUDAT’s
services will address the
full lifecycle of research
data
38. OpenAIRE training and
support materials
• Briefing papers, factsheets,
Webinars, workshops,
FAQs
• Information on:
• Open Research Data Pilot
• Creating a data management
plan
• Selecting a data repository
https://www.openaire.eu/opendatapilot
https://www.openaire.eu/support
39. www.eudat.eu www.openaire.eu
Thanks – any questions?
Contact us:
Tony Ross-Hellauer, OpenAIRE: ross-hellauer@sub.uni-goettingen.de
Sarah Jones, EUDAT: Sarah.Jones@glasgow.ac.uk
Acknowledgements:
Thanks to EUDAT colleagues Mark van de Sanden and Christine Staiger
for slides.
Content has also been repurposed from the DataONE Educational
modules, ‘Data Management’ and ‘Data Sharing’ Retrieved from
https://www.dataone.org/education-modules
Editor's Notes
There are four main topics that we will discuss:
Why manage data - The changing data landscape, looking at what issues this brings.
Brief overview of evolution of EC’s RDM policies
Secondly, we discuss considerations to make when managing and sharing data
Finally we’ll touch on EUDAT and OpenAIRE services to show how support is provided throughout the lifecycle
So let’s begin by looking at the changing data landscape.
There’s been a data explosion.
1. 90% of all the data in the world has been generated over the last 2 years.
2. Scientific data output is currently increasing at an annual rate of 30%.
As the amount of data being created now is growing exponentially, the biggest challenge is being able to navigate and use it. This is why data management is critical.
Digital data are fragile. There are lots of ways in which data can be lost. Hardware and software can fail, formats can become obsolete, you can lose the knowledge and skills needed to understand the data, and you can lose the investment needed to keep the data accessible.Despite significant investment, data is not being managed effectively
The current estimated total global spend on research and development is $1.5 trillion, which could be at risk.
Much of the data generated is lost – in one study, the odds of sourcing datasets declined by 17% each year.
The same study found 80% of datasets over 20 years old not available.
Many experimentally established "facts" don't seem to hold up to repeated investigation. Several studies have shown alarming numbers of published papers that don’t stand up to scrutiny.
Over half of psychology studies fail reproducibility test (61/100) – Nosek et al, Science, 2015
Causes of reproducibility not well understood – but can say that it is obvious that where the original data is available, accountability is increased – able to review where questions arise.
There are lots of reasons to manage research data. Ultimately though, it’s to make your research easier. If data are properly documented and organised, you can stop yourself drowning in irrelevant stuff and find the data when you need it – for example to validate findings. By managing your data you can also more easily share it with others to get more credit and impact. You may also be required to explain how you will manage your data by your funder or university.
Well-managed data opens up opportunities for re-use, integration and new science
Let’s move on to the considerations to make when managing and sharing data
Introduced at the start of 2015, covering just seven work programme areas, the Horizon 2020 Open Research Data Pilot has been a big success. In the first six months of the pilot, about a third of projects (65.4%, 431 signed grant agreements) that were part of the pilot chose to opt out. The most common reasons for opting out were: (1) concerns over intellectual property (37%), (2) the project did not expect to generate any data (18%), and privacy/data protection concerns (18%). Of those projects that were not originally part of the pilot, 11.9% (3268 projects) nonetheless have voluntarily opted in.
Introduced at the start of 2015, covering just seven work programme areas, the Horizon 2020 Open Research Data Pilot has been a big success. In the first six months of the pilot, about a third of projects (65.4%, 431 signed grant agreements) that were part of the pilot chose to opt out. The most common reasons for opting out were: (1) concerns over intellectual property (37%), (2) the project did not expect to generate any data (18%), and privacy/data protection concerns (18%). Of those projects that were not originally part of the pilot, 11.9% (3268 projects) nonetheless have voluntarily opted in.
Let’s move on to the considerations to make when managing and sharing data
This research data lifecycle is taken from the UK Data Archive. It shows you the different processes and activities you’ll go through.
Creating data: This is when you’ll design the research, write Data Management Plans, negotiate consent agreements, find any existing data you want to reuse, collect/capture your data and create any associated metadata
Processing data: When processing your data, you’ll be entering, transcribing, checking, validating and cleaning it, you may also need to anonymise your data, you should describe it and make sure it’s properly managed and stored.
Analysing data: when you analyse your data you’ll be interpreting it and creating derived data and outputs, you’ll probably also author publications and prepare the data for deposit and sharing.
Preserving data: data repositories play a key role in preserving data: they will make sure it’s properly stored and archived, they will migrate the formats and storage medium and create associated metadata and documentation to explain any changes made
Access to data: it may be that you share your data via a repository or handle access requests yourself. Either way, you need to establish copyright, decide who can have access and promote the data.
Re-using data: data can be re-used in follow-up studies, new research, research reviews, to evidence findings or for teaching and learning. Try to keep an open mind about the different ways in which your data could be re-used and make it as open as possible.
A Data Management Plan is often written early on in the research process to determine what data will be created and how it will be managed. Sometime you are asked for a DMP as part of a grant application, but they are useful to write regardless as it helps to develop consistent procedures from the outset.
Metadata is needed to locate and understand the data. When you are deciding what information to capture, think about what others would need in order to find, evaluate, understand, and reuse your data. Also get others to check your metadata to improve the quality and make sure it’s understandable to others. Standards should be used where possible.
To make sure their data can be understood by themselves, their community and others, researchers should create metadata and documentation.
Metadata is basic descriptive information to help identify and understand the structure of the data e.g. title, author...
Documentation provides the wider context. It’s useful to share the methodology / workflow, software and any information needed to understand the data e.g. explanation of abbreviations or acronyms
There are lots of standards that can be used. The DCC started a catalogue of disciplinary metadata standards which is now being taken forward as an international initiative via an RDA working group
There are lots of places you can store your data. You’re best to use managed services where possible as they’re more resilient. If you store data on standalone computers, memory sticks or in the cloud, be mindful of the risk of loss or security breaches.
If you’re responsible for backing up your own data, you want to ensure there are multiple copies, on different media with at least 1 offsite. Where possible though, you should use managed services so the backup is done automatically for you.
Remember that backup and preservation are not the same thing (though the terms are often used interchangeably).
Backups are performed regularly to take periodic snapshots of the data for the short to medium term, whereas archiving is preserving the final version of the data for the long-term.
You should make sure your data are backed-up during the active phase of research and that any data needed for the long-term are archived.
It is also important to share your data where possible, particularly to evidence your findings.
This article reflects on an inadvertent error in a economics paper by Reinhadt and Rogoff. Missing some rows out of an average gave drastically different results – what was published suggested that countries with 90% debt ratios see their economies shrink by 0.1%. Instead, it should have found that they grow by 2.2% – less than those with lower debt ratios, but not a spiralling collapse. This mistake wasn’t picked up on initially as the data hadn’t been shared. The mistake fed into government policy as the findings were used as justification for austerity measures in the UK and various other countries in the EU.
Naturally, researchers may worry that the data will be taken out of context, misinterpreted or used inappropriately. They may also be concerned about maintaining the confidentiality and security of sensitive data. Business concerns may arise as well - will data users give proper credit and acknowledgement to the scientist? Will the scientist lose a competitive advantage by sharing this valuable resource?
There are lots of reasons why researchers may be reluctant to share data, so what is the solution?
Each of these issues can, in great part, be addressed by providing rich data documentation known as ‘metadata’.
By providing metadata, the research scientist establishes the purpose, methods, sources and parameters of the data. As such, data users are given the information necessary to appropriately apply, protect and cite the data. If the metadata contains information about proprietary data processing or analysis techniques, the competitive advantage can be maintained by creating a second, more generalized, metadata record for public distribution.
To make your data shareable, you should create robust metadata and seek a second a second opinion on this to ensure it’s understandable to others. Also include reference information so others can find your data and give you credit. The metadata should be published online and packaged up with your data to deposit in repositories.
Guidance from the DCC can also help researchers to understand data licensing. This guide outlines the pros and cons of each approach e.g. the limitations of some CC options
The OA guidelines under Horizon 2020 point to CC-0 or CC-BY as a straightforward and effective way to make it possible for others to mine, exploit and reproduce the data. See p11 at: http://ec.europa.eu/research/participants/data/ref/h2020/grants_manual/hi/oa_pilot/h2020-hi-oa-pilot-guide_en.pdf
It might not be possible to preserve and share all your data, so you may need to make a selection. Some factor to consider could be what has to be kept, for example for legal reasons or to evidence findings, what is potentially useful to others or can’t be recreated. You may also be under obligation to destroy certain data due to consent agreements or commercial non-disclosure restrictions.
The Digital Curation Centre has guidance on how to select what data to keep.
Let’s close by looking briefly at the EUDAT service suite and how it helps with data management and sharing
EUDAT offers a pan-European solution, providing a generic set of data services. These are being built in close collaboration with user communities.
The services assist researchers to store, manage and process the data through-out the active phase of research, and also help to archive data and make it discoverable to others.
The B2DROP service helps you to syncronrise and exchange research data like Dropbox; B2STAGE helps you get data to computation when processing and analysing data; B2SAFE helps you to replicate the data safely; B2SHARE is a repository to archive the data and share it with others; and B2FIND is a cataloguing service that allows you and others to find relevant data.
Catch-all repository
Multiple data types
Publications
Long tail of research data
Citable data (DOI)
Links to funding, pubs, data, software
Should happen automatically thanks to our data-literature interlinking services
But where it doesn’t, you