Openness is the becoming more crucial feature of science.
Science is to be open because of healthiness of science itself and its public relationship. Openness in science is not only papers but also data now. Scholarly communication should adapt to the openness of papers and data in science. It is a long experience for openness of papers, i.e., open access of papers, but openness of data is relatively new and more challenging since data is diverse in contents and quantity and manifold in use. One of the key technology for sharing papers and data is identifier. Identifiers can remove ambiguity of digital objects and their attributions, realize smarter use of data, and make more linkages to other information. DOI (Digital Object Identifier) is the most successful identifiers in this area, which is now extending their target to data. Japan Link Center (JaLC) now starts the experimental use of DOI for data. The other identifier to be noted is ORCID (Open Researcher and Contributor Identifier), which is designed and used for identifying researchers worldwide. Embeding these idenfiers in metadata can make more suitable scholarly commucation to openness of science.
DataCite and Campus Data Services
Paul Bracke, Associate Dean for Digital Programs and Information Services, Purdue University
Research libraries are increasingly interested in developing data services for their campuses. There are many perspectives, however, on how to develop services that are responsive to the many needs of scientists; sensitive to the concerns of scientists who are not always accustomed to sharing their data; and that are attractive to campus administrators. This presentation will discuss the development of campus-based data services programs, the centrality of data citation to these efforts, and the ways in which engagement with DataCite can enhance local programs.
Openness is the becoming more crucial feature of science.
Science is to be open because of healthiness of science itself and its public relationship. Openness in science is not only papers but also data now. Scholarly communication should adapt to the openness of papers and data in science. It is a long experience for openness of papers, i.e., open access of papers, but openness of data is relatively new and more challenging since data is diverse in contents and quantity and manifold in use. One of the key technology for sharing papers and data is identifier. Identifiers can remove ambiguity of digital objects and their attributions, realize smarter use of data, and make more linkages to other information. DOI (Digital Object Identifier) is the most successful identifiers in this area, which is now extending their target to data. Japan Link Center (JaLC) now starts the experimental use of DOI for data. The other identifier to be noted is ORCID (Open Researcher and Contributor Identifier), which is designed and used for identifying researchers worldwide. Embeding these idenfiers in metadata can make more suitable scholarly commucation to openness of science.
DataCite and Campus Data Services
Paul Bracke, Associate Dean for Digital Programs and Information Services, Purdue University
Research libraries are increasingly interested in developing data services for their campuses. There are many perspectives, however, on how to develop services that are responsive to the many needs of scientists; sensitive to the concerns of scientists who are not always accustomed to sharing their data; and that are attractive to campus administrators. This presentation will discuss the development of campus-based data services programs, the centrality of data citation to these efforts, and the ways in which engagement with DataCite can enhance local programs.
S. Venkataraman (DCC) talks about the basics of Research Data Management and how to apply this when creating or reviewing a Data Management Plan (DMP). He discusses data formats and metadata standards, persistent identifiers, licensing, controlled vocabularies and data repositories.
link to : dcc.ac.uk/resources
Functional and Architectural Requirements for Metadata: Supporting Discovery...Jian Qin
The tremendous growth in digital data has led to an increase in metadata initiatives for different types of scientific data, as evident in Ball’s survey (2009). Although individual communities have specific needs, there are shared goals that need to be recognized if systems are to effectively support data sharing within and across all domains. This paper considers this need, and explores systems requirements that are essential for metadata supporting the discovery and management of scientific data. The paper begins with an introduction and a review of selected research specific to metadata modeling in the sciences. Next, the paper’s goals are stated, followed by the presentation of valuable systems requirements. The results include a base-model with three chief principles: principle of least effort, infrastructure service, and portability. The principles are intended to support “data user” tasks. Results also include a set of defined user tasks and functions, and applications scenarios.
How Portable Are the Metadata Standards for Scientific Data?Jian Qin
The one-covers-all approach in current metadata standards for scientific data has serious limitations in keeping up with the ever-growing data. This paper reports the findings from a survey to metadata standards in the scientific data domain and argues for the need for a metadata infrastructure. The survey collected 4400+ unique elements from 16 standards and categorized these elements into 9 categories. Findings from the data included that the highest counts of element occurred in the descriptive category and many of them overlapped with DC elements. This pattern also repeated in the elements co-occurred in different standards. A small number of semantically general elements appeared across the largest numbers of standards while the rest of the element co-occurrences formed a long tail with a wide range of specific semantics. The paper discussed implications of the findings in the context of metadata portability and infrastructure and pointed out that large, complex standards and widely varied naming practices are the major hurdles for building a metadata infrastructure.
EZID: Easy dataset identification & management
Joan Starr, Manager, Strategic and Project Planning and EZID Service Manager, California Digital Library
Data and data curation are assuming a growing role today’s research library. New approaches are needed both to address the resulting challenges and take advantage of the emerging opportunities. Long-term identifiers represent one such tool. In this presentation, Joan Starr will introduce identifiers and an application designed to make them easy to create and manage: EZID. She will provide a closer look at two identifier types: DOIs and ARKs, and discuss what bringing an identifier service to your institution might mean.
This slideshow was used in a Preparing Your Research Data for the Future course taught in the Medical Sciences Division, University of Oxford, on 2015-06-08. It provides an overview of some key issues, focusing on long-term data management, sharing, and curation.
Going Full Circle: Research Data Management @ University of PretoriaJohann van Wyk
Presentation delivered at the eResearch Africa Conference, held 23-27 November 2014, at the University of Cape Town, Cape Town, South Africa. Various approaches to Research Data Management at Higher Education Institutions focus on an aspect or two of the research data cycle. At the University of Pretoria the approach has been to support researchers throughout the research process covering the whole research data cycle. The idea is to facilitate/capture the research data throughout the research cycle. This will give context to the data and will add provenance to the data. The University of Pretoria uses the UK Data Archive’s research data cycle model, to align its Research Data Management project-development. This model identifies the stages of a research data cycle as: creating data, processing data, analysing data, preserving data, giving access to data, and reusing data. This paper will give a short overview of the chronological development of research data management at the University of Pretoria. The overview will also highlight findings of two surveys done at the University, one in 2009 and one in 2013. This will be followed by a discussion of a number of pilot projects at the University, and how the needs of researchers involved in these projects are being addressed in a number of the stages of the research data cycle. The discussion will also give a short overview of how the University plans to support those stages not currently being addressed. The second part of the presentation will focus on the projects and technology (software and hardware) used. The University of Pretoria has adopted an Enterprise Content Management (ECM) approach to manage its Research Data. ECM is not a singular platform or system but rather a set of strategies, tools and methodologies that interoperate with each other to create a comprehensive management tool. These sets create an all-encompassing process addressing document, web, records and digital asset management. At the University of Pretoria we address all these processes with different software suites and tools to create a complete management system. Each process presented its own technical challenges. These had to be addressed, while keeping in mind the end objective of supporting researchers throughout the whole research process and data life cycle. Various platforms and standards have been adopted to meet the University of Pretoria’s criteria. To date three processes have been addressed namely, the capturing of data during the research process, the dissemination of data and the preservation of data.
A presentation given by Manjula Patel (UKOLN) at the Repository Curation Environments (RECURSE) Workshop held at the 4th International Digital Curation Conference, Edinburgh, 1st December 2008,
http://www.dcc.ac.uk/events/dcc-2008/programme/
S. Venkataraman (DCC) talks about the basics of Research Data Management and how to apply this when creating or reviewing a Data Management Plan (DMP). He discusses data formats and metadata standards, persistent identifiers, licensing, controlled vocabularies and data repositories.
link to : dcc.ac.uk/resources
Functional and Architectural Requirements for Metadata: Supporting Discovery...Jian Qin
The tremendous growth in digital data has led to an increase in metadata initiatives for different types of scientific data, as evident in Ball’s survey (2009). Although individual communities have specific needs, there are shared goals that need to be recognized if systems are to effectively support data sharing within and across all domains. This paper considers this need, and explores systems requirements that are essential for metadata supporting the discovery and management of scientific data. The paper begins with an introduction and a review of selected research specific to metadata modeling in the sciences. Next, the paper’s goals are stated, followed by the presentation of valuable systems requirements. The results include a base-model with three chief principles: principle of least effort, infrastructure service, and portability. The principles are intended to support “data user” tasks. Results also include a set of defined user tasks and functions, and applications scenarios.
How Portable Are the Metadata Standards for Scientific Data?Jian Qin
The one-covers-all approach in current metadata standards for scientific data has serious limitations in keeping up with the ever-growing data. This paper reports the findings from a survey to metadata standards in the scientific data domain and argues for the need for a metadata infrastructure. The survey collected 4400+ unique elements from 16 standards and categorized these elements into 9 categories. Findings from the data included that the highest counts of element occurred in the descriptive category and many of them overlapped with DC elements. This pattern also repeated in the elements co-occurred in different standards. A small number of semantically general elements appeared across the largest numbers of standards while the rest of the element co-occurrences formed a long tail with a wide range of specific semantics. The paper discussed implications of the findings in the context of metadata portability and infrastructure and pointed out that large, complex standards and widely varied naming practices are the major hurdles for building a metadata infrastructure.
EZID: Easy dataset identification & management
Joan Starr, Manager, Strategic and Project Planning and EZID Service Manager, California Digital Library
Data and data curation are assuming a growing role today’s research library. New approaches are needed both to address the resulting challenges and take advantage of the emerging opportunities. Long-term identifiers represent one such tool. In this presentation, Joan Starr will introduce identifiers and an application designed to make them easy to create and manage: EZID. She will provide a closer look at two identifier types: DOIs and ARKs, and discuss what bringing an identifier service to your institution might mean.
This slideshow was used in a Preparing Your Research Data for the Future course taught in the Medical Sciences Division, University of Oxford, on 2015-06-08. It provides an overview of some key issues, focusing on long-term data management, sharing, and curation.
Going Full Circle: Research Data Management @ University of PretoriaJohann van Wyk
Presentation delivered at the eResearch Africa Conference, held 23-27 November 2014, at the University of Cape Town, Cape Town, South Africa. Various approaches to Research Data Management at Higher Education Institutions focus on an aspect or two of the research data cycle. At the University of Pretoria the approach has been to support researchers throughout the research process covering the whole research data cycle. The idea is to facilitate/capture the research data throughout the research cycle. This will give context to the data and will add provenance to the data. The University of Pretoria uses the UK Data Archive’s research data cycle model, to align its Research Data Management project-development. This model identifies the stages of a research data cycle as: creating data, processing data, analysing data, preserving data, giving access to data, and reusing data. This paper will give a short overview of the chronological development of research data management at the University of Pretoria. The overview will also highlight findings of two surveys done at the University, one in 2009 and one in 2013. This will be followed by a discussion of a number of pilot projects at the University, and how the needs of researchers involved in these projects are being addressed in a number of the stages of the research data cycle. The discussion will also give a short overview of how the University plans to support those stages not currently being addressed. The second part of the presentation will focus on the projects and technology (software and hardware) used. The University of Pretoria has adopted an Enterprise Content Management (ECM) approach to manage its Research Data. ECM is not a singular platform or system but rather a set of strategies, tools and methodologies that interoperate with each other to create a comprehensive management tool. These sets create an all-encompassing process addressing document, web, records and digital asset management. At the University of Pretoria we address all these processes with different software suites and tools to create a complete management system. Each process presented its own technical challenges. These had to be addressed, while keeping in mind the end objective of supporting researchers throughout the whole research process and data life cycle. Various platforms and standards have been adopted to meet the University of Pretoria’s criteria. To date three processes have been addressed namely, the capturing of data during the research process, the dissemination of data and the preservation of data.
A presentation given by Manjula Patel (UKOLN) at the Repository Curation Environments (RECURSE) Workshop held at the 4th International Digital Curation Conference, Edinburgh, 1st December 2008,
http://www.dcc.ac.uk/events/dcc-2008/programme/
Building Open Research Infrastructure with PIDsETH-Bibliothek
Learn more about ORCID, how it enables connections between persistent identifiers to increase transparency and trust in research information and how to get involved.
VIVO and persistent identifiers: Integrating ORCID_08152013Rebecca Bryant, PhD
Title: VIVO and persistent identifiers: Integrating ORCID
Presented at the VIVO 2013 conference in St. Louis, MO, 08/15/13
Presenters:
Rebecca Bryant, PhD, ORCID, Bethesda, MD, USA
Hal Warren, American Psychological Association, Washington DC
Simeon Warner, PhD, Cornell University, Ithaca, NY
Abstract:
Since the launch of the ORCID Registry in October 2012, thousands of researchers have claimed their ORCID iD. Organizations have been embedding ORCID identifiers in manuscript submission systems, in funding applications, and adding them to university profile systems. Even before launch, the VIVO ontology had incorporated an ORCID field. In this panel, we will provide an overview of the ORCID registry and adoption, and demonstrate how the American Psychological Association (APA) has integrated ORCID identifiers into its VIVO system and developed an application to populate ORCID records with demographic and publication attributes from APA VIVO RDF files. The ORCID data are packaged as a JSON object stored as a URI in the VIVO record. This serves as a cross-check for ORCID assertions from the publisher of works claimed and allows APA to use VIVO to extend valid provenance assertions for publications in a Linked Open Data Trust Framework. We will discuss the application of this use case for other VIVO implementations and other researcher profiling systems, focusing on integrations at universities.
RDA Presentation by Hilary Hanahoe at Open Science 2020 event, Pisa 8th April - Sharing data across technologies, disciplines and countries, what is it, how does it work, how and why you should get involved
Who's the Author? Identifier soup - ORCID, ISNI, LC NACO and VIAFSimeon Warner
Identifiers, including ORCID, ISNI, LC NACO and VIAF, are playing an increasing role in library authority work. Well describe changes to cataloging practices to leverage identifiers. We'll then tell a short story of the how and why of ORCID identifiers for researchers, and relationships with other person identifiers. Finally, we'll discuss the use of identifiers as part of moves toward linked data cataloging being explored in Linked Data for Libraries work (in the LD4L Labs and LD4P projects).
IDS Project: Promoting library excellence through community and technologyTim Bowersox
An overview of the IDS Project, a library cooperative in NY state that promotes library excellence through community and technology. For more info, visit http://idsproject.org
Presented at Journal Paper Track, The Web Conference, Lyon, France, April 15, 2018
https://doi.org/10.1145/3184558.3186234
Abstract: Linked Open Data (LOD) technology enables web of data and exchangeable knowledge graphs through the Internet. However, the change in knowledge is happened everywhere and every time, and it becomes a challenging issue of linking data precisely because the misinterpretation and misunderstanding of some terms and concepts may be dissimilar under different context of time and different community knowledge. To solve this issue, we introduce an approach to the preservation of knowledge graph, and we select the biodiversity domain to be our case studies because knowledge of this domain is commonly changed and all changes are clearly documented. Our work produces an ontology, transformation rules, and an application to demonstrate that it is feasible to present and preserve knowledge graphs and provides open and accurate access to linked data. It covers changes in names and their relationships from different time and communities as can be seen in the cases of taxonomic knowledge.
We propose Crop Vocabulary(CVO) as a basis of the core vocabulary of crop names that becomes the guidelines for data interoperability between agricultural ICT systems on the food chain. Since a single species is treated in different ways, there are many different types of crop names. So, we organize the crop name discriminated by properties such as scientific name, planting method, edible part and registered cultivar information. Also, Crop Vocabulary is also linked to existing vocabularies issued by Japanese government agency and international organization such as AGROVOC. It is expected to use in the data format in the agricultural ICT system.
Presented in 45th Asia Pacific Advanced Network (APAN45) Meeting, Singapore (2018)
Presented as the invited talk at International Workshop on kNowledge eXplication for Industry (kNeXI2017). In this talk, I explain the experience and lesson learnt how to build ontologies. I am currently building the agriculture activity ontology (AAO). It describes classification and properties of various activities in the agriculture domain. It is formalized with Description Logics.
Presented at the Interest Group on Agricultural Data (IGAD) ,3 April, 2017, Barcelona, Spain
Abstract: n this talk, we present the current status of our agriculture ontologies that are developed to accelerate the data use in agriculture.
The agriculture activity ontology formalizes the activities in agriculture. We have developed it for three years. Now we are developing its applications. One application is to exchange formats between different farmer management systems. Another ontology is the crop ontology that standardizes the names of crops. The structure is simple but has links to many other standards in distribution industry, food industry and so on.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
PHP Frameworks: I want to break free (IPC Berlin 2024)
Working with Global Infrastructure at a National Level
1. Working with Global
Infrastructure at a National Level
Hideaki Takeda
Chair, Joint Steering Committee, Japan Link Center
Professor, National Institute of Informatics
http://japanlinkcenter.org/
1Joint Global Infrastructure Conference (ORCID/Crossref/DataCite), June 15, Seoul, South Korea
http://orcid.org/0000-0002-2909-7163
takeda@nii.ac.jp
8. Global Infrastructure for Scholarly Communication
ID
ID ID
ID
ID
ID
ID
ID
IDID
ID
• ID for
• Article
• Data
• Researcher
• Institutions, affiliation
• Funding agency, funded project
• Academic society
• Topic
• …
9. Global Infrastructure for Scholarly Communication
• So the global activity is going on.
• IDF, CrossRef, DataCite, ORCID …
• Why is the nation-wide activity needed?
• Realize interoperability over difference by nation
9
10. Global Infrastructure for Scholarly Communication
• Realize interoperability over difference by nation
• Difference from
• Language
• Scholarly Systems
• Scholarly Organization
• Scholarly Culture
• Interoperability for
• ID Systems
• Metadata
• Systems
• Information Flow
10
11. Japan Link Center (JaLC)
• Founded in March 2012
• Aimed to register DOIs for academic contents produced
in Japan or in Japanese, to circulate information in Japan and
overseas.
• Controlled by four national organizations:
Japan Science and Technology Agency (JST)
National Institute for Materials Science (NIMS)
National Institute of Informatics (NII)
National Diet Library (NDL)
• Operated by JST
• Membership system
29 Regular members
(Academic societies, Publishers, University libraries, etc)
1200+ Associate members
(978 associate members under JST, 144 associate members under NII)
• External coordination
JaLC is a member of CrossRef and DataCite.
11
12. Role of JaLC as a national service
• 1. Offer various ways for registration procedures
• 2. Offer the total service for PID Registration
• 3. Connecting various content holders and users
• 4. Pick up and implement various requests for DOI
• 5. Nation-wide outreach of DOI for various sectors
• 6. Locally suitable business model
12
13. Offer various ways for registration procedures
• Support various ways to register DOIs
• Small institutions are incapable to register DOI by themselves
• Consolidate the existing information flow
• Current Implementation
• via J-Stage (JST service for academic E-journals)
• via IRDB (Institutional Repository Aggregator)
• Via Japan Medical Abstracts Society
13
14. Organizational structure of DOI Registration
14
Members
Associate
members
International DOI Foundation
(IDF)
Registration
Agencies
etc.
etc.
etc.
Academic
societies
etc.Universities
DOI
Registrant
IDF
NDL29 members
1,518 members
15. Organizational structure of DOI Registration
15
Members
Associate
members
International DOI Foundation
(IDF)
Registration
Agencies
etc.
etc.
etc.
Academic
societies
etc.Universities
DOI
Registrant
IDF
NDL
Journals Institutional Repositories
16. Offer various ways for registration procedures
• via J-Stage (JST service for academic E-journals)
• J-Stage: 1,290 academic societies / 2,191 titles / 3,283,542 articles
• DOI Registration 1,258 assoc. members / 2,007 titles / 2,880,097 articles
(97.5%) (91.6%) (87.7%)
16
Academic Society
Academic Society
Academic Society
Academic Society
Academic Society
E-journal Platform
Journal editing
17. 17
Improving the number of access by connecting with external services
Other services
meta data
Collaboration with other services
Search engines
18. Offer various ways for registration procedures
• via IRDB (Institutional Repository Aggregator)
• IRDB: 621 institutions / 1,984,896 items
• DOI Registration 254 institutions / 118,619 items
(40.9%) (6.0%)
18
Institutional Repository
Institutional Repository
Institutional Repository
Institutional Repository
Institutional Repository
IR Aggregator & IR Search Service
harvesting
University
IRDB
19. Offer the total service for PID Registration
19
Members
Associate
members
International DOI Foundation
(IDF)
Registration
Agencies
etc.
etc.
Academic
societies
etc.Universities
DOI
Registrant
IDF
NDL
20. Offer the total service for PID Registration
20
IDF
Articles
DOI & JaLCメタデータ
JaLC Mem.JaLC Mem.JaLC Mem.
JaLC Assoc. JaLC Assoc. JaLC Assoc.
JaLC Mem.
LHS
DOI, URL
Research Data
JaLC Assoc.
DOI, URL DOI, URL
DOI、Crossref Metadata
DOI、DataCite Metadata
JaLC Metadata
22. ID for Researchers
• ORCID ID
• e-rad/KAKEN ID:
• Provided for grant application by MEXT (Ministry of Education, …)
• Almost all researchers in universities are registered
22
UniversitiesResearchers
e-rad/KAKEN ID
Gov. Funders
ORCID ID
24. Scholarly Information Infrastructure in NII
KAKEN
NII-ELS
IInstitutional
repository
Catalog information Research Information
More than
650 institutions
J-Stage
(JST) NDL
Linkage to other DB services Univ.
Library
More than
1,200 libraries
JAIRO
JSPS MEXT
Metadata and links of
Japanese journal
articles
19 M records
CiNii
Articles
Full-text of
journal articles
4.26 M records
Metadata and links of
Japanese institutional
repositories
2.22 M records
学協会
学協会Academic
Societies
Project reports of MEXT
supported scientific
researches
760 K record
compilation
digitizationintegration
digitization
compilation
NACSIS-CAT
Note: The record
numbers are as of
March 2016
Repositories
Cloud Service
CiNii
Books
Catalog of materials held by
universities
Bibliographic info 11 M records
Holding information 130 M records
Journal articles Journal articles
Scholarly information is disseminated through various portals provided by NII,
in which the information is compiled with the collaboration with universities.
Universities and Research Institutions
24
ID/Metadata for Object/Person/Project
25. Research Data Infrastructure for Open Science in NII
25
Discovery Service
Publication Platform
Research Data Management System
DOI
Subject
Repository
Metadata Management
● Linking Func between Article and Data
● Researcher and Research Project
Identification and Management Func
● Data Exchange with International
Discovery Service
Access Control Metadata Mng
Journal
Article Supplemental
Data
Institutional Research Data Mng
Hot
Storage
Hot
Storage
Hot
Storage
Cold
Storage
Cold
Storage
Cold
Storage
Data Depositor
ArchiveExp/Store
Search/Find
Data User
Article
● High Speed Access using SINET5
● Data Sharing Func using
Virtual NW and ID Federation
● Effective Data Storage Switcher
● Data oriented Self-Archiving Func
● Versioning and auto-Packaging Func
● User Dependent Personal Data
Pseudonym Func
Research Data Repository
Private Shared Public
RDM Platform
Discovery Service
International
Metadata
Aggregator
Storage Area for Long-term Preservation
Re-use
Metadata Aggregation
Exp Data
User Flow
Data Flow
for Data
for Data
Research Data Mng User Interface
26. Summary
• Interoperability is the key for global scholarly communication
• Participate the global activity
• Bridge the gap between the global and local situations
• ID Level
• Metadata Level
• Data Level
• System Level
26
https://ja.wikipedia.org/wiki/%E3%83%A9%E3%82%A6%E3%83%B3%E3%83%89%E3%82%A2%E3%83%90%E3%82%A6%E3%83%88#/media/File:Swindon_Magic_Roundabout.svg