Presented at Linked Open Data: current practice in libraries and archives (Cataloguing & Indexing Group in Scotlland 3rd Linked Open Data Conference), Edinburgh, 18 Nov 2013
This document discusses the LODAC project which aims to connect museum data as linked open data. It describes gathering data from 14 Japanese museums and other sources, standardizing the data, and integrating it while maintaining links to original sources. The integrated data includes information on artworks, creators, and museums. Creator data is matched across sources to link artists to their works. The goals are to publish the data as reusable linked open data, create connections between cultural heritage information and other domains, and allow for new discoveries and uses of museum data through open connectivity.
Adoption of RDA DTR and PID in Deep Carbon Observatory Data PortalXiaogang (Marshall) Ma
The Deep Carbon Observatory (DCO) community is building a cyber-enabled platform for linked science, made available to the community by a multi-institutional data portal. Persistent identifiers and domain specific data types have been identified as key technological issues the portal must address. This presentation focuses on the DCO portal’s planned adoption of RDA DTR and PID methodologies and technologies as a means to address the DCO community's need for persistently identifiable and understandable data type information.
DOI registration with DataCite - COOPEUS, ENVRI, EUDAT workshop 2013Frauke Ziedorn
DataCite is an international organization that provides Digital Object Identifiers (DOIs) to help researchers cite and discover research data on the internet. It was founded in 2009 and now has 17 members from 12 countries. The German National Library of Science and Technology (TIB) was the first DOI registration agency for research data in 2005. DataCite registers DOIs for research data, non-textual materials, and grey literature. It requires clients to commit to data persistence, provide metadata, and landing pages to make data discoverable through the DOI system.
The document discusses open data and challenges with publishing open data. It introduces Entryscape Catalog as a solution for easily, explicitly, and quickly publishing open data through intuitive interfaces with minimum manual work. Entryscape Catalog allows describing data through standard-based forms, publishing data one item at a time or all at once, uploading existing non-open data, and creating APIs from tabular data with a click.
Metadata is data that describes other data. It provides information about digital documents and other content to help identify, manage, retrieve, and understand their usage. There are different types of metadata including administrative, descriptive, structural, technical, and preservation metadata. Common metadata standards include Dublin Core, MARC, and Schema.org which define elements and schemas for describing content in libraries, databases, and on websites.
IP LodB project (for more details see iplod.io ) capitalizes on LOD database thinking, to build bridges between patented information and scientific knowledge, whilst focusing on individuals who codify new knowledge and their connected organizations, including those who apply patents in new products and services.
As main outputs the IP LodB produced an intellectual property rights (IPR) linked open data (LOD) map (IP LOD map), and has tested the linkability of the European patent (EP) LOD database, whilst increasing the uniqueness of data using different harmonization techniques.
These slides were developed for NIPO workshop
Alphabet soup: CDM, VRA, CCO, METS, MODS, RDF - Why Metadata MattersNew York University
This presentation given to University of Iowa Libraries on Nov. 17, 2014, discussing 1) the alphabet soup of metadata standards, e.g. CDM, VRA, CCO, METS, MODS, RDF, including sample tagging and their applications for digital libraries, and 2) why metadata matters. It does not address metadata issues and tools for metadata creation, extraction, transformation, quality control, syndication and ingest.
This document discusses the LODAC project which aims to connect museum data as linked open data. It describes gathering data from 14 Japanese museums and other sources, standardizing the data, and integrating it while maintaining links to original sources. The integrated data includes information on artworks, creators, and museums. Creator data is matched across sources to link artists to their works. The goals are to publish the data as reusable linked open data, create connections between cultural heritage information and other domains, and allow for new discoveries and uses of museum data through open connectivity.
Adoption of RDA DTR and PID in Deep Carbon Observatory Data PortalXiaogang (Marshall) Ma
The Deep Carbon Observatory (DCO) community is building a cyber-enabled platform for linked science, made available to the community by a multi-institutional data portal. Persistent identifiers and domain specific data types have been identified as key technological issues the portal must address. This presentation focuses on the DCO portal’s planned adoption of RDA DTR and PID methodologies and technologies as a means to address the DCO community's need for persistently identifiable and understandable data type information.
DOI registration with DataCite - COOPEUS, ENVRI, EUDAT workshop 2013Frauke Ziedorn
DataCite is an international organization that provides Digital Object Identifiers (DOIs) to help researchers cite and discover research data on the internet. It was founded in 2009 and now has 17 members from 12 countries. The German National Library of Science and Technology (TIB) was the first DOI registration agency for research data in 2005. DataCite registers DOIs for research data, non-textual materials, and grey literature. It requires clients to commit to data persistence, provide metadata, and landing pages to make data discoverable through the DOI system.
The document discusses open data and challenges with publishing open data. It introduces Entryscape Catalog as a solution for easily, explicitly, and quickly publishing open data through intuitive interfaces with minimum manual work. Entryscape Catalog allows describing data through standard-based forms, publishing data one item at a time or all at once, uploading existing non-open data, and creating APIs from tabular data with a click.
Metadata is data that describes other data. It provides information about digital documents and other content to help identify, manage, retrieve, and understand their usage. There are different types of metadata including administrative, descriptive, structural, technical, and preservation metadata. Common metadata standards include Dublin Core, MARC, and Schema.org which define elements and schemas for describing content in libraries, databases, and on websites.
IP LodB project (for more details see iplod.io ) capitalizes on LOD database thinking, to build bridges between patented information and scientific knowledge, whilst focusing on individuals who codify new knowledge and their connected organizations, including those who apply patents in new products and services.
As main outputs the IP LodB produced an intellectual property rights (IPR) linked open data (LOD) map (IP LOD map), and has tested the linkability of the European patent (EP) LOD database, whilst increasing the uniqueness of data using different harmonization techniques.
These slides were developed for NIPO workshop
Alphabet soup: CDM, VRA, CCO, METS, MODS, RDF - Why Metadata MattersNew York University
This presentation given to University of Iowa Libraries on Nov. 17, 2014, discussing 1) the alphabet soup of metadata standards, e.g. CDM, VRA, CCO, METS, MODS, RDF, including sample tagging and their applications for digital libraries, and 2) why metadata matters. It does not address metadata issues and tools for metadata creation, extraction, transformation, quality control, syndication and ingest.
A Higher-Order Data Flow Model for Heterogeneous Big DataSimon Price
Paper presentation at IEEE Big Data 2013 conference, Santa Clara, California. We introduce a data flow model that supports highly parallelisable design patterns and also has useful properties for analysing data serially over extended time periods without requiring traditional Big Data computing facilities.
The model ranges over a class of higher-order relations which are sufficiently expressive to represent a wide variety of unstructured, semi-structured and structured data. Using JSONMatch, our web service implementation of the model, we show that the combination of this model and higher-order representation provides a powerful and extensible framework that is particularly well suited to analysing Big Variety data in a web application context.
The document discusses the ten key principles of Cataloging Cultural Objects (CCO) for cataloging objects in museum collections. It focuses on Principle 1, which is to establish logical relationships between work, collection, and image records. An example is provided of a stereoscope card depicting the Taj Mahal that would have collection, work, and image records linked together. Relationships can be indicated through record links or controlled vocabularies. Principle 2 is also covered, which is to include all required CCO elements in records, such as title, creator, and date. CCO elements are aligned with standards like VRA Core.
LIS 653, Session 4-B: Introduction to Descriptive Metadata Dr. Starr Hoffman
This document provides an overview of descriptive metadata. It defines metadata as "data about data" and discusses common metadata formats like MARC and Dublin Core. It explains how metadata standards, controlled vocabularies, encoding rules, and schemas all work together to structure descriptive data about resources. Specific metadata types are also outlined, including descriptive, administrative, technical, and use metadata.
Dekho queries allow users to pull information from various data sources to display relevant information to end users. There are different types of queries like search, result viewing, and integration queries. Best practices include optimizing queries for performance, selecting only necessary fields, and using multiple queries to provide relevant information to users without overloading them. Queries need to be defined in Dekho Studio and can include spatial elements to filter results based on selections in maps.
The document discusses different methods for embedding semantic information in HTML pages, including microformats which use standard HTML tags and attributes to describe information, and RDFa which uses XML attributes to carry metadata. It provides examples of using microformats and RDFa to add semantic annotations and explores tools for processing and consuming these semantic data formats. The conclusion questions which technology, microformats or RDFa, web developers will embrace more in the future for adding semantics to web pages.
Introduction | Categories for Description of Works of Art | CDWA-LITE Kymberly Keeton
The CDWA-Lite XML schema was created by three organizations to describe cultural works and their related information. It aims to provide core descriptive metadata for works while allowing flexibility in the level of detail provided. The schema includes 22 elements, with 19 for descriptive metadata, 3 for administrative metadata, and 9 required elements. It utilizes controlled vocabularies, XML syntax rules, and semantic elements. Records encoded in CDWA-Lite can be harvested through the Open Archives Initiative protocol.
The document discusses plans by the Japan Link Center (JaLC) to expand its Digital Object Identifier (DOI) registration services. Currently JaLC registers DOIs for journal articles and will soon add additional content types like books, theses, reports, and research data. To gain experience with registering DOIs for research data, JaLC will conduct an experimental project involving participant organizations. The project aims to establish workflows for stable research data DOI registration and integration with DataCite standards. Testing is scheduled to begin in the fall. The addition of data and other content will help JaLC further its goal of supporting all researcher activities through persistent identification with DOIs.
This document provides an introduction to non-relational (NOSQL) databases. It discusses object-based databases and storing objects in SQL databases. It also covers different NOSQL data models like Neo4j which is a graph database, and MongoDB which is a document-oriented database. The document describes some key differences between SQL and MongoDB concepts like databases, tables, collections, documents, and indexes. It provides an example of embedding documents and linking in MongoDB as an alternative to table joins in SQL.
Introducción a la web semántica - Linkatu - irekia 2012Alberto Labarga
This document provides an introduction to linked data journalism and the use of semantic web technologies like RDF, SPARQL, and linked open data. It explains key concepts such as using RDF triples to represent knowledge, querying linked data with SPARQL, and examples of data sources like DBpedia and data marketplaces.
This document discusses the DiJeSt project which aims to transform catalogues and authority files into machine-readable linked open data. It analyzes the Israeli National Library's authority file and bibliography of the Hebrew book, mapping their data using OpenRefine. The project develops an RDF data model using standard ontologies to represent entities like people, places, and books. This linked data can then be used to build applications like eLinda and interactive maps of book data. The vision is to connect this data to other linked open data resources and expand the model to represent additional types of information.
Presented by Howard Goldstein at the Annual Conference of the Visual Resources Association, April 3rd - April 6th, 2013, in Providence, Rhode Island.
Session #4: Every Asset, Everywhere: Perspectives on Digital Asset Management
ORGANIZER / MODERATOR : Elisa Lanzi, Smith College
PRESENTERS:
Howard Goldstein, HR Goldstein Consulting, Digital Imaging Strategies
Elisa Lanzi, Smith College, Director of Digital Strategies and Services
Noah Richman, SRP Phoenix, Media Librarian
Are you hearing the terms “DAM, Digital Preservation, and Media Repositories” more and more at your institution? As our organizations increasingly depend on digital content for all areas of business, the need for enterprise-wide digital asset management is being expressed loud and clear. While cultural institutions are just beginning to implement systems for managing and preserving assets, other media communities (broadcasting, advertising, publishing, etc.) have broad experience in this area. This session brings together panelists for a multi-point perspective on DAM and its impact on the Visual Resources community. Topics include: Digital Asset Management systems demystified; Metadata, taxonomy and DAM; and Digital Preservation. In addition, an open dialogue on VR and organizational change with DAM will be moderated by Lanzi.
The document discusses challenges in preserving and sharing 3D digital objects across changes in file formats and software over time. It describes a case study of a 3D object stored in an AutoCAD DWG file from 1994 that can no longer be opened with modern AutoCAD. The document evaluates options for preservation formats, like IGES and STEP, and dissemination formats, noting that no single format currently supports full feature preservation or translation. It concludes that original files should be kept, multiple preservation formats used, and that open preservation and dissemination formats are still needed to reduce information loss over time.
The document discusses the Names Project, which aims to create a name authority service to help improve metadata and search across institutional repositories. It notes that repositories currently have issues with inconsistent author names leading to incomplete search results. The project conducted scoping work and is building a prototype database based on the FRAD model that assigns unique identifiers to individuals and institutions. This would allow repositories and search services to link information about authors more precisely across repositories.
Entity Linking, Link Prediction, and Knowledge Graph CompletionJennifer D'Souza
A survey presented at the International Winter School on Knowledge Graphs and Semantic Web 2020 http://www.kgswc.org/winter-school/; November 2020; DOI: 10.13140/RG.2.2.12523.77603
This document discusses .NET database objects and how to work with databases in .NET applications. It describes how to create a Windows Forms application with a dataset and table adapter to interact with a SQL database. It explains primary keys, foreign keys, and how to perform common database operations like selecting, inserting, updating and deleting data using .NET objects like the dataset, table adapter, and binding navigator.
Preseted at OR2017 - Brisbane
Panel Discussion: COAR Next Generation Repositories: Results and Recommendations
The presentation focus on the recommended technologies to implement in Repository platforms
The nearly ubiquitous deployment of repository systems in higher education and research institutions provides the foundation for a distributed, globally networked infrastructure for scholarly communication. However, repository platforms are still using technologies and protocols designed almost twenty years ago, before the boom of the Web and the dominance of Google, social networking, semantic web and ubiquitous mobile devices.
To that end, in April 2016, COAR launched a working group to identify the technologies and architectures of the next generation of repositories. There are two threads to our work: (1) increase the exposure by repositories of uniform behaviors that can be used by machine agents to fuel novel scholarly applications that reach beyond the scope of a single repository and that enable to smoothly embed repository content into mainstream web applications. (2) integrate with existing scholarly infrastructures, specifically those aimed at identification, as a means to solidly embed repositories in the overall scholarly communication landscape.
This panel will present the results of the COAR Next Generation Repositories Working Group including our vision, design assumptions, use cases, architectural and technical recommendations, and next steps. The session will also include time for audience discussion and feedback.
RDA implementation, University of St Andrews Cataloguing Department / Iain Ve...CIGScotland
Presented at the RDA for Implementers Conference, 27 May 2015 at the National Library of Scotland, Edinburgh. Organised by the Cataloguing & Indexing Group in Scotland
A Higher-Order Data Flow Model for Heterogeneous Big DataSimon Price
Paper presentation at IEEE Big Data 2013 conference, Santa Clara, California. We introduce a data flow model that supports highly parallelisable design patterns and also has useful properties for analysing data serially over extended time periods without requiring traditional Big Data computing facilities.
The model ranges over a class of higher-order relations which are sufficiently expressive to represent a wide variety of unstructured, semi-structured and structured data. Using JSONMatch, our web service implementation of the model, we show that the combination of this model and higher-order representation provides a powerful and extensible framework that is particularly well suited to analysing Big Variety data in a web application context.
The document discusses the ten key principles of Cataloging Cultural Objects (CCO) for cataloging objects in museum collections. It focuses on Principle 1, which is to establish logical relationships between work, collection, and image records. An example is provided of a stereoscope card depicting the Taj Mahal that would have collection, work, and image records linked together. Relationships can be indicated through record links or controlled vocabularies. Principle 2 is also covered, which is to include all required CCO elements in records, such as title, creator, and date. CCO elements are aligned with standards like VRA Core.
LIS 653, Session 4-B: Introduction to Descriptive Metadata Dr. Starr Hoffman
This document provides an overview of descriptive metadata. It defines metadata as "data about data" and discusses common metadata formats like MARC and Dublin Core. It explains how metadata standards, controlled vocabularies, encoding rules, and schemas all work together to structure descriptive data about resources. Specific metadata types are also outlined, including descriptive, administrative, technical, and use metadata.
Dekho queries allow users to pull information from various data sources to display relevant information to end users. There are different types of queries like search, result viewing, and integration queries. Best practices include optimizing queries for performance, selecting only necessary fields, and using multiple queries to provide relevant information to users without overloading them. Queries need to be defined in Dekho Studio and can include spatial elements to filter results based on selections in maps.
The document discusses different methods for embedding semantic information in HTML pages, including microformats which use standard HTML tags and attributes to describe information, and RDFa which uses XML attributes to carry metadata. It provides examples of using microformats and RDFa to add semantic annotations and explores tools for processing and consuming these semantic data formats. The conclusion questions which technology, microformats or RDFa, web developers will embrace more in the future for adding semantics to web pages.
Introduction | Categories for Description of Works of Art | CDWA-LITE Kymberly Keeton
The CDWA-Lite XML schema was created by three organizations to describe cultural works and their related information. It aims to provide core descriptive metadata for works while allowing flexibility in the level of detail provided. The schema includes 22 elements, with 19 for descriptive metadata, 3 for administrative metadata, and 9 required elements. It utilizes controlled vocabularies, XML syntax rules, and semantic elements. Records encoded in CDWA-Lite can be harvested through the Open Archives Initiative protocol.
The document discusses plans by the Japan Link Center (JaLC) to expand its Digital Object Identifier (DOI) registration services. Currently JaLC registers DOIs for journal articles and will soon add additional content types like books, theses, reports, and research data. To gain experience with registering DOIs for research data, JaLC will conduct an experimental project involving participant organizations. The project aims to establish workflows for stable research data DOI registration and integration with DataCite standards. Testing is scheduled to begin in the fall. The addition of data and other content will help JaLC further its goal of supporting all researcher activities through persistent identification with DOIs.
This document provides an introduction to non-relational (NOSQL) databases. It discusses object-based databases and storing objects in SQL databases. It also covers different NOSQL data models like Neo4j which is a graph database, and MongoDB which is a document-oriented database. The document describes some key differences between SQL and MongoDB concepts like databases, tables, collections, documents, and indexes. It provides an example of embedding documents and linking in MongoDB as an alternative to table joins in SQL.
Introducción a la web semántica - Linkatu - irekia 2012Alberto Labarga
This document provides an introduction to linked data journalism and the use of semantic web technologies like RDF, SPARQL, and linked open data. It explains key concepts such as using RDF triples to represent knowledge, querying linked data with SPARQL, and examples of data sources like DBpedia and data marketplaces.
This document discusses the DiJeSt project which aims to transform catalogues and authority files into machine-readable linked open data. It analyzes the Israeli National Library's authority file and bibliography of the Hebrew book, mapping their data using OpenRefine. The project develops an RDF data model using standard ontologies to represent entities like people, places, and books. This linked data can then be used to build applications like eLinda and interactive maps of book data. The vision is to connect this data to other linked open data resources and expand the model to represent additional types of information.
Presented by Howard Goldstein at the Annual Conference of the Visual Resources Association, April 3rd - April 6th, 2013, in Providence, Rhode Island.
Session #4: Every Asset, Everywhere: Perspectives on Digital Asset Management
ORGANIZER / MODERATOR : Elisa Lanzi, Smith College
PRESENTERS:
Howard Goldstein, HR Goldstein Consulting, Digital Imaging Strategies
Elisa Lanzi, Smith College, Director of Digital Strategies and Services
Noah Richman, SRP Phoenix, Media Librarian
Are you hearing the terms “DAM, Digital Preservation, and Media Repositories” more and more at your institution? As our organizations increasingly depend on digital content for all areas of business, the need for enterprise-wide digital asset management is being expressed loud and clear. While cultural institutions are just beginning to implement systems for managing and preserving assets, other media communities (broadcasting, advertising, publishing, etc.) have broad experience in this area. This session brings together panelists for a multi-point perspective on DAM and its impact on the Visual Resources community. Topics include: Digital Asset Management systems demystified; Metadata, taxonomy and DAM; and Digital Preservation. In addition, an open dialogue on VR and organizational change with DAM will be moderated by Lanzi.
The document discusses challenges in preserving and sharing 3D digital objects across changes in file formats and software over time. It describes a case study of a 3D object stored in an AutoCAD DWG file from 1994 that can no longer be opened with modern AutoCAD. The document evaluates options for preservation formats, like IGES and STEP, and dissemination formats, noting that no single format currently supports full feature preservation or translation. It concludes that original files should be kept, multiple preservation formats used, and that open preservation and dissemination formats are still needed to reduce information loss over time.
The document discusses the Names Project, which aims to create a name authority service to help improve metadata and search across institutional repositories. It notes that repositories currently have issues with inconsistent author names leading to incomplete search results. The project conducted scoping work and is building a prototype database based on the FRAD model that assigns unique identifiers to individuals and institutions. This would allow repositories and search services to link information about authors more precisely across repositories.
Entity Linking, Link Prediction, and Knowledge Graph CompletionJennifer D'Souza
A survey presented at the International Winter School on Knowledge Graphs and Semantic Web 2020 http://www.kgswc.org/winter-school/; November 2020; DOI: 10.13140/RG.2.2.12523.77603
This document discusses .NET database objects and how to work with databases in .NET applications. It describes how to create a Windows Forms application with a dataset and table adapter to interact with a SQL database. It explains primary keys, foreign keys, and how to perform common database operations like selecting, inserting, updating and deleting data using .NET objects like the dataset, table adapter, and binding navigator.
Preseted at OR2017 - Brisbane
Panel Discussion: COAR Next Generation Repositories: Results and Recommendations
The presentation focus on the recommended technologies to implement in Repository platforms
The nearly ubiquitous deployment of repository systems in higher education and research institutions provides the foundation for a distributed, globally networked infrastructure for scholarly communication. However, repository platforms are still using technologies and protocols designed almost twenty years ago, before the boom of the Web and the dominance of Google, social networking, semantic web and ubiquitous mobile devices.
To that end, in April 2016, COAR launched a working group to identify the technologies and architectures of the next generation of repositories. There are two threads to our work: (1) increase the exposure by repositories of uniform behaviors that can be used by machine agents to fuel novel scholarly applications that reach beyond the scope of a single repository and that enable to smoothly embed repository content into mainstream web applications. (2) integrate with existing scholarly infrastructures, specifically those aimed at identification, as a means to solidly embed repositories in the overall scholarly communication landscape.
This panel will present the results of the COAR Next Generation Repositories Working Group including our vision, design assumptions, use cases, architectural and technical recommendations, and next steps. The session will also include time for audience discussion and feedback.
RDA implementation, University of St Andrews Cataloguing Department / Iain Ve...CIGScotland
Presented at the RDA for Implementers Conference, 27 May 2015 at the National Library of Scotland, Edinburgh. Organised by the Cataloguing & Indexing Group in Scotland
RDA development and implementation overview / Gordon DunsireCIGScotland
Presented at the RDA for Implementers Conference, 27 May 2015 at the National Library of Scotland, Edinburgh. Organised by the Cataloguing & Indexing Group in Scotland
The document describes two memory exercises:
1) The Wallet Exercise - Take a digit from your ID or bank card and visualize images representing each digit on a wall to memorize the number.
2) Multiplication Tables Exercise - Pick 14 locations in a room and visualize numbers/images at each to represent the 14 multiplication table, allowing easy recall of multiplication facts up to 14. Visualizing creative images and interacting with them physically helps commit the information to memory.
The document provides background information on Charles Dickens and his works. It discusses the Victorian period in England during which Dickens wrote. It then summarizes the plots and characters of three of Dickens' novels: A Tale of Two Cities, David Copperfield, and the social criticism found in his later works as England underwent a deep economic crisis. The document uses tables and lists to concisely outline the key people and events in the novels.
This document describes the career opportunities and fellowship training available in pediatric infectious diseases. It discusses what a career as a pediatric infectious disease specialist entails, including treating various infections and allocating time to patient care, research, and teaching. The fellowship program at Baylor College of Medicine and Texas Children's Hospital is designed to provide extensive clinical experience over three years, as well as training in research, quality improvement, and professional development. Graduates of this program have diverse career opportunities in pediatric infectious diseases.
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive function. Exercise causes chemical changes in the brain that may help protect against mental illness and improve symptoms.
The document summarizes highlights from the 2015 Chicago Auto Show, including new vehicles showcased by various automakers. Fiat Chrysler unveiled the 2016 Ram Laramie Limited pickup truck. Ford showed the updated 2016 Police Interceptor. General Motors introduced the facelifted 2016 Chevrolet Equinox crossover. Honda unveiled refreshed 2016 Acura RDX and the new 2016 Honda Pilot and Ridgeline. Hyundai-Kia displayed the Trail'ster electric concept vehicle. Mitsubishi presented the GC-PHEV plug-in hybrid concept. Toyota previewed the redesigned 2016 Toyota Avalon sedan.
A pesar de las dificultades que enfrenta el país, los colombianos deben ser optimistas y apoyar a su nación. Colombia tiene muchas cosas por las que vale la pena sentirse orgulloso, como su música, deportes, comida y paisajes naturales. Los colombianos son personas resilientes que se levantan cada vez que caen.
This document lists and describes the top 10 smartphones of 2013. It provides details on the operating systems, processors, storage capacities, cameras and other key specifications of phones such as the Nokia Lumia 920, Samsung Galaxy Note II, Google Nexus 4, Samsung Galaxy S3, HTC First, LG Optimus G, iPhone 5, Samsung Galaxy S4, and Sony Xperia Z. It names the HTC One as the number one smartphone of the year, highlighting its quad-core processor, Android operating system, 32GB/64GB storage, 2GB RAM, 4G connectivity, 13MP camera and other premium specifications.
This document provides information on various topics for learning English using media. It discusses speaking, pronunciation, vocabulary, British and American English, visual dictionary, reading, writing, grammar, and MindMaple which includes grammar, listening and reading. For each topic it provides a brief definition and example to illustrate the concept. The overall document aims to outline resources and methods for improving English language skills using different media tools.
Crowdsourced Georeferencing for Map Library Collections / Chris Fleet, Senior...CIGScotland
Chris Fleet describes the role crowdwourcing can play in providing accurate georeferencing for historic map collections online. Presented at the 6th annual Metadata & Web 2.0 seminar organised by the Cataloguing and Indexing Group in Scotland, held at the National Library of Scotland, 21 June 2013
Integrating Heterogeneous Data Sources in the Web of DataFranck Michel
These are the slides of a 40mn presentation I've made at the CNRS Software Development days (JDEV 2017), in Marseille (France), July 5th, 2017.
Here is the Webcast, in French: https://webcast.in2p3.fr/videos-integrer_des_sources_de_donnees_heterogenes_dans_le_web_de_donnees
Multimedia Data Navigation and the Semantic Web (SemTech 2006)Bradley Allen
The document describes a system for faceted navigation of multimedia content using semantic web technologies. It discusses using ontologies expressed in RDF(S) and OWL to represent metadata, BBC rush footage used as a case study, and visual facets for color, texture and combinations that were generated through MPEG-7 feature extraction and self-organizing map clustering. The system allows retrieval of clips and shots based on textual and visual facet filtering of the RDF represented multimedia data.
Mapping cross-domain metadata to the Europeana Data Model (EDM) - EDM introd...Valentine Charles
- The document introduces the Europeana Data Model (EDM), which was created to allow Europeana to ingest metadata from various sources and domains while maintaining granularity and semantics.
- EDM uses standards like Dublin Core, CIDOC-CRM, and RDF to distinguish cultural heritage objects from their representations and metadata, and to represent relationships between objects and contextual information.
- EDM profiles allow communities to build on EDM to meet their specific needs while maintaining interoperability, and it has been adopted by projects beyond Europeana seeking interoperable metadata.
This document discusses how ISBD (International Standard Bibliographic Description) metadata can be expressed using semantic web standards like RDF, URIs, and namespaces to make the metadata machine-readable. Key points include: metadata is expressed as subject-predicate-object triples using URIs; namespaces provide bases for URIs and can be abbreviated; common vocabularies define predicates and properties for labels, scopes, etc.; and an application profile specifies how a useful "record" can be constructed by defining which triples from which namespaces are involved, their sequence, repeatability, and mandatory status.
20160818 Semantics and Linkage of Archived Catalogsandrea huang
1. The document discusses representing archive catalog data as linked data using semantic web technologies. It involves mapping catalog metadata from XML and CSV formats to RDF and linking to external vocabularies.
2. A system is presented that converts archive catalogs to linked data, stores it using CKAN and provides SPARQL querying. It allows browsing catalog records, performing spatial and temporal queries.
3. An ontology called voc4odw is introduced for organizing open data. It is based on the R4R ontology and aims to semantically enrich catalog records by linking objects, events, places and times using common vocabularies.
Presentation at ELAG 2011, European Library Automation Group Conference, Prague, Czech Republic. 25th May 2011
http://elag2011.techlib.cz/en/815-lifting-the-lid-on-linked-data/
Presentation slide for this:
Kei Kurakawa, Toward universal information access on the digital object cloud, In book of abstracts of International Workshop on Data Science - Present & Future of Open Data & Open Science -, p.57-59, November 12-15, 2018, Mishima Citizens Cultural Hall & Joint Support-Center for Data Science Research, Mishima, Shizuoka, Japan
Beyond Kaggle: Solving Data Science Challenges at ScaleTuri, Inc.
This document summarizes a presentation on entity resolution and data deduplication using Dato toolkits. It discusses key concepts like entity resolution, challenges in entity resolution like missing data and data integration from multiple sources, and provides an example dataset of matching Amazon and Google products. It also outlines the preprocessing steps, describes using a nearest neighbors algorithm to find duplicate records, and shares some resources on entity resolution.
RO-Crate: packaging metadata love notes into FAIR Digital ObjectsCarole Goble
Abstract
slides available at: https://zenodo.org/record/7147703#.Y7agoxXP2F4
The Helmholtz Metadata Collaboration aims to make the research data [and software] produced by Helmholtz Centres FAIR for their own and the wider science community by means of metadata enrichment [1]. Why metadata enrichment and why FAIR? Because the whole scientific enterprise depends on a cycle of finding, exchanging, understanding, validating, reproducing), integrating and reusing research entities across a dispersed community of researchers.
Metadata is not just “a love note to the future” [2], it is a love note to today’s collaborators and peers. Moreover, a FAIR Commons must cater for the metadata of all the entities of research – data, software, workflows, protocols, instruments, geo-spatial locations, specimens, samples, people (well as traditional articles) – and their interconnectivity. That is a lot of metadata love notes to manage, bundle up and move around. Notes written in different languages at different times by different folks, produced and hosted by different platforms, yet referring to each other, and building an integrated picture of a multi-part and multi-party investigation. We need a crate!
RO-Crate [3] is an open, community-driven, and lightweight approach to packaging research entities along with their metadata in a machine-readable manner. Following key principles - “just enough” and “developer and legacy friendliness - RO-Crate simplifies the process of making research outputs FAIR while also enhancing research reproducibility and citability. As a self-describing and unbounded “metadata middleware” framework RO-Crate shows that a little bit of packaging goes a long way to realise the goals of FAIR Digital Objects (FDO)[4], and to not just overcome platform diversity but celebrate it while retaining investigation contextual integrity.
In this talk I will present the why, and how Research Object packaging eases Metadata Collaboration using examples in big data and mixed object exchange, mixed object archiving and publishing, mass citation, and reproducibility. Some examples come from the HMC, others from EOSC, USA and Australia, and from different disciplines.
Metadata is a love note to the future, RO-Crate is the delivery package.
[1] https://helmholtz-metadaten.de/en
[2] Scott, Jason The Metadata Mania, http://ascii.textfiles.com/archives/3181, June 2011
[3] Soiland-Reyes, Stian et al. “Packaging Research Artefacts with RO-Crate”. Data Science, 2022; 5(2):97-138, DOI: 10.3233/DS-210053
[4] De Smedt K, Koureas D, Wittenburg P. “FAIR Digital Objects for Science: From Data Pieces to Actionable Knowledge Units”. Publications. 2020; 8(2):21. https://doi.org/10.3390/publications8020021
The application of linked data concepts and technologies promises great benefits for the retrieval and access of library and cultural heritage resources, but requires new ways of thinking by professionals and end-users. This presentation will use linked data examples from RDA: resource description and access, and will discuss:
- A basic introduction to data structures: triples, chains, and clusters
- What is a linked data record?
- Global, multilingual linked data
- Linking data from multiple sources
- The Semantic Web: a paradigm shift?
Linked Open Data Fundamentals for Libraries, Archives and Museumstrevorthornton
This document provides an overview of linked open data concepts for libraries, archives, and museums. It discusses what linked open data is, potential benefits for cultural institutions, and technical concepts like URIs, HTTP, RDF, ontologies, and SPARQL. The document also covers publishing linked open data by establishing URIs for resources and using content negotiation. Trust and attribution of linked data sources are addressed. Open data licensing, including options from Creative Commons, is also summarized.
Data Equivalence
Mark Parsons, Lead Project Manager, Senior Associate Scientist, National Snow and Ice Data Center
Data citation, especially using persistent identifiers like Digital Object Identifiers (DOIs), is an increasingly accepted scientific practice. Recently, several, respected organizations have developed guidelines for data citation. The different guidelines are largely congruent in that they agree on the basic practice and elements of data citation, especially for relatively static, whole data collections. There is less agreement on the more subtle nuances of data citation that are sometimes necessary to ensure precise reference and scientific reproducibility--the core purpose of data citation. We need to be sure that if you follow a data reference you get to the precise data that were used or at least their scientific equivalent. Identifiers such as DOIs are necessary but not sufficient for the precise, detailed, references necessary. This talk discusses issues around data set versioning, micro-citation, and scientific equivalence. I propose some interim solutions and suggest research strategies for the future.
This document provides an overview of the Resource Description Framework (RDF). It begins with background information on RDF including URIs, URLs, IRIs and QNames. It then describes the RDF data model, noting that RDF is a schema-less data model featuring unambiguous identifiers and named relations between pairs of resources. It also explains that RDF graphs are sets of triples consisting of a subject, predicate and object. The document also covers RDF syntax using Turtle and literals, as well as modeling with RDF. It concludes with a brief overview of common RDF tools including Jena.
Panel presentation to a graduate class at the University of Arizona School of Information Resources and Library Science. Invited by Dr. Jana Bradley. July 2006.
The students in an LS 566 Metadata class were asked to index a collection of six images from the 1992 University of Alabama Crimson Tide football season. The class chose a modified Dublin Core metadata schema containing 15 elements to describe the images. Students worked in groups to define the elements and map them to the images. They used Omeka software to extract image text and connect files to element fields. The indexing process took longer than expected due to a tornado that damaged the university.
Presentation in the First Workshop on Digital Information Management. The workshop is organized by the Laboratory on Digital Libraries and Electronic Publication, Department of Archives and Library Sciences, Ionian University, Greece and aims to create a venue for unfolding research activity on the general field of Information Science. The workshop features sessions for the dissemination of the research results of the Laboratory members, as well as tutorial sessions on interesting issues.
Similar to How to publish local metadata as linked data / Gordon Dunsire (20)
From Scottish Bibliographies Online to National Bibliography of Scotland : Re...CIGScotland
The National Library of Scotland presented on reinventing the National Bibliography of Scotland for the 21st century. They discussed the history of bibliographies in Scotland. They conducted a review which found a need to align the bibliography with their strategy and open data principles. After stakeholder consultation, they are implementing the National Bibliography of Scotland as an extension of their catalog, leveraging new technologies and linked open data approaches to maintain it as Scotland's published cultural record.
The future of cataloguing: a CIGS World Cafe WorkshopCIGScotland
Alan Danskin from the British Library gives a scene setting presentation, including the history of cataloguing, the longevity of good metadata, and the British Library's Metadata Strategy (2019-23). This put everyone in the right frame of mind before the interactive discussion workshop on the Future of Cataloguing.
Everyone, everywhere, everything : then, now and the future / Gill Hamilton, ...CIGScotland
Review of the changing face of cataloguing and resource discovery over the years, and a vision of the future. Presented at the CIG Scotland seminar 'Resource Discovery : from catalogues to discovery services' at the National Library of Scotland, Edinburgh, 21st March 2018
What do you want to discover today? / Janet Aucock, University of St AndrewsCIGScotland
Overview of resource discovery in libraries today. Presented at the CIG Scotland seminar 'Resource Discovery : from catalogues to discovery services' at the National Library of Scotland, Edinburgh, 21st March 2018
From student to graduate trainee : a user perspective / Liz Antel, Graduate L...CIGScotland
An insight into the user experience of modern library resource discovery. Presented at the CIG Scotland seminar 'Resource Discovery : from catalogues to discovery services' at the National Library of Scotland, Edinburgh, 21st March 2018.
"We want something like Google ... why do we get so many results?" : implemen...CIGScotland
Description of Durham developed unified resource discovery, and the challenges and rewards of integrating library, archival, museum and archaeological collections.
Presented at the CIG Scotland seminar 'Resource Discovery : from catalogues to discovery services' at the National Library of Scotland, Edinburgh, 21st March 2018
Researching the user experience of the National Library of Scotland eResource...CIGScotland
Research into how users discover, use and perceive the Library's collection of e-resources. Presented at the CIG Scotland seminar 'Resource Discovery : from catalogues to discovery services' at the National Library of Scotland, Edinburgh, 21st March 2018
Where did you come from, where will you go? : bibliographic data and union ca...CIGScotland
Introduction to the development of the National Bibliographic Knowledgebase. Presented at the CIG Scotland seminar 'Resource Discovery : from catalogues to discovery services' at the National Library of Scotland, Edinburgh, 21st March 2018
Engaging the crowd : old hands, modern minds : evolving an on-line manuscript...CIGScotland
Presented at the CIG Scotland seminar 'Somewhere over the Rainbow: our metadata online, past, present & future' (Metadata & Web 2.0 Series) at the National Library of Scotland, Edinburgh, 5th April 2017
5Rights : enabling children and young people to access the digital world crea...CIGScotland
Young Scot, Scottish Government and 5Rights introduce Scotland's 5Rights Youth Commission - a diverse group of young people passionate about their digital rights. We will hear from two young people what their "5Rights" mean to them, and how children and young people can be empowered to access technology creatively, knowledgeably, and fearlessly.
Presented at the CIG Scotland seminar 'Somewhere over the Rainbow: our metadata online, past, present & future' (Metadata & Web 2.0 Series) at the National Library of Scotland, Edinburgh, 5th April 2017
Managing your Digital Footprint : Taking control of the metadata and tracks a...CIGScotland
Find out how personal metadata, social media posts, and online activity make up an individual's "Digital Footprint", why they matter, and hear some advice on how to better manage digital tracks and traces. Nicola Osborne will draw on recent University of Edinburgh research on students' digital footprints, which is also the subject of the new #DFMOOC free online.
Presented at the CIG Scotland seminar 'Somewhere over the Rainbow: our metadata online, past, present & future' (Metadata & Web 2.0 Series) at the National Library of Scotland, Edinburgh, 5th April 2017
Playing with metadata / Gavin Willshaw, Scott Renton (University of Edinburgh)CIGScotland
Learn about Edinburgh University Library's metadata games platform, a crowdsourcing initiative which has improved descriptive metadata and become a vital engagement tool both within and beyond the library. Hear how they have developed their games in collaboration with Tiltfactor, a Dartmouth College-based research group which explores games design for social change, and learn what they're doing with crowd-sourced data.
Presented at the CIG Scotland seminar 'Somewhere over the Rainbow: our metadata online, past, present & future' (Metadata & Web 2.0 Series) at the National Library of Scotland, Edinburgh, 5th April 2017
How to effectively archive Olympic and Paralympic websites / Helena Byrne (Br...CIGScotland
This document discusses how the British Library effectively archived Olympic and Paralympic websites from 2010-2016 through an international collaboration. Key steps included planning metadata requirements, collecting nominations through a shared spreadsheet, and addressing issues like incomplete URLs. Over 4,800 nominations were received from 125 countries, resulting in a collection of over 4,000 archived websites. Clear communication, cultural awareness, and regular updates were important to the success of the project.
The Statistical Accounts of Scotland / Vivienne Mayo (EDINA)CIGScotland
Vivienne describes how information from the past has found a new lease of life in the recently re-launched Statistical Accounts of Scotland. Presented at the CIG Scotland seminar 'Somewhere over the Rainbow: our metadata online, past, present & future' (Metadata & Web 2.0 Series) at the National Library of Scotland, Edinburgh, 5th April 2017
Dianne Pennington of Strathclyde University will move beyond the bounds of bibliographic description as she discusses her research about emotions shared by music fans online and how they might be used as metadata for new approaches to search and retrieval.
Presented at the CIG Scotland seminar 'Somewhere over the Rainbow : our metadata online, past, present & future' (Metadata & Web 2.0 Series) at the National Library of Scotland, Edinburgh, 5th April 2017
State of Artificial intelligence Report 2023kuntobimo2016
Artificial intelligence (AI) is a multidisciplinary field of science and engineering whose goal is to create intelligent machines.
We believe that AI will be a force multiplier on technological progress in our increasingly digital, data-driven world. This is because everything around us today, ranging from culture to consumer products, is a product of intelligence.
The State of AI Report is now in its sixth year. Consider this report as a compilation of the most interesting things we’ve seen with a goal of triggering an informed conversation about the state of AI and its implication for the future.
We consider the following key dimensions in our report:
Research: Technology breakthroughs and their capabilities.
Industry: Areas of commercial application for AI and its business impact.
Politics: Regulation of AI, its economic implications and the evolving geopolitics of AI.
Safety: Identifying and mitigating catastrophic risks that highly-capable future AI systems could pose to us.
Predictions: What we believe will happen in the next 12 months and a 2022 performance review to keep us honest.
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...sameer shah
"Join us for STATATHON, a dynamic 2-day event dedicated to exploring statistical knowledge and its real-world applications. From theory to practice, participants engage in intensive learning sessions, workshops, and challenges, fostering a deeper understanding of statistical methodologies and their significance in various fields."
The Ipsos - AI - Monitor 2024 Report.pdfSocial Samosa
According to Ipsos AI Monitor's 2024 report, 65% Indians said that products and services using AI have profoundly changed their daily life in the past 3-5 years.
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data LakeWalaa Eldin Moustafa
Dynamic policy enforcement is becoming an increasingly important topic in today’s world where data privacy and compliance is a top priority for companies, individuals, and regulators alike. In these slides, we discuss how LinkedIn implements a powerful dynamic policy enforcement engine, called ViewShift, and integrates it within its data lake. We show the query engine architecture and how catalog implementations can automatically route table resolutions to compliance-enforcing SQL views. Such views have a set of very interesting properties: (1) They are auto-generated from declarative data annotations. (2) They respect user-level consent and preferences (3) They are context-aware, encoding a different set of transformations for different use cases (4) They are portable; while the SQL logic is only implemented in one SQL dialect, it is accessible in all engines.
#SQL #Views #Privacy #Compliance #DataLake
Influence of Marketing Strategy and Market Competition on Business Plan
How to publish local metadata as linked data / Gordon Dunsire
1. How to publish local metadata as
linked data
Gordon Dunsire
Presented at Linked Open Data: current practice
in libraries and archives (3rd Linked Open Data
Conference), Edinburgh, 18 Nov 2013
2. Overview
Choosing and finding elements for publishing
linked data
Losing data; dumbing down
Creating local elements
Case study: NLS DOD
Developing local elements and URIs
Extracting data triples
Demo of proof-of-concept interface
3. Choosing elements
Classes (types) and properties (attributes and
relationships) that represent the context and
scope of linked data
Elements should be available to applications
“Global” element >= local element
[meaning/semantic]
< results in semantic incoherency + low trust/quality
rating
> results in loss of precision, loss of information
“dumbing down” (Dublin Core)
4. Finding elements
Pick and mix
Many element sets for bibliographic data are
available
General: Dublin Core, FOAF, SKOS …
Specific: BibO, FRBR, ISBD, RDA …
With significant overlap …
Shades of meaning = semantic granularity
Searchable registries: joinup, LOV, OMR …
Or grow your own
5. Creating elements
Structure of local records derived from a local
schema
Home grown or modified standard
Simple methodology to define structure as a
local element set
Use data dictionary/documentation as basis
for element labels and definitions
Assign URIs to each element
Make element set available to applications
6. Case study: NLS DOD
Digital Object Database (National Library of
Scotland)
Metadata for digitized resources
Local schema expressed in a relational database
structure
Normalized (3rd form)
Standard method for linked data representation
Standard design template used to derive element
labels
Element set for descriptive metadata published
using Open Metadata Registry (OMR)
8. Extracting data triples
Process the metadata records to output data
values
Values are the objects of data triples with
predicates taken from the local element set
X nlsdod:hasTitle “Remarkable map”
X nlsdod:hasDescription “Map of an unknown island”
…
Subjects of triples are URIs for the individual
thing being described
Derived from database “object” identifier
10. Data triples extracted from the DOD
Subject
Predicate
d:D74547188 tDescription-title
Object
“Anti-aircraft gunner”
d:D74547188 tDescription-description “A soldier is sitting on
the edge of a …”
d:D74547188 hasDateOrEvent
k:D3014
d:D74547188 hasKeyword
k:D3764
k:D3014
tKeyword-keyword
“Ypres, 3rd Battle of,
Ieper, Belgium, 1917”
k:D3764
tKeyword-keyword
“Holes”
Subject and object instance URIs derived
from resource or record identifier
Internal links reflect database structure
11. “Authority” strings assigned local things (URIs)
local:rowURI
local:hasKeywordLabel
“stuff”
global:URI
Local things can then be matched to global things
12. External mappings
Mappings from local to global things stored as
turtle files
Separate from extracted local triples
New mappings can be added ad hoc
When resources available
Prioritised by local usage/focus/context
E.g. “Haig, Douglas, Sir, 1861-1928”
E.g. “Ypres, 3rd Battle of, Ieper, Belgium, 1917”
dbPedia “Battle of Passchendaele”
Maps to external element sets support
interoperability
13. Demo Interface
All data is stored and processed as RDF triples
Local data held in MySQL triple store (ARC2)
External data is accessed live online
PHP script uses ARC2 functions
Everything runs on a laptop, so here are some
screenshots …