Europeana provides access to digital resources from a wide range of cultural heritage institutions all across Europe. In order to support Europeana, a wide network of organizations collaborates in data integration activities. The European Library plays the role of library-domain aggregator for Europeana, and its activities include also being a gateway to the collections and data of Europe’s national and research libraries, operating on the principle of open data for re-use.
The Europeana Network addresses its data integration challenges by leveraging on Linked Data and the Semantic Web. Its approach to data integration is based in a single data model, the Europeana Data Model, which embraces the Semantic Web principles to integrate the various data models and ontologies used in cultural heritage data.
The paradigm of Linked Data, brings many new challenges to libraries. The generic nature of data representation used in Linked Data, while allowing any community to manipulate the data, also opens many paths for implementation, with no clear optimal choice for libraries. The European Library leverages on its operational infrastructure to make library data available. It maintains The European Library Open Dataset, which is derived from the data aggregated from member libraries, and made available under the Creative Commons CC0 1.0 Universal license, in order to promote and facilitate its reuse by any community.
Extensive linking is performed in the preparation of The European Library Open Dataset. It relies on Information Extraction and Data Mining to establish links to external open datasets, covering the most prominent entities types present in library data: persons, corporate bodies, places, concepts, intellectual works and manifestations.
The European Library also applies a linked data approach for intellectual property rights clearance processes, for supporting mass digitization projects. This approach is applied in the within the European ARROW rights infrastructure .
DanteSources is a focused digital library endowed with web services that allow visualizing information on Dante Alighieri’s primary sources in form of charts and tables. The visualized charts can be exported in various well-known formats like PDF and JPEG, but the data can be also exported in CSV format, to lend them to further analyses. DanteSources makes information about Dante’s primary sources available in digital format for the first time. Having the information about primary sources dispersed on paper books makes it difficult to
systematically overview how the cultural background of Dante evolved in time. On the other hand, the automatic visualization of data allows understanding the development of Dante’s cultural background in comparison with the different phases of his biography.
This presentation was delivered by Gloria Gonzalez of Zepheira during the NISO Virtual Conference, BIBFRAME & Real World Applications of Linked Bibliographic Data, held on June 15, 2016.
Günter Mühlberger (University of Innsbruck, AT): The READ project. Objectives, tasks and partner organisations
co:op-READ-Convention Marburg
Technology meets Scholarship, or how Handwritten Text Recognition will Revolutionize Access to Archival Collections.
With a special focus on biographical data in archives
Hessian State Archives Marburg Friedrichsplatz 15, D - 35037 Marburg
19-21 January 2016
An approach to identify how much a Linked Data dataset is biased, using statistical methods and the links between datasets. 28/11/2014 @EKAW2014, Linköping, Sweden
Bridging Informal MOOCs & Formal English for Academic Purposes Programmes wit...Alannah Fitzgerald
Presented at the Teaching and Language Corpora (TaLC) Conference in Lancaster on July 23, 2014. Based on collaborative work with the FLAX Language Project (Shaoqun Wu and Ian Witten) and the Language Centre at Queen Mary University of London (Martin Barge, William Tweddle, Saima Sherazi).
DanteSources is a focused digital library endowed with web services that allow visualizing information on Dante Alighieri’s primary sources in form of charts and tables. The visualized charts can be exported in various well-known formats like PDF and JPEG, but the data can be also exported in CSV format, to lend them to further analyses. DanteSources makes information about Dante’s primary sources available in digital format for the first time. Having the information about primary sources dispersed on paper books makes it difficult to
systematically overview how the cultural background of Dante evolved in time. On the other hand, the automatic visualization of data allows understanding the development of Dante’s cultural background in comparison with the different phases of his biography.
This presentation was delivered by Gloria Gonzalez of Zepheira during the NISO Virtual Conference, BIBFRAME & Real World Applications of Linked Bibliographic Data, held on June 15, 2016.
Günter Mühlberger (University of Innsbruck, AT): The READ project. Objectives, tasks and partner organisations
co:op-READ-Convention Marburg
Technology meets Scholarship, or how Handwritten Text Recognition will Revolutionize Access to Archival Collections.
With a special focus on biographical data in archives
Hessian State Archives Marburg Friedrichsplatz 15, D - 35037 Marburg
19-21 January 2016
An approach to identify how much a Linked Data dataset is biased, using statistical methods and the links between datasets. 28/11/2014 @EKAW2014, Linköping, Sweden
Bridging Informal MOOCs & Formal English for Academic Purposes Programmes wit...Alannah Fitzgerald
Presented at the Teaching and Language Corpora (TaLC) Conference in Lancaster on July 23, 2014. Based on collaborative work with the FLAX Language Project (Shaoqun Wu and Ian Witten) and the Language Centre at Queen Mary University of London (Martin Barge, William Tweddle, Saima Sherazi).
FLAX: Flexible Language Acquisition with Open Data-Driven LearningAlannah Fitzgerald
Presented at the International Semantic Web Conference (ISWC) in Riva del Garda, Trentino, Italy as part of the Vici LinkedUp Challenge for tools and demos that use open data for educational purposes on October 20th, 2014.
Sebastian Colutto (University of Innsbruck, AT): Transkribus. A virtual research environment for the transcription and recognition of historical documents
co:op-READ-Convention Marburg
Technology meets Scholarship, or how Handwritten Text Recognition will Revolutionize Access to Archival Collections.
With a special focus on biographical data in archives
Hessian State Archives Marburg Friedrichsplatz 15, D - 35037 Marburg
19-21 January 2016
Basilis Gatos (Computational Intelligence Laboratory, Institute of Informatics and Telecommunications, National Center for Scientific Research “Demokritos”, GR): Hard Tasks in the Background. Layout analysis
co:op-READ-Convention Marburg
Technology meets Scholarship, or how Handwritten Text Recognition will Revolutionize Access to Archival Collections.
With a special focus on biographical data in archives
Hessian State Archives Marburg Friedrichsplatz 15, D - 35037 Marburg
19-21 January 2016
These slides review the Europeana Libraries project, a 2011-2 project funded by the European Commission in the field of digital libraries. The project sought to develop the structure of The European Library, and help aggregate metadata from research libraries in Europe.
What is the state of natural language processing for Danish in 2018? This reviews language technology in Denmark this year. Present at a "Puzzle of Danish" workshop.
Dissertation as a document provides data on new knowledge, but also – encodes important scientometrical information. A study of social structure of science through data found in dissertations and theses provides bibliometrical data for study of national style of science. The pilot study, described below encourages the library community to improve their documentation in this area, in particular, the notation of supervisors and institutions within a bibliographical record. It is proposed that the CBD argue for stricter standards of library/archive record of dissertation.
Along with the dissertation data in the new IsisCB platform the social structure of the history of science community might be analysed. Scientometrical study of dissertation abstracts at a local level (ie., Lithuania) will provide a model for future studies of scholarly communication at global level.
LIBER Webinar: Research Data Services Survey LIBER Europe
These are the slides that accompanied the LIBER Webinar on Research Data Services survey (1 December 2016). Find out more about the survey here: http://libereurope.eu/blog/2016/10/13/research-data-services-europes-academic-research-libraries
An overview of work currently being done by the Digital Manuscript Technology group. This presentation was given to the 2013 CLIR fellows in medieval data curation, and is a synthesis of earlier presentations, some of which were co-authored with Robert Sanderson.
Slides for Culture Hack panel @SXSW2013 : http://schedule.sxsw.com/2013/events/event_IAP4580
Some slides re-used from Harry Verwayen (http://www.slideshare.net/hverwayen/business-model-innovation-open-data) and Julia Fallon
Visualizing the Transcribe Bentham Corpus
Frédérique Mélanie, Estelle Tieberghien, Pablo Ruiz Fabo,
Thierry Poibeau
LATTICE Lab: ENS – CNRS – U Paris 3, PSL – USPC
Tim Causer, Melissa Terras
UCL Bentham Project, UCL Digital Humanities
UCLDH Seminar, December 2016
Wikidata, a target for Europeana's semantic strategy - GLAM-WIKI 2015Antoine Isaac
"Wikidata, a target for Europeana's semantic strategy"/ Presentation at the GLAM-Wiki conference with Valentine Charles, Hugo Manguinhas, Antoine Isaac, Vladimir Alexiev http://nl.wikimedia.org/wiki/GLAM-WIKI_2015/
Introduction to the Europeana hackathon in PoznanDavid Haskiya
My introductory slides for the Poznan hackathon. Covering the hackathon roadshow, its links to the EU Digital Agenda, the Europeana API, the Europeana Linked Open Data Pilot and the kind of content we have in Europeana
Word Occurrence Based Extraction of Work Contributors from Statements of Resp...The European Library
This paper addresses the identification of all contributors of an intellectual work, when they are recorded in bibliographic data but in unstructured form. National bibliographies are very reliable on representing the first author of a work, but frequently, secondary contributors are represented in the statements of responsibility that are transcribed by the cataloguer from the book into the bibliographic records. The identification of work contributors mentioned in statements of responsibility is a typical motivation for the application of information extraction techniques. This paper presents an approach developed for the specific application scenario of the ARROW rights infrastructure being deployed in several European countries to assist in the determination of the copyright status of works that may not be under public domain. Our approach performed reliably in most languages and bibliographic datasets of at least one million records, achieving precision and recall above 0.97 on five of the six evaluated datasets. We conclude that the approach can be reliably applied to other national bibliographies and languages.
FLAX: Flexible Language Acquisition with Open Data-Driven LearningAlannah Fitzgerald
Presented at the International Semantic Web Conference (ISWC) in Riva del Garda, Trentino, Italy as part of the Vici LinkedUp Challenge for tools and demos that use open data for educational purposes on October 20th, 2014.
Sebastian Colutto (University of Innsbruck, AT): Transkribus. A virtual research environment for the transcription and recognition of historical documents
co:op-READ-Convention Marburg
Technology meets Scholarship, or how Handwritten Text Recognition will Revolutionize Access to Archival Collections.
With a special focus on biographical data in archives
Hessian State Archives Marburg Friedrichsplatz 15, D - 35037 Marburg
19-21 January 2016
Basilis Gatos (Computational Intelligence Laboratory, Institute of Informatics and Telecommunications, National Center for Scientific Research “Demokritos”, GR): Hard Tasks in the Background. Layout analysis
co:op-READ-Convention Marburg
Technology meets Scholarship, or how Handwritten Text Recognition will Revolutionize Access to Archival Collections.
With a special focus on biographical data in archives
Hessian State Archives Marburg Friedrichsplatz 15, D - 35037 Marburg
19-21 January 2016
These slides review the Europeana Libraries project, a 2011-2 project funded by the European Commission in the field of digital libraries. The project sought to develop the structure of The European Library, and help aggregate metadata from research libraries in Europe.
What is the state of natural language processing for Danish in 2018? This reviews language technology in Denmark this year. Present at a "Puzzle of Danish" workshop.
Dissertation as a document provides data on new knowledge, but also – encodes important scientometrical information. A study of social structure of science through data found in dissertations and theses provides bibliometrical data for study of national style of science. The pilot study, described below encourages the library community to improve their documentation in this area, in particular, the notation of supervisors and institutions within a bibliographical record. It is proposed that the CBD argue for stricter standards of library/archive record of dissertation.
Along with the dissertation data in the new IsisCB platform the social structure of the history of science community might be analysed. Scientometrical study of dissertation abstracts at a local level (ie., Lithuania) will provide a model for future studies of scholarly communication at global level.
LIBER Webinar: Research Data Services Survey LIBER Europe
These are the slides that accompanied the LIBER Webinar on Research Data Services survey (1 December 2016). Find out more about the survey here: http://libereurope.eu/blog/2016/10/13/research-data-services-europes-academic-research-libraries
An overview of work currently being done by the Digital Manuscript Technology group. This presentation was given to the 2013 CLIR fellows in medieval data curation, and is a synthesis of earlier presentations, some of which were co-authored with Robert Sanderson.
Slides for Culture Hack panel @SXSW2013 : http://schedule.sxsw.com/2013/events/event_IAP4580
Some slides re-used from Harry Verwayen (http://www.slideshare.net/hverwayen/business-model-innovation-open-data) and Julia Fallon
Visualizing the Transcribe Bentham Corpus
Frédérique Mélanie, Estelle Tieberghien, Pablo Ruiz Fabo,
Thierry Poibeau
LATTICE Lab: ENS – CNRS – U Paris 3, PSL – USPC
Tim Causer, Melissa Terras
UCL Bentham Project, UCL Digital Humanities
UCLDH Seminar, December 2016
Wikidata, a target for Europeana's semantic strategy - GLAM-WIKI 2015Antoine Isaac
"Wikidata, a target for Europeana's semantic strategy"/ Presentation at the GLAM-Wiki conference with Valentine Charles, Hugo Manguinhas, Antoine Isaac, Vladimir Alexiev http://nl.wikimedia.org/wiki/GLAM-WIKI_2015/
Introduction to the Europeana hackathon in PoznanDavid Haskiya
My introductory slides for the Poznan hackathon. Covering the hackathon roadshow, its links to the EU Digital Agenda, the Europeana API, the Europeana Linked Open Data Pilot and the kind of content we have in Europeana
Word Occurrence Based Extraction of Work Contributors from Statements of Resp...The European Library
This paper addresses the identification of all contributors of an intellectual work, when they are recorded in bibliographic data but in unstructured form. National bibliographies are very reliable on representing the first author of a work, but frequently, secondary contributors are represented in the statements of responsibility that are transcribed by the cataloguer from the book into the bibliographic records. The identification of work contributors mentioned in statements of responsibility is a typical motivation for the application of information extraction techniques. This paper presents an approach developed for the specific application scenario of the ARROW rights infrastructure being deployed in several European countries to assist in the determination of the copyright status of works that may not be under public domain. Our approach performed reliably in most languages and bibliographic datasets of at least one million records, achieving precision and recall above 0.97 on five of the six evaluated datasets. We conclude that the approach can be reliably applied to other national bibliographies and languages.
Innovative methods for data integration: Linked Data and NLPariadnenetwork
Linked Data (LD) + Natural Language Processing (NLP)
Two technologies that open up new possibilities for semantic integration of archaeological datasets and fieldwork reports.
Overview
•Illustrative early examples
- a flavour of progress and challenges to date
•NLP of grey literature (English – Dutch)
•Mapping between multilingual vocabularies
Designing a multilingual knowledge graph - DCMI2018Antoine Isaac
Presentation for the paper "Designing a multilingual knowledge graph as service for cultural heritage" at the DCMI2018 conference https://www.dublincore.org/conferences/2018/abstracts/#559
Building an ecosystem of networked referencesHugo Manguinhas
Over the past five years, the amount of contextual entities in Europeana’s metadata has grown considerably. These entities are provided as references as part of the metadata delivered by Europeana or selected by Europeana semantic automatic enrichment. Pursuing their efforts towards the creation of a semantic network around cultural heritage objects, Europeana and its partners providers and aggregators are investigating ways to better exchange vocabulary data and manage co-references/alignments between vocabularies. In this presentation we will explore the potential of tools such as OpenSkos and Cultuurlink for supporting the building of networked references.
Presented at the 6th DBpedia Community Meeting in The Hague 2016, see http://wiki.dbpedia.org/meetings/TheHague2016
Information Extraction from EuroParliament and UK Parliament dataWim Peters
These slides describe the work done at the CLARIN talk of Europe Creative Camp, in which groups from various countries worked with EuroParliament speeches.
Our work covers term extraction, term organisation and term linking between the Europarliament and UK Parliament data sets.
Information Extraction in the TalkOfEurope Creative CampWim Peters
The CLARIN Talk of Europe Creative Camp event in March 2015 invited people to work on the EuroParliament data of the Talk of Europe data set (http://linkedpolitics.ops.few.vu.nl/home)
Our work during that event covers the conceptualization of the content of two data sets:
- English EuroParliament speeches from the Talk of Europe data set and
- UK Parliament speeches.
We performed term extraction, term organisation and the linking of terminology between these two data sets. the results were
Finding the annotation needs of the botanical community in a digital libraryWilliam Ulate
The Center for Biodiversity Informatics at the Missouri Botanical Garden and Saint Louis University are analyzing the web annotation needs of the botanical community to develop a prototype of how those needs may be met within a digital library platform. We want to assess the practicality of existing tools to satisfy the technical, economic, and operational needs of botanical users to annotate. This will inform on requisites, best practices, and further developments for a research project to integrate an annotation tool within a virtual library. We surveyed 14 members of 10 different institutions in the botanical and scientific communities. We included both, those who currently annotate online as well as those who have only annotated offline (e.g. print or analog), in order to better understand the functionality needed to encourage and support online annotation activities. The answers to this survey were analyzed in the context of an annotation tool in a digital library and a prioritized list of annotation needs for users of a botanical virtual library was produced, taking into account the minimal and recommended functionality required to comply with the users requirements. Preliminary results from the report of the in-depth user assessments of annotation needs in the specific domain of botanists are shared with the attendees. Advances in the definition of a prototype are also shown.
Similar to Linked Data and cultural heritage data: an overview of the approaches from Europeana and The European Library (20)
Joining The European Library, Adam Sofronijevic, University of BelgradeThe European Library
Joining The European Library: Members' Perspectives
University of Belgrade, University library "Svetozar Markovic
The European Library Annual Meeting 2013
16-17 September 2013
Amsterdam, The Netherlands
Nutraceutical market, scope and growth: Herbal drug technologyLokesh Patil
As consumer awareness of health and wellness rises, the nutraceutical market—which includes goods like functional meals, drinks, and dietary supplements that provide health advantages beyond basic nutrition—is growing significantly. As healthcare expenses rise, the population ages, and people want natural and preventative health solutions more and more, this industry is increasing quickly. Further driving market expansion are product formulation innovations and the use of cutting-edge technology for customized nutrition. With its worldwide reach, the nutraceutical industry is expected to keep growing and provide significant chances for research and investment in a number of categories, including vitamins, minerals, probiotics, and herbal supplements.
This presentation explores a brief idea about the structural and functional attributes of nucleotides, the structure and function of genetic materials along with the impact of UV rays and pH upon them.
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Sérgio Sacani
We characterize the earliest galaxy population in the JADES Origins Field (JOF), the deepest
imaging field observed with JWST. We make use of the ancillary Hubble optical images (5 filters
spanning 0.4−0.9µm) and novel JWST images with 14 filters spanning 0.8−5µm, including 7 mediumband filters, and reaching total exposure times of up to 46 hours per filter. We combine all our data
at > 2.3µm to construct an ultradeep image, reaching as deep as ≈ 31.4 AB mag in the stack and
30.3-31.0 AB mag (5σ, r = 0.1” circular aperture) in individual filters. We measure photometric
redshifts and use robust selection criteria to identify a sample of eight galaxy candidates at redshifts
z = 11.5 − 15. These objects show compact half-light radii of R1/2 ∼ 50 − 200pc, stellar masses of
M⋆ ∼ 107−108M⊙, and star-formation rates of SFR ∼ 0.1−1 M⊙ yr−1
. Our search finds no candidates
at 15 < z < 20, placing upper limits at these redshifts. We develop a forward modeling approach to
infer the properties of the evolving luminosity function without binning in redshift or luminosity that
marginalizes over the photometric redshift uncertainty of our candidate galaxies and incorporates the
impact of non-detections. We find a z = 12 luminosity function in good agreement with prior results,
and that the luminosity function normalization and UV luminosity density decline by a factor of ∼ 2.5
from z = 12 to z = 14. We discuss the possible implications of our results in the context of theoretical
models for evolution of the dark matter halo mass function.
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Ana Luísa Pinho
Functional Magnetic Resonance Imaging (fMRI) provides means to characterize brain activations in response to behavior. However, cognitive neuroscience has been limited to group-level effects referring to the performance of specific tasks. To obtain the functional profile of elementary cognitive mechanisms, the combination of brain responses to many tasks is required. Yet, to date, both structural atlases and parcellation-based activations do not fully account for cognitive function and still present several limitations. Further, they do not adapt overall to individual characteristics. In this talk, I will give an account of deep-behavioral phenotyping strategies, namely data-driven methods in large task-fMRI datasets, to optimize functional brain-data collection and improve inference of effects-of-interest related to mental processes. Key to this approach is the employment of fast multi-functional paradigms rich on features that can be well parametrized and, consequently, facilitate the creation of psycho-physiological constructs to be modelled with imaging data. Particular emphasis will be given to music stimuli when studying high-order cognitive mechanisms, due to their ecological nature and quality to enable complex behavior compounded by discrete entities. I will also discuss how deep-behavioral phenotyping and individualized models applied to neuroimaging data can better account for the subject-specific organization of domain-general cognitive systems in the human brain. Finally, the accumulation of functional brain signatures brings the possibility to clarify relationships among tasks and create a univocal link between brain systems and mental functions through: (1) the development of ontologies proposing an organization of cognitive processes; and (2) brain-network taxonomies describing functional specialization. To this end, tools to improve commensurability in cognitive science are necessary, such as public repositories, ontology-based platforms and automated meta-analysis tools. I will thus discuss some brain-atlasing resources currently under development, and their applicability in cognitive as well as clinical neuroscience.
Richard's aventures in two entangled wonderlandsRichard Gill
Since the loophole-free Bell experiments of 2020 and the Nobel prizes in physics of 2022, critics of Bell's work have retreated to the fortress of super-determinism. Now, super-determinism is a derogatory word - it just means "determinism". Palmer, Hance and Hossenfelder argue that quantum mechanics and determinism are not incompatible, using a sophisticated mathematical construction based on a subtle thinning of allowed states and measurements in quantum mechanics, such that what is left appears to make Bell's argument fail, without altering the empirical predictions of quantum mechanics. I think however that it is a smoke screen, and the slogan "lost in math" comes to my mind. I will discuss some other recent disproofs of Bell's theorem using the language of causality based on causal graphs. Causal thinking is also central to law and justice. I will mention surprising connections to my work on serial killer nurse cases, in particular the Dutch case of Lucia de Berk and the current UK case of Lucy Letby.
What is greenhouse gasses and how many gasses are there to affect the Earth.moosaasad1975
What are greenhouse gasses how they affect the earth and its environment what is the future of the environment and earth how the weather and the climate effects.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.Sérgio Sacani
The return of a sample of near-surface atmosphere from Mars would facilitate answers to several first-order science questions surrounding the formation and evolution of the planet. One of the important aspects of terrestrial planet formation in general is the role that primary atmospheres played in influencing the chemistry and structure of the planets and their antecedents. Studies of the martian atmosphere can be used to investigate the role of a primary atmosphere in its history. Atmosphere samples would also inform our understanding of the near-surface chemistry of the planet, and ultimately the prospects for life. High-precision isotopic analyses of constituent gases are needed to address these questions, requiring that the analyses are made on returned samples rather than in situ.
Cancer cell metabolism: special Reference to Lactate PathwayAADYARAJPANDEY1
Normal Cell Metabolism:
Cellular respiration describes the series of steps that cells use to break down sugar and other chemicals to get the energy we need to function.
Energy is stored in the bonds of glucose and when glucose is broken down, much of that energy is released.
Cell utilize energy in the form of ATP.
The first step of respiration is called glycolysis. In a series of steps, glycolysis breaks glucose into two smaller molecules - a chemical called pyruvate. A small amount of ATP is formed during this process.
Most healthy cells continue the breakdown in a second process, called the Kreb's cycle. The Kreb's cycle allows cells to “burn” the pyruvates made in glycolysis to get more ATP.
The last step in the breakdown of glucose is called oxidative phosphorylation (Ox-Phos).
It takes place in specialized cell structures called mitochondria. This process produces a large amount of ATP. Importantly, cells need oxygen to complete oxidative phosphorylation.
If a cell completes only glycolysis, only 2 molecules of ATP are made per glucose. However, if the cell completes the entire respiration process (glycolysis - Kreb's - oxidative phosphorylation), about 36 molecules of ATP are created, giving it much more energy to use.
IN CANCER CELL:
Unlike healthy cells that "burn" the entire molecule of sugar to capture a large amount of energy as ATP, cancer cells are wasteful.
Cancer cells only partially break down sugar molecules. They overuse the first step of respiration, glycolysis. They frequently do not complete the second step, oxidative phosphorylation.
This results in only 2 molecules of ATP per each glucose molecule instead of the 36 or so ATPs healthy cells gain. As a result, cancer cells need to use a lot more sugar molecules to get enough energy to survive.
Unlike healthy cells that "burn" the entire molecule of sugar to capture a large amount of energy as ATP, cancer cells are wasteful.
Cancer cells only partially break down sugar molecules. They overuse the first step of respiration, glycolysis. They frequently do not complete the second step, oxidative phosphorylation.
This results in only 2 molecules of ATP per each glucose molecule instead of the 36 or so ATPs healthy cells gain. As a result, cancer cells need to use a lot more sugar molecules to get enough energy to survive.
introduction to WARBERG PHENOMENA:
WARBURG EFFECT Usually, cancer cells are highly glycolytic (glucose addiction) and take up more glucose than do normal cells from outside.
Otto Heinrich Warburg (; 8 October 1883 – 1 August 1970) In 1931 was awarded the Nobel Prize in Physiology for his "discovery of the nature and mode of action of the respiratory enzyme.
WARNBURG EFFECT : cancer cells under aerobic (well-oxygenated) conditions to metabolize glucose to lactate (aerobic glycolysis) is known as the Warburg effect. Warburg made the observation that tumor slices consume glucose and secrete lactate at a higher rate than normal tissues.
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...Sérgio Sacani
Since volcanic activity was first discovered on Io from Voyager images in 1979, changes
on Io’s surface have been monitored from both spacecraft and ground-based telescopes.
Here, we present the highest spatial resolution images of Io ever obtained from a groundbased telescope. These images, acquired by the SHARK-VIS instrument on the Large
Binocular Telescope, show evidence of a major resurfacing event on Io’s trailing hemisphere. When compared to the most recent spacecraft images, the SHARK-VIS images
show that a plume deposit from a powerful eruption at Pillan Patera has covered part
of the long-lived Pele plume deposit. Although this type of resurfacing event may be common on Io, few have been detected due to the rarity of spacecraft visits and the previously low spatial resolution available from Earth-based telescopes. The SHARK-VIS instrument ushers in a new era of high resolution imaging of Io’s surface using adaptive
optics at visible wavelengths.
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
Linked Data and cultural heritage data: an overview of the approaches from Europeana and The European Library
1.
2. Linked Data and cultural heritage
data: an overview of the
approaches from Europeana and
The European Library
Nuno Freire
Chief data officer
The European Library
Pacific Neighbourhood Consortium
2014 Annual Conference
Taipei, October 2014
3. Outline
Introduction and Context
• The European Library
• Europeana
• The data model for metadata exchange in the
Europeana network
Linked Data at The European Library
• Managing and linking person names
• Managing and linking place names
• Managing and linking concepts
5. What is The European
Library?
Project started 1996, full operational service
from 2005
European hub of metadata, collections and
increasing amount of full text
Membership of national and research libraries of
47 Council of Europe states
Non-profit, owned and managed by member
libraries
6.
7. What does The European
Library offer?
Experienced
European
project partner
Large-scale
aggregation
Infrastructure
Data and
digital content
of Europe’s
libraries
Data
distribution
Data
enrichment
Linked open
data
9. EUROPEANA - Europe’s cultural heritage
portal
32.6m records from
2,300 European
galleries, museums,
archives and libraries
Books, newspapers,
journals, letters, diaries,
archival papers
Paintings, maps,
drawings, photographs
Music, spoken word,
radio broadcasts
Film, newsreels,
television
Curated exhibitions
31 languages
10. The European Library as
libraries aggregator to Europeana
Domain Aggregators National initiatives
Audiovisual
collections
National Aggregators
Regional Aggregators
Archives
Thematic collections
Libraries
e.g.
Culture
Grid,
Culture.fr
e.g. Musées
Lausannois
e.g. The
European
Library
e.g. APEX
e.g. EUScreen,
European Film
Gateway
e.g. Judaica Europeana,
Europeana Fashion
11. Metadata in the Europeana Context
Provides a portal for users to access that data
• Metadata, previews and links to source
Makes the metadata freely available for anyone to re-use
• Under Creative Commons Zero (CC0) -public domain dedication
Makes metadata available via an API
Makes metadata available as Linked Open Data
• http://data.europeana.eu/
12. Europeana Data Model: a Collaborative Effort
Cross-community development
Involving library, archive and museum experts
Ca. 60 participants
http://pro.europeana.eu/edm-documentation
13. Europeana Data Model: general principles
• A cross domain approach
• Supporting the common semantics of cultural domains
• Addressing the requirements of the Europeana portal
• Adheres to the modeling principles of the Web of Data
• Available as an OWL ontology and XML schema
• Allows finer-grained models of the different domains to be at least
partly interoperable at the semantic level
• Allows metadata to retain their original expressivity and richness
14. Linked Data at
The European Library
Managing and linking person names
15. Which data from VIAF is used at The
European Library
Name variants
Various forms of the name of the person or organization.
May include the complete name, abbreviated names,
acronyms, etc.
Date of birth/death
The dates of birth and death of the person
Nationalities
The nationalities of a person or organization.
16. How data from VIAF is used
in The European Library
Name variants
• For matching of names across records and data sources
• Improves the identification of all publications of a work, the
identification of publications in books-in-print databases, and the
identification of the contributor in the rights-holders databases.
Date of birth/death
• Used for determining the public domain status.
• Used for matching confirmation and disambiguation of
homonyms across data sources
Nationalities
• Used, in some countries, for determining the public domain
status of the work.
17. The matching process
VIAF data used for matching,
disambiguation, and match probability
18. Matching work contributors with VIAF
Names are matched by similarity
Confirmation of the correctness of a name
match is taken from other matching data
• The dates of birth and death
• The title of the work is compared against the list
of titles available in VIAF
• All the contributors of the work are matched
against the list of known co-authors in VIAF
• The publisher(s) of the work are matched against
the list of known publishers in VIAF
A match is only chosen if enough supporting
evidence is found
19. Contributor names in statements of
responsibility
“French Canadian freely arranged by Katherine K. Davis”.
“ed. by Peter Noever ; with a forew. by Frank O. Gehry; and contrib. by
Coop Himmelblau.”
“W. Lange, A.C. Zeven and N.G. Hogenboom, editors”
“by Pamela and Neal Priestland”
“Vicente Aleixandre ; estudio previo, selección y notas de Leopoldo de
Luis”
20. The approach
To approach the problem as a Named Entity Recognition
task in text that may not be grammatically correct, thus
lacking lexical evidence
Some requirements from the ARROW context
• Easily applicable to several languages
• The outcomes of the recognition task must be explainable
Design decisions
• Exploring the structured data within national bibliographies
• By analysis of the frequency of word occurrences in names of
persons, and in other textual data
• Using word occurrence frequency allows to
• bypass the need for building training sets
• be able to provide simpler explanations of the name recognition
results
21. The process – bibliographic record
processing
The named entity recognition is performed for a
record as follows:
• Statement of responsibility is tokenized
• The person names are recognized by comparing the
tokens with the dictionaries
• The recognized names are compared against the
names of the contributors present in the structured
fields of the record.
• If no similar name exists in the record, the contributor
is added to the record in a structured data field
22. Evaluation data set
(size of bibliographies and evaluation samples)
National Bibliography Total
records
Main
language
Evaluation sample
Statements of
responsibility
Referred
Persons
British Library 13.4 million English 205 328
German National
Library 9.4 million German 200 378
National Library of the
Netherlands 3.2 million Dutch 200 335
National Library of
Greece 0.4 million Greek 297 379
Central Institute for the
Union Catalogue of
12.4 million Italian 224 297
Italian Libraries
Royal Library of
Belgium 1 million French and
Dutch 203 387
Total: 1329 2104
23. Evaluation results
Dataset
Exact match
metric
Partial match
metric
Precision Recall Precision Recall
British Library 0.981 0.979 0.991 0.991
German National Library 0.975 0.934 0.992 0.992
National Library of the
Netherlands 0.973 0.875 0.977 0.979
National Library of
Greece 0.656 0.414 0.758 0.868
Central Institute for the
Union Catalogue of
0.97 0.896 0.971 0.973
Italian Libraries
Royal Library of Belgium 0.981 0.959 0.981 0.982
Overall: 0.948 0.837 0.958 0.963
24. Linked Data at
The European Library
Managing and linking place names
25. The approach for place name
linking
• We process the complete metadata elements
• The alignment is performed with Geonames
• Using the RDF dump of Geonames
• A generic approach not using any language
specific information
• The words themselves are not used as evidence
• We use only characteristics of the words (capitalization, size,
etc)
• Wordnets, part-of-speech analysis, morphological
analysis, etc., are not used.
• … in order to allow the use of this approach in a
language independent manner
26. Resolution of the place names
• This task aims to find a single entity in the geographic
ontology for aligning with the place name
• The first step of this task is to find all possible
candidates for the resolution in the geographic
ontology
• Uses a heuristic based predictive model:
• Assigns a probability for each resolution candidate as match
or non_match
• An alignment is established if a minimum probability
threshold for the class match is achieved.
27. Which information supports the place
name resolution
Feature Description
Number of words The number of words in the place name.
Name match If the recognized place name matched: the main name of the
place, an alternate name, etc.
Exact name
match
If the recognized place name matched exactly the place
name.
Relative
population
Relative population of the candidate in comparison with other
candidates.
Geographic
feature type
The type of geographic feature: continent, country, city, etc.
Related places
found
The number of other place names found in the
administrative hierarchy.
Relative related
places
The relative number of administrative divisions found in the
subject heading
In source country If it is located in one of the source countries of the subject
heading system.
28. Linked Data at
The European Library
Managing and linking concepts
29. Linking Subject Indexing and
Classification Data
The context
• The centralization of bibliographic metadata enables
resource access under a unified knowledge organization
system
The challenges
• Diversity of languages
• Diversity of knowledge organization systems in use across
European libraries
• Heterogeneous levels of details in subject information
Current status at The European Library
• Use of alignments between ontologies:
• Alignments were created manually or semi-automatically
• Alignments in use include: CERIF, MACS (LCSH,
RAMEAU, SWD), UDC and DDC
30. References
Further details may be consulted in the following publications:
•Freire, N, 2014, 'Word Occurrence Based Extraction of Work Contributors from Statements
of Responsibility'. International Journal on Digital Libraries: Volume 14, Issue 3 (2014),
Page 141-148. DOI: 10.1007/s00799-014-0113-3.
•Charles, V., Freire, N, Antoine, I., 2014, 'Links, languages and semantics: linked data
approaches in The European Library and Europeana', in 'Linked Data in Libraries: Let's
make it happen!' IFLA 2014 Satellite Meeting on Linked Data in Libraries.
•Freire, N, Muhr, M, 2013, 'Use of Authorities Open Data in the ARROW Rights
Infrastructure' in proceeding of the DC-2013 Linking to the Future Conference, 2013.
•Freire, N, 2013, 'Visualization and navigation of knowledge in pan-European resources: the
case of The European Library' in proceedings of International UDC Seminar on
Classification & Visualization: interfaces to knowledge.
•N. Freire, et al., "Author Consolidation across European National Bibliographies and
Academic Digital Repositories", 11th International Conference on Current Research
Information Systems, 2012.
•N. Freire, J. Borbinha, P. Calado, "A Language Independent Approach for Aligning Subject
Heading Systems with Geographic Ontologies", International Conference on Dublin Core
and Metadata Applications 2011, 2011.
•N. Freire, J. Borbinha, P. Calado, B. Martins, "A Metadata Geoparsing System for Place
Name Recognition and Resolution in Metadata Records", ACM/IEEE Joint Conference on
Digital Libraries, 2011.
31. Thank you
Nuno Freire
Chief data officer
nuno.freire@theeuropeanlibrary.org
Editor's Notes
The European Library has a long history as an online service, starting as a project in 1996 and launching as a full operational service in 2005. Its membership base comprises all the national libraries of Europe, based on the Council of Europe’s 47 states, and a significant number of research libraries. The service is owned and run by its members, with representatives from the pan-European library organisations, including LIBER, on the board. The European Library is a single online gateway to the resources of Europe’s libraries.
The European Library has a very large dataset of some 200 million bibliographic records, representing Europe’s bibliography; 26 million digitised object;, millions of pages of digitised text. The European Library aggregates all types of data and content provided by libraries, including respository text and data. It has a large-scale aggregation infrastructure, and ingests, indexes, enriches and clusters significant amounts of data and content. Metadata is open and distributed as data dumps and APIs and placed in the workflows and systems used by researchers. The data is also provided as linked open datasets. The European Library is an experienced partner in European projects, including project co-ordination. Let’s see some examples.
At a working level, we operate in a network of aggregators. We can’t work directly with 2,200 organisations, so we rely on aggregators to
collect data, harmonise it, and deliver to Europeana.
Aggregators are important because they share a background with the organisations whose content they bring together, so there is close understanding.The aggregation model enables Europeana to collect huge quantities of data from thousands of providers, through only a handful of channels.