The LEMO Annotation Framework is a web-based multimedia annotation system that uses semantic web standards like RDF and OWL. It supports annotating images, audio, and video. LEMO aims to provide a uniform annotation model and fragment identification approach across different media types by reusing standards like Annotea, MPEG-21, and exposing annotations as web resources through RESTful interfaces.
Proto-GIS and the Birth of Digital Mappingseltzoid
This document provides an overview of the origins and development of digital mapping, or "proto-GIS", from the 1960s through the present. It discusses key individuals and organizations involved in early digital mapping efforts, including William Bunge's theoretical geography work in the 1960s, the OSS map division's work during WWII collecting maps from the public, and Marie Tharp and Bruce Heezen's mapping of ocean floor topography in the 1950s-60s. It also examines themes of data collection and surveillance anxieties over time, vulnerabilities of algorithms to hacking and spoofing, and the need for critical histories of geographic computation.
The general aim of LinkedCulture is to describe how the information need of the Tussen Kunst & Kitsch (Antiques Roadshow) viewers can be satisfied from both their couch and on-the-go, supporting both passive and more active needs. Linking to external information and content, such as Europeana, museum collections but also auction information has been incorporated in these scenarios
11.5.14 Presentation Slides, “Fedora 4.0 in Action at Penn State and Stanford”DuraSpace
Hot Topics: The DuraSpace Community Webinar Series
Series 9: Early Advantage: Introducing New Fedora 4.0 Repositories
Curated by David Wilcox, Fedora Product Manager, DuraSpace
“Fedora 4.0 in Action at Penn State and Stanford”
Wednesday, November 5, 1:00-2:00pm ET
Presented by:
David Wilcox, Fedora Product Manager, DuraSpace
Adam Wead, Developer, Pennsylvania State University and Tom Cramer, Chief Technology Strategist and Associate Director of Digital Library Systems and Services, Stanford University
The document summarizes the Open Annotation Collaboration (OAC) Model which aims to enable sharing of annotations across platforms and content collections. It presents the OAC goals, the baseline data model where an annotation associates a body resource about a target resource, additional properties that can be included, different annotation types, how bodies can be embedded inline, and how fragments and media fragments can identify portions of resources. It discusses how constrained targets can describe segments that can't use fragment URIs.
Annotating with Annotations - PHPBenelux June/2012Rafael Dohms
What exactly are annotations? How can they change the way you code and make life simpler? Annotations allow you to implement new functionality into code using "notes" this allows you to easily maintain your own architecture but benefit from external tools. Let's look at how annotations are used today in PHP and how we can develop our own solutions based on the existing libraries.
Assemblers translate assembly language into machine code object files. Linkers merge object files and library routines into executable files by resolving references and assigning memory locations. Loaders bring executables into memory and start program execution by initializing registers and jumping to the main routine.
Proto-GIS and the Birth of Digital Mappingseltzoid
This document provides an overview of the origins and development of digital mapping, or "proto-GIS", from the 1960s through the present. It discusses key individuals and organizations involved in early digital mapping efforts, including William Bunge's theoretical geography work in the 1960s, the OSS map division's work during WWII collecting maps from the public, and Marie Tharp and Bruce Heezen's mapping of ocean floor topography in the 1950s-60s. It also examines themes of data collection and surveillance anxieties over time, vulnerabilities of algorithms to hacking and spoofing, and the need for critical histories of geographic computation.
The general aim of LinkedCulture is to describe how the information need of the Tussen Kunst & Kitsch (Antiques Roadshow) viewers can be satisfied from both their couch and on-the-go, supporting both passive and more active needs. Linking to external information and content, such as Europeana, museum collections but also auction information has been incorporated in these scenarios
11.5.14 Presentation Slides, “Fedora 4.0 in Action at Penn State and Stanford”DuraSpace
Hot Topics: The DuraSpace Community Webinar Series
Series 9: Early Advantage: Introducing New Fedora 4.0 Repositories
Curated by David Wilcox, Fedora Product Manager, DuraSpace
“Fedora 4.0 in Action at Penn State and Stanford”
Wednesday, November 5, 1:00-2:00pm ET
Presented by:
David Wilcox, Fedora Product Manager, DuraSpace
Adam Wead, Developer, Pennsylvania State University and Tom Cramer, Chief Technology Strategist and Associate Director of Digital Library Systems and Services, Stanford University
The document summarizes the Open Annotation Collaboration (OAC) Model which aims to enable sharing of annotations across platforms and content collections. It presents the OAC goals, the baseline data model where an annotation associates a body resource about a target resource, additional properties that can be included, different annotation types, how bodies can be embedded inline, and how fragments and media fragments can identify portions of resources. It discusses how constrained targets can describe segments that can't use fragment URIs.
Annotating with Annotations - PHPBenelux June/2012Rafael Dohms
What exactly are annotations? How can they change the way you code and make life simpler? Annotations allow you to implement new functionality into code using "notes" this allows you to easily maintain your own architecture but benefit from external tools. Let's look at how annotations are used today in PHP and how we can develop our own solutions based on the existing libraries.
Assemblers translate assembly language into machine code object files. Linkers merge object files and library routines into executable files by resolving references and assigning memory locations. Loaders bring executables into memory and start program execution by initializing registers and jumping to the main routine.
The document describes the manifest file for a Building Block, which provides directives to the Blackboard Learning System about how to render links for the Building Block. The manifest defines several types of links for basic information, content handling, applications, and modules. It also describes how to localize text in the manifest by including resource bundle files for different languages.
An Architecture for Open Cross-Media Annotation ServicesBeat Signer
Presentation given at WISE 2009, Tenth International Conference on Web Information Systems Engineering, Poznan, Poland, October 2009
ABSTRACT: The emergence of new media technologies in combination with enhanced information sharing functionality oered by the Web provides new possibilities for cross-media annotations. This in turn raises new challenges in terms of how a true integration across dierent types of media can be achieved and how we can develop annotation services that are sufficiently flexible and extensible to cater for new document formats as they emerge. We present a general model for cross-media annotation services and describe how it was used to define an architecture that supports extensibility at the data level as well as within authoring and visualisation tools.
The document discusses key concepts related to memory models in C#, including:
1. The compilation process involves lexical analysis, parsing, semantic analysis, optimization, and code generation.
2. Value types are stored on the stack while reference types are stored on the heap.
3. The garbage collector performs memory management by freeing up unused memory on the heap.
A Controlled Natural Language Interface for Semantic MediaWikiJie Bao
This document proposes using a controlled natural language interface for the semantic wiki Semantic MediaWiki. It aims to improve usability and expressivity. Key points:
- Using a controlled natural language instead of formal logic improves ease of use for non-experts and allows knowledge input without thinking in "subject-property-object" terms.
- An ontology meta-model extends Semantic MediaWiki to support the full range of OWL/RDF constructs like class domains and ranges.
- Forms, templates, and a natural language generation module allow editing knowledge in controlled natural language and translating between the wiki, ontology meta-model, and RDF formats.
- The approach supports multiple controlled natural
TYPO3 is developing a new framework called TYPO3 5.0 that will provide the foundation for the CMS and other applications. The framework uses components, packages, AOP, MVC and other design patterns to improve modularity, reuse and separation of concerns. It also leverages existing open source frameworks where appropriate. The goal is to develop a flexible yet powerful platform that inspires continued collaboration.
SophiaConf2010 Présentation des Retours d'expériences de la Conférence du 08 ...TelecomValley
SophiaConf2010 Présentation des Retours d'expériences de la Conférence du 08 Juillet - HTML 5, une plateforme contemporaine pour le Web : Stefano Crosta, Chief Technical Officer de SLICE FACTORY ; Raphaël Troncy, Maître de Conférences à Eurecom.
Open Data Mashups: linking fragments into mosaicsphduchesne
This document discusses open data mashups and linking data fragments into mosaics. It covers linked data standards and representation formats like RDF and JSON-LD. It also discusses formalizing URI fragments for different media types and dimensions like text, time, spatial coordinates and more. The presentation demonstrates a mosaics model for building documents from heterogeneous data fragments and embedding them while maintaining their original contexts. Potential use cases include disaster management, open science, fact checking and data journalism.
The document discusses Roslyn, a .NET compiler platform that provides open-source C# and Visual Basic compilers with rich code analysis APIs. It outlines the compiler pipeline including parsing, binding, and IL emission, as well as APIs for syntax trees, symbols, and emitting. The document also describes how Roslyn enables language services like auto-completion, navigation, and refactoring tools.
This document provides a Bash style guide and coding standard for writing Bash scripts. It outlines best practices for script structure, formatting, commenting, and testing. Key recommendations include limiting line length to 88 characters, using indentation to indicate code blocks, including introductory and section comments, and thoroughly testing scripts for syntax, functionality, and edge cases. The goal is to create scripts that are easy to understand, maintain, and modify.
Doug Cutting on the State of the Hadoop EcosystemCloudera, Inc.
Doug Cutting, Apache Hadoop Co-founder, explains how the growth of the Hadoop ecosystem has made Hadoop a much more powerful machine, and how the continued expansion will lead to great things.
This document outlines a technique for executing arbitrary binaries in memory on a Mac OS X machine by defeating address space layout randomization (ASLR). It discusses:
1. The structure of Mach-O files and how binaries are executed by the kernel and dynamic linker in Mac OS X.
2. A proposed attack that embeds a shellcode and crafted stack in a binary to impersonate the kernel and load a new binary into memory.
3. How ASLR works in Mac OS X Leopard and a method to retrieve the base address of the libSystem library to adjust symbol addresses and defeat ASLR, enhancing the attack.
4. A demonstration of the technique executing Nmap and Safari
Typography + Text Editing in the Text layout FrameworkHuyen Tue Dao
The document provides an overview of the Text Layout Framework (TLF) and how it can be used for advanced text formatting and layout. TLF delivers more typographic control, multi-lingual support, and text selection/editing capabilities compared to regular HTML text. It has a model-view-controller structure built on the Flash Text Engine, representing text as hierarchical FlowElements that can be formatted and rendered via containers on stage. Formatting is applied via style properties on FlowElements or by importing/exporting markup. While Flex components simplify usage, developers can also directly assemble text flows, containers, and interaction managers.
- Semantic publishing enhances the meaning and automated discovery of published articles by enriching them with metadata. This allows articles to be more easily linked, integrated, and analyzed.
- SPAR and EARMARK are ontologies that can be used to semantically annotate TEI documents. SPAR defines terms for scholarly objects, and EARMARK defines how to annotate text ranges without modifying the source document.
- An example is provided of how SPAR and EARMARK could be used to annotate a TEI version of Shakespeare's The Tempest, linking text excerpts to definitions in DBPedia and recording different annotators' interpretations.
The document discusses key concepts in C# programming including literals, variables, data types, and constants. It defines literals as value constants assigned to variables, variables as named locations to store data that can change during program execution, and data types as determining the type of data a variable can hold and its properties. The document also covers common C# keywords, rules for naming variables, value and reference types, default variable values, and the use of constants.
Wiki systems have developed over the past years as lightweight, community-editable, web-based hypertext systems. With the emergence of semantic wikis such as Semantic MediaWiki, these collections of interlinked documents have also gained a dual role as ad-hoc RDF graphs. However, their roots lie in the limited hypertext capabilities of the World Wide Web: embedded links, without support for features like composite objects or transclusion. Collaborative editing on wikis has been hampered by redundancy; much of the effort spent on Wikipedia is used keeping content synchronised and organised. We have developed a model for a system, which we have prototyped and are evaluating, which reintroduces ideas from the field of hypertext to help alleviate this burden.
Presented at BlogTalk 2009.
This document provides an introduction to Cassandra, a distributed database management system. It begins with an overview of Cassandra and how it compares to traditional databases. Key aspects discussed include that Cassandra uses a simple query language, scales out through clustering rather than up on larger servers, does not require a fixed database schema, and is eventually consistent. The document then covers Cassandra's data model, architecture, configuration, usage, performance considerations and tuning. Real-world experiences with Cassandra in a Twitter analytics application are also shared.
Decentralized Finance (DeFi) - Understanding Risks in an Emerging Financial P...Bernhard Haslhofer
Decentralized finance (DeFi) protocols allow for permissionless and transparent financial services on distributed ledgers using smart contracts. However, DeFi comes with various technical, economic, and systemic risks. At the technical level, smart contract vulnerabilities and network congestion could be exploited. Behavior like front-running on public transaction data allows some agents to profit. Systemic risks also exist if protocols become interconnected in complex ways. Understanding both individual protocols and their compositions is important to help address these risks.
Token Systems, Payment Channels, and Corporate CurrenciesBernhard Haslhofer
This document provides a summary of token systems, payment channels, and corporate currencies. It begins with defining key concepts like tokens, token systems, and cryptocurrencies. It then discusses Ethereum token standards like ERC-20 and ERC-721 and provides examples. Payment channels are introduced as a solution for blockchain scalability, and implementations like the Lightning Network and Raiden Network are summarized. Finally, it discusses corporate currencies like Libra and compares them to cryptocurrencies. The document aims to explain and demystify recent developments in the crypto space.
The document describes the manifest file for a Building Block, which provides directives to the Blackboard Learning System about how to render links for the Building Block. The manifest defines several types of links for basic information, content handling, applications, and modules. It also describes how to localize text in the manifest by including resource bundle files for different languages.
An Architecture for Open Cross-Media Annotation ServicesBeat Signer
Presentation given at WISE 2009, Tenth International Conference on Web Information Systems Engineering, Poznan, Poland, October 2009
ABSTRACT: The emergence of new media technologies in combination with enhanced information sharing functionality oered by the Web provides new possibilities for cross-media annotations. This in turn raises new challenges in terms of how a true integration across dierent types of media can be achieved and how we can develop annotation services that are sufficiently flexible and extensible to cater for new document formats as they emerge. We present a general model for cross-media annotation services and describe how it was used to define an architecture that supports extensibility at the data level as well as within authoring and visualisation tools.
The document discusses key concepts related to memory models in C#, including:
1. The compilation process involves lexical analysis, parsing, semantic analysis, optimization, and code generation.
2. Value types are stored on the stack while reference types are stored on the heap.
3. The garbage collector performs memory management by freeing up unused memory on the heap.
A Controlled Natural Language Interface for Semantic MediaWikiJie Bao
This document proposes using a controlled natural language interface for the semantic wiki Semantic MediaWiki. It aims to improve usability and expressivity. Key points:
- Using a controlled natural language instead of formal logic improves ease of use for non-experts and allows knowledge input without thinking in "subject-property-object" terms.
- An ontology meta-model extends Semantic MediaWiki to support the full range of OWL/RDF constructs like class domains and ranges.
- Forms, templates, and a natural language generation module allow editing knowledge in controlled natural language and translating between the wiki, ontology meta-model, and RDF formats.
- The approach supports multiple controlled natural
TYPO3 is developing a new framework called TYPO3 5.0 that will provide the foundation for the CMS and other applications. The framework uses components, packages, AOP, MVC and other design patterns to improve modularity, reuse and separation of concerns. It also leverages existing open source frameworks where appropriate. The goal is to develop a flexible yet powerful platform that inspires continued collaboration.
SophiaConf2010 Présentation des Retours d'expériences de la Conférence du 08 ...TelecomValley
SophiaConf2010 Présentation des Retours d'expériences de la Conférence du 08 Juillet - HTML 5, une plateforme contemporaine pour le Web : Stefano Crosta, Chief Technical Officer de SLICE FACTORY ; Raphaël Troncy, Maître de Conférences à Eurecom.
Open Data Mashups: linking fragments into mosaicsphduchesne
This document discusses open data mashups and linking data fragments into mosaics. It covers linked data standards and representation formats like RDF and JSON-LD. It also discusses formalizing URI fragments for different media types and dimensions like text, time, spatial coordinates and more. The presentation demonstrates a mosaics model for building documents from heterogeneous data fragments and embedding them while maintaining their original contexts. Potential use cases include disaster management, open science, fact checking and data journalism.
The document discusses Roslyn, a .NET compiler platform that provides open-source C# and Visual Basic compilers with rich code analysis APIs. It outlines the compiler pipeline including parsing, binding, and IL emission, as well as APIs for syntax trees, symbols, and emitting. The document also describes how Roslyn enables language services like auto-completion, navigation, and refactoring tools.
This document provides a Bash style guide and coding standard for writing Bash scripts. It outlines best practices for script structure, formatting, commenting, and testing. Key recommendations include limiting line length to 88 characters, using indentation to indicate code blocks, including introductory and section comments, and thoroughly testing scripts for syntax, functionality, and edge cases. The goal is to create scripts that are easy to understand, maintain, and modify.
Doug Cutting on the State of the Hadoop EcosystemCloudera, Inc.
Doug Cutting, Apache Hadoop Co-founder, explains how the growth of the Hadoop ecosystem has made Hadoop a much more powerful machine, and how the continued expansion will lead to great things.
This document outlines a technique for executing arbitrary binaries in memory on a Mac OS X machine by defeating address space layout randomization (ASLR). It discusses:
1. The structure of Mach-O files and how binaries are executed by the kernel and dynamic linker in Mac OS X.
2. A proposed attack that embeds a shellcode and crafted stack in a binary to impersonate the kernel and load a new binary into memory.
3. How ASLR works in Mac OS X Leopard and a method to retrieve the base address of the libSystem library to adjust symbol addresses and defeat ASLR, enhancing the attack.
4. A demonstration of the technique executing Nmap and Safari
Typography + Text Editing in the Text layout FrameworkHuyen Tue Dao
The document provides an overview of the Text Layout Framework (TLF) and how it can be used for advanced text formatting and layout. TLF delivers more typographic control, multi-lingual support, and text selection/editing capabilities compared to regular HTML text. It has a model-view-controller structure built on the Flash Text Engine, representing text as hierarchical FlowElements that can be formatted and rendered via containers on stage. Formatting is applied via style properties on FlowElements or by importing/exporting markup. While Flex components simplify usage, developers can also directly assemble text flows, containers, and interaction managers.
- Semantic publishing enhances the meaning and automated discovery of published articles by enriching them with metadata. This allows articles to be more easily linked, integrated, and analyzed.
- SPAR and EARMARK are ontologies that can be used to semantically annotate TEI documents. SPAR defines terms for scholarly objects, and EARMARK defines how to annotate text ranges without modifying the source document.
- An example is provided of how SPAR and EARMARK could be used to annotate a TEI version of Shakespeare's The Tempest, linking text excerpts to definitions in DBPedia and recording different annotators' interpretations.
The document discusses key concepts in C# programming including literals, variables, data types, and constants. It defines literals as value constants assigned to variables, variables as named locations to store data that can change during program execution, and data types as determining the type of data a variable can hold and its properties. The document also covers common C# keywords, rules for naming variables, value and reference types, default variable values, and the use of constants.
Wiki systems have developed over the past years as lightweight, community-editable, web-based hypertext systems. With the emergence of semantic wikis such as Semantic MediaWiki, these collections of interlinked documents have also gained a dual role as ad-hoc RDF graphs. However, their roots lie in the limited hypertext capabilities of the World Wide Web: embedded links, without support for features like composite objects or transclusion. Collaborative editing on wikis has been hampered by redundancy; much of the effort spent on Wikipedia is used keeping content synchronised and organised. We have developed a model for a system, which we have prototyped and are evaluating, which reintroduces ideas from the field of hypertext to help alleviate this burden.
Presented at BlogTalk 2009.
This document provides an introduction to Cassandra, a distributed database management system. It begins with an overview of Cassandra and how it compares to traditional databases. Key aspects discussed include that Cassandra uses a simple query language, scales out through clustering rather than up on larger servers, does not require a fixed database schema, and is eventually consistent. The document then covers Cassandra's data model, architecture, configuration, usage, performance considerations and tuning. Real-world experiences with Cassandra in a Twitter analytics application are also shared.
Decentralized Finance (DeFi) - Understanding Risks in an Emerging Financial P...Bernhard Haslhofer
Decentralized finance (DeFi) protocols allow for permissionless and transparent financial services on distributed ledgers using smart contracts. However, DeFi comes with various technical, economic, and systemic risks. At the technical level, smart contract vulnerabilities and network congestion could be exploited. Behavior like front-running on public transaction data allows some agents to profit. Systemic risks also exist if protocols become interconnected in complex ways. Understanding both individual protocols and their compositions is important to help address these risks.
Token Systems, Payment Channels, and Corporate CurrenciesBernhard Haslhofer
This document provides a summary of token systems, payment channels, and corporate currencies. It begins with defining key concepts like tokens, token systems, and cryptocurrencies. It then discusses Ethereum token standards like ERC-20 and ERC-721 and provides examples. Payment channels are introduced as a solution for blockchain scalability, and implementations like the Lightning Network and Raiden Network are summarized. Finally, it discusses corporate currencies like Libra and compares them to cryptocurrencies. The document aims to explain and demystify recent developments in the crypto space.
Dr. Bernhard Haslhofer presented on measurements and analytics techniques for cryptocurrency networks. He discussed network abstractions like transaction networks and address networks that can be used to cluster addresses and analyze cryptocurrency flows. As an example application, he described a ransomware study using the GraphSense analytics platform that identified the Locky ransomware family as generating the highest revenues, estimated at over $7 million USD, by tracing cryptocurrency transactions from seed ransomware addresses.
Post-Bitcoin Cryptocurrencies, Off-Chain Transaction Channels, and Cryptocur...Bernhard Haslhofer
The document provides an overview of post-Bitcoin cryptocurrencies, off-chain transaction channels, and cryptocurrency analytics techniques. It discusses the realities of Bitcoin compared to initial promises and expectations, such as de-facto centralization and lack of anonymity. Privacy-enhancing cryptocurrencies like Monero and Zcash are introduced that aim to provide untraceable and unlinkable transactions. Off-chain payment channels like the Lightning Network are described as a solution to Bitcoin's scalability problems by moving transactions off-chain. The document concludes by outlining goals and approaches for cryptocurrency analytics to better understand the structure and dynamics of cryptocurrency ecosystems.
Insight Into Cryptocurrencies - Methods and Tools for Analyzing Blockchain-ba...Bernhard Haslhofer
This document discusses cryptocurrency analytics and summarizes a presentation on analyzing blockchain-based ecosystems. The presentation covers goals of cryptocurrency analytics like macroscopic and microscopic analysis of ecosystems. It describes approaches like using blockchain address graphs and clustering addresses. Statistics on the Bitcoin blockchain are provided. Implementation details are discussed along with a cryptocurrency analytics tool. Stakeholders like science, public authorities, and fintech are examined along with use cases for ransomware studies, law enforcement, anti-money laundering compliance, and evaluating blockchain technology. Future research directions include cybercrime, financial crime forensics, and off-chain transaction channels.
O Bitcoin Where Art Thou? An Introduction to Cryptocurrency AnalyticsBernhard Haslhofer
This document discusses cryptocurrency analytics and summarizes a presentation on the topic. It introduces some of the most frequent questions received about cryptocurrency and outlines the goals of developing scalable quantitative methods and tools to better understand cryptocurrency ecosystems through micro-level analysis of individual transactions and macro-level analysis of actors and their relationships. Example analytics methods are described, including blockchain analytics techniques like address clustering and transaction network analysis.
Mind the Gap - Data Science Meets Software EngineeringBernhard Haslhofer
This document summarizes a talk on combining data science and software engineering approaches. It discusses how the two fields approach problems differently, with software engineering focusing on implementing features and ensuring quality through testing, while data science focuses on evaluating models and metrics. The document proposes a solution of defining goals, collecting ground truth data, implementing models and functions, testing and evaluating them, analyzing errors, and deploying services based on metrics. This "metrics driven software engineering" approach aims to bridge the gaps between the two fields.
GraphSense - Real-time Insight into Virtual Currency EcosystemsBernhard Haslhofer
GraphSense is a project to develop algorithms and tools for gaining insight into evolving virtual currency transaction graphs. The project will construct graphs from open and private data sources on bitcoin transactions, apply scalable anomaly detection and machine learning algorithms, and provide a dashboard for analytics. The 24-month project is led by AIT - Austrian Institute of Technology and funded by FFG - IKT der Zukunft.
BITCOIN - De-anonymization and Money Laundering Detection StrategiesBernhard Haslhofer
This document discusses strategies for de-anonymizing bitcoin transactions and detecting money laundering on the bitcoin blockchain. It describes cross-referencing the blockchain with external data like forum posts to link user information to transactions. It also discusses learning the peer-to-peer network topology by observing which servers forward transaction information. Methods for scanning the blockchain for known money laundering patterns and reverse engineering mixing services are presented. Finally, transaction blacklisting is proposed as a prevention strategy.
Bitcoin - Introduction, Technical Aspects and Ongoing DevelopmentsBernhard Haslhofer
This document provides an introduction to Bitcoin, covering its technical aspects and ongoing developments. It begins with introductions from two presenters, Bernhard Haslhofer and Aljosha Judmayer. The agenda then outlines an overview of Bitcoin, including its introduction, technical aspects such as how the blockchain works, and ongoing developments like alternative applications and recent initiatives from MIT. It also briefly discusses Bitcoin's history with crime and potential applications beyond digital currency.
Maphub und Pelagios: Anwendung von Linked Data in den Digitalen Geisteswissen...Bernhard Haslhofer
In recent years, scientists at the Austrian Institute of Technology have been involved in numerous projects in the digital humanities area. In this talk, Dr. Bernhard Haslhofer will present two of them, both having a strong focus on applying the Linked Open Data method on datasets produced throughout the project. The first is Maphub (http://maphub.github.io/), an open source Web application which allows users to create annotations on historical maps, link these annotations with other Web sources (e.g., Wikipedia), and share annotations as Linked Open Data following the Open Annotation model. The second is Pelagios (http://pelagios-project.blogspot.co.at/), a community initiative that aims to facilitate better linking between online resources documenting the past, based on the places they refer to. To date, Pelagios interconnects 900.000+ heterogeneous digital objects - literature, archaeology, epigraphy, cartography - from 40+ international partners. The current focus of the project is to annotate Early Geospatial Documents - documents that use written or visual representation to describe geographic space prior to the European discovery of the Americas in 1492, and make the annotations available as (Linked) Open Data.
This document discusses open data and the OpenGLAM network. It provides an overview of OpenGLAM, which aims to make cultural heritage data openly available and reusable without restrictions. The document outlines the benefits of open data, such as transparency and engagement. It also describes how OpenGLAM can help cultural institutions make their resources discoverable by providing advice, examples of open collections, and connecting local open data groups.
The document provides an overview of open data principles and techniques for publishing and linking structured data on the web. It discusses key open data principles like availability and universal participation. It then covers two main techniques: linked (open) data, which uses semantic web technologies like RDF and URIs to interlink data; and microdata, which embeds structured data in HTML using tags. The presenter provides examples of open government data and linked open datasets. The goal is to create a single globally connected data space on the web through open sharing of structured data using these techniques.
This document summarizes a study on semantic tagging of historical maps. Researchers conducted an experiment comparing four tagging conditions: label-based tagging, suggestive tagging, semantic tagging, and semantic tagging with context display. They found that semantic tagging did not affect the number or types of tags created by users or their workload, as measured by a task load index, compared to other tagging techniques. Semantic tagging provided unambiguous semantic relationships to defined web resources without negative impacts on the tagging process.
Bernhard Haslhofer presented on Maphub, an open platform for mapping and annotating historical maps. Maphub allows users to participate by annotating maps with semantic tags linked to concepts in Wikipedia. These annotations add interpretive context to maps and help connect related concepts. The annotations are stored as structured data following the Open Annotation data model. By enabling collaborative annotation, Maphub aims to unlock greater engagement with and understanding of cultural heritage maps.
The document discusses the OpenGLAM initiative and OpenGLAM-AT working group in Austria. It provides background on OpenGLAM's goal of promoting open access to cultural heritage data held by galleries, libraries, archives, and museums. It outlines recent developments that expanded open access to cultural works, including Europeana releasing CC0-licensed metadata and the Getty Trust launching an open content program. The document proposes next steps for OpenGLAM-AT, such as a kickoff event, participating in discussions around open data guidelines, and creating awareness through blogging and meetups.
Semantic Tagging for old maps...and other things on the WebBernhard Haslhofer
This document discusses semantic tagging of historical maps and other digital objects on the web. It introduces Maphub, a platform for annotating and semantically tagging digitized historical maps. Maphub allows users to add comments, semantic tags linked to controlled vocabularies, and geo-references to maps. The presentation describes Maphub's features, how semantic tagging works in the system by linking visual tags to unique URIs, and plans to implement semantic tagging plugins for other annotation tools and support additional knowledge organization systems. The benefits of semantic tagging are collecting structured metadata and connecting maps and annotations to other web resources through semantic links.
Semantic Tagging for old maps...and other things on the Web
The LEMO Annotation Framework
1. The LEMO Annotation Framework
Open Annotations Meeting
Berkeley, 10/12/2009
Bernhard Haslhofer
Bernhard Haslhofer
2. What is LEMO?
• A Web-based, multimedia annotation framework,
consisting of
• a back-end storage
• front-ends for image, audio, and video annotation
• Demo at: http://dme.arcs.ac.at/annotation/econnect-
annotation-showcase.html
Bernhard Haslhofer 2
3.
4. Driving Requirements...
1. objects to be annotated are multimedia
‣ need a uniform model; no content-type specific solution
2. annotations address specific content parts (fragments)
‣ need a uniform fragment identification approach
3. shift towards the Web
‣ need to expose annotation data on the Web (Linked Data)
Bernhard Haslhofer 4
5. Reuse of existing standards
• Semantic Web standards (RDF/S, OWL, SPARQL)
• RESTful Web Service Design
• Annotea data model & protocol
• XPointer, MPEG-21, SVG, etc. for fragment
identification
• ....
Bernhard Haslhofer 5
6. Why not use Annotea as it is?
• Annotea is for the annotation of Web pages; LEMO
should also support annotation of embedded content
items
• Annotea doesn’t assign datatypes to its properties;
LEMO uses OWL’s (data)typing features
• Annotea doesn’t support multimedia fragment
identification; LEMO uses MPEG-21 (multimedia)
fragment identification
Bernhard Haslhofer 6
7. A Uniform Annotation Model
• annotation profiles; cf. application profiles for metadata
interoperability
• an Annotea-based core model
• with content- and annotation-type specific model
extensions
• implemented as light-weight software components
(add-ons)
Bernhard Haslhofer 7
8. A Uniform Annotation Model
Annotea Annotation Schema: http://www.w3.org/2000/10/annotation-ns#
Annotation
annotates author body created modified context related
subClassOf subPropertyOf subPropertyOf subPropertyOf subPropertyOf subPropertyOf subPropertyOf
annotates author label created modified fragment
date date any
string string
Time Time URI
Annotation
Annotation Core Schema: http://lemo.mminf.univie.ac.at/annotation-core#
Key Symbols:
RDFS/OWL RDFS OWL OWL XML Schema
domain range
Class Property ObjectProperty DatatypeProperty Datatype
Bernhard Haslhofer 8
9. A Uniform Annotation Model
Annotation Relationship Schema: Text Annotation Schema:
http://lemo.mminf.univie.ac.at/ann-relationship# http://lemo.mminf.univie.ac.at/ann-text#
Relation- Text
ship Annotation
isLinkedTo title description
subClassOf subCla
ssOf
subPropertyOf subPropertyOf
annotates author label created modified fragment
date date any
string string
Time Time URI
Annotation
Annotation Core Schema: http://lemo.mminf.univie.ac.at/annotation-core#
Bernhard Haslhofer 9
10. Uniform Fragment Identification
• introduce a fragment element with URI range
• follow MPEG-21 media pointer scheme
• an annotation can define multiple fragments
• URI-encoding of fragments
• minimal consensus
• content-type specific extensions
Bernhard Haslhofer 10
11. Uniform Fragment Identification
12
< rdf : RDF Fragment
xmlns : a =" http :// lemo . mminf . univie . ac . at / cation can int
annotation - core #"
..." > for all approa
ized and wide
...
useful to reta
<a : annotates > http :// www . univie . ac . at / test . mpg identifier can
</ a : annotates > fragment part
<a : fragment >
http :// www . univie . ac . at / test . mpg # mp (~ time requested res
( ’ npt ’ , ’30 ’ , ’40 ’) ) losing the exa
</ a : fragment >
relationship t
... By using
</ rdf : RDF >
improving the
representation
Listing 1 Time fragment of a video expressed according to lieve, as Geurt
the MPEG-21 fragment identification specification tous use of UR
interoperable,
Bernhard Haslhofer 11 annotations.
12. <a : annotates > http :// www . univie . ac . at / test . mpg identifier can not be processed by a
</ a : annotates > fragment part of the respective UR
<a : fragment >
http :// www . univie . ac . at / test . mpg # mp (~ time requested resource is returned. Wi
( ’ npt ’ , ’30 ’ , ’40 ’) ) losing the exact fragment, this beh
</ a : fragment >
relationship to the resource as a wh
... By using this simple method in
improving the interoperability of fra
Uniform Fragment Identification
</ rdf : RDF >
representations in diverse annotatio
Listing 1 Time fragment of a video expressed according to lieve, as Geurts et al [20] have conclu
the MPEG-21 fragment identification specification tous use of URIs will help to solve the
interoperable, explicit links between
annotations.
< rdf : RDF
xmlns : a =" http :// lemo . mminf . univie . ac . at /
annotation - core #"
xmlns : x =" http :// lemo . mminf . univie . ac . at / 4.4 Exposing Annotations as Web R
annotation - video #"
..." >
Since we follow a REST-based appro
... URIs (e.g., http://www.example.o
are in fact dereferencable URIs, whi
<a : annotates > http :// www . univie . ac . at / test . mpg
</ a : annotates > by humans and machines. Therefor
<a : fragment > tation framework fulfills the first and
http :// www . univie . ac . at / test . mpg # mp (~ time
( ’ npt ’ , ’30 ’ , ’40 ’) ) principles, as described in Section 3
</ a : fragment > To fulfill the third principle, LEM
<x : time_fragment xmlns : mpeg21 =" http :// lemo .
mminf . univie . ac . at / an not ati on_m peg 21 #" >
expose annotation data in different
< mpeg21 : uri_fid > http :// www . univie . ac . at / Humans typically access Web resour
test . mpg # mp (~ time ( ’ npt ’ , ’30 ’ , ’40 ’) ) which in turn requires an (X)HTM
</ mpeg21 : uri_fid >
< mpeg21 : time_scheme > npt </ mpeg21 : order to display the returned inform
time_scheme > requirement by relying on content n
< mpeg21 : start_time >30 </ mpeg21 : start_time >
< mpeg21 : end_time >40 </ mpeg21 : end_time > a built-in HTTP feature. Figure 5 il
</ x : time_fragment > tations can be retrieved in various fo
...
the appropriate mime-type in the H
header field. LEMO forwards client r
</ rdf : RDF > annotation (e.g., http://example.c
Listing 2 Alternative representation within the add-on to the appropriate physical represen
model example.com/annotations/html/1
and http://example.com/annotati
Bernhard Haslhofer 12 ing an HTTP 303 See Other respon
13. Annotations as Web Resources
Resource
1
+ URI
annotates
0..* Annotation
Bernhard Haslhofer 13
14. Annotations as Web Resources
Dereference an annotation URI, requesting HTML content
GET http://example.com/annotations/1
Accept: text/html
303 See Other
Location http://example.com/annotations/html/1
GET http://example.com/annotations/html/1
Client LEMO
Accept: text/html
200 OK
<HTML>
...
</HTML>
Dereference an annotation URI, requesting RDF content
GET http://example.com/annotations/1
Accept: application/rdf+xml
303 See Other
Location http://example.com/annotations/rdf/1
GET http://example.com/annotations/rdf/1
Client LEMO
Accept: application/rdf+xml
200 OK
<RDF>
...
</RDF>
Bernhard Haslhofer 14
15. Current LEMO developments
• integration into Europeana portal
• enhancing multimedia support on user-interface
• integration with semi-automatic (multimedia)
metadata extraction techniques
• investigating adoption of fragment identification
working groups recommendations
Bernhard Haslhofer 15