Tabular data represents a large amount of published data on the web. The W3C CSV on the Web working group aims to improve interoperability of CSV and similar tabular formats by developing specifications for metadata, data models, and conversions between CSV, JSON, and RDF. The specifications define methods for parsing CSV into structured models, associating CSV with metadata to provide data typing and relationships, and converting between tabular and other common data formats.
Tutorial on RDFa, to be held at ISWC2010 in Shanghai, China. (I was supposed to hold the tutorial but last minute issues made it impossible for me to travel there...)
Open Standards for the Semantic Web: XML / RDF(S) / OWL / SOAPPieter De Leenheer
This lecture elaborates on RDF, RDFS, and SOAP starting from a short recap of XML, and the history of the W3C and the development of "open standard recommendations". We also compare RDF triples with DOGMA lexons. We finalise by listing shortcomings of RDFS regarding semantics, and give short overview of the history of OWL as one answer to this. A full elaboration on OWL and description logic is for another lecture.
Invited talk at USEWOD2014 (http://people.cs.kuleuven.be/~bettina.berendt/USEWOD2014/)
A tremendous amount of machine-interpretable information is available in the Linked Open Data Cloud. Unfortunately, much of this data remains underused as machine clients struggle to use the Web. I believe this can be solved by giving machines interfaces similar to those we offer humans, instead of separate interfaces such as SPARQL endpoints. In this talk, I'll discuss the Linked Data Fragments vision on machine access to the Web of Data, and indicate how this impacts usage analysis of the LOD Cloud. We all can learn a lot from how humans access the Web, and those strategies can be applied to querying and analysis. In particular, we have to focus first on solving those use cases that humans can do easily, and only then consider tackling others.
Tutorial on RDFa, to be held at ISWC2010 in Shanghai, China. (I was supposed to hold the tutorial but last minute issues made it impossible for me to travel there...)
Open Standards for the Semantic Web: XML / RDF(S) / OWL / SOAPPieter De Leenheer
This lecture elaborates on RDF, RDFS, and SOAP starting from a short recap of XML, and the history of the W3C and the development of "open standard recommendations". We also compare RDF triples with DOGMA lexons. We finalise by listing shortcomings of RDFS regarding semantics, and give short overview of the history of OWL as one answer to this. A full elaboration on OWL and description logic is for another lecture.
Invited talk at USEWOD2014 (http://people.cs.kuleuven.be/~bettina.berendt/USEWOD2014/)
A tremendous amount of machine-interpretable information is available in the Linked Open Data Cloud. Unfortunately, much of this data remains underused as machine clients struggle to use the Web. I believe this can be solved by giving machines interfaces similar to those we offer humans, instead of separate interfaces such as SPARQL endpoints. In this talk, I'll discuss the Linked Data Fragments vision on machine access to the Web of Data, and indicate how this impacts usage analysis of the LOD Cloud. We all can learn a lot from how humans access the Web, and those strategies can be applied to querying and analysis. In particular, we have to focus first on solving those use cases that humans can do easily, and only then consider tackling others.
Performance comparison: Multi-Model vs. MongoDB and Neo4jArangoDB Database
Native multi-model databases combine different data models like documents or graphs in one tool and even allow to mix them in a single query. How can this concept compete with a pure document store like MongoDB or a graph database like Neo4j? I myself and a lot of folks in the community asked that question.
So here are some benchmark results.
Searching Relational Data with Elasticsearchsirensolutions
Second Galway Data Meetup, 29th April 2015
Elasticsearch was originally developed for searching flat documents. However, as real world data is inherently more complex, e.g., nested json data, relational data, interconnected documents and entities, Elasticsearch quickly evolves to support more advanced search scenarios. In this presentation, we will review existing features and plugins to support such scenarios, discuss their advantages and disadvantages, and understand which one is more appropriate for a particular scenario.
Linked Data in Use: Schema.org, JSON-LD and hypermedia APIs - Front in Bahia...Ícaro Medeiros
In my talk I walk throgh Semantic Web initiatives, like RDF and SPARQL, linked data principles, discuss some implementation and adoption issues and talk about semantic annotation in HTML. Semantic annotation using the Schema.org vocabulary is demonstrated using both HTML 5 Microdata or JSON-LD input. There is a strong highlight in benefits seen in Google search results with Rich Snippets, Actions in Email, and Google Now with real examples.
Recently a new breed of "multi-model" databases has emerged. They are a document store, a graph database and a key/value store combined in one program. Therefore they are able to cover a lot of use cases which otherwise would need multiple different database systems.
This approach promises a boost to the idea of "polyglot persistence", which has become very popular in recent years although it creates some friction in the form of data conversion and synchronisation between different systems. This is, because with a multi-model database one can enjoy the benefits of polyglot persistence without the disadvantages.
In this talk I will explain the motivation behind the multi-model approach, discuss its advantages and limitations, and will then risk to make some predictions about the NoSQL database market in five years time, which I shall only reveal during the talk.
OrientDB: Unlock the Value of Document Data RelationshipsFabrizio Fortino
a) A general introduction of graph databases and OrientDB,
b) Why connected data has more value than just data,
c)How to "have fun" with OrientDB combining documents with graphs via SQL,
d) A use case on how OrientDB has helped to raise standards in Irish Public Office.
On OrientDB: NOSQL document databases provide an elegant way to deal with data in different shapes enabling developers to create better and faster products quickly. The main goal of these systems is to find the most efficient solution to manage data itself. With the Big Data Explosion we need to deal with a myriad of highly interconnected information. The challenge now is not only on how to store data but on how to manage, analyse, traverse and use your data within the context of relationships. Graph databases shine at maintaining highly connected data and is the fastest growing category in database management systems: 2014 registered an increase of 250% in terms of adoption and Forrester Research predicts that more than a quarter of enterprises will be using graphs by 2017. OrientDB combines more than one NOSQL model offering the unique flexibility of modelling data in the form of either documents, or graphs, while incorporating object oriented programming as a way of encapsulating relationships.
A set of slides that provides a high-level overview of the W3C Linked Data Platform specification presented at the 4th Linked Data in Architecture and Construction Workshop.
For more detailed and technical version of the presentation, please refer to
http://www.slideshare.net/nandana/learning-w3c-linked-data-platform-with-examples
LDAC 2016 programme
http://smartcity.linkeddata.es/LDAC2016/#programme
Comparison with storing data using NoSQL(CouchDB) and a relational database.eross77
This presentation is intended to show those familiar with relational databases how a NoSQL database can make their jobs easier with loosely structured data.
An introduction to Graph databases and in particular Neo4j, including where Neo4j lives on the CAP Scale in relation to other databases, the Graph data model and a very quick introduction to the Cypher Query Language.
Performance comparison: Multi-Model vs. MongoDB and Neo4jArangoDB Database
Native multi-model databases combine different data models like documents or graphs in one tool and even allow to mix them in a single query. How can this concept compete with a pure document store like MongoDB or a graph database like Neo4j? I myself and a lot of folks in the community asked that question.
So here are some benchmark results.
Searching Relational Data with Elasticsearchsirensolutions
Second Galway Data Meetup, 29th April 2015
Elasticsearch was originally developed for searching flat documents. However, as real world data is inherently more complex, e.g., nested json data, relational data, interconnected documents and entities, Elasticsearch quickly evolves to support more advanced search scenarios. In this presentation, we will review existing features and plugins to support such scenarios, discuss their advantages and disadvantages, and understand which one is more appropriate for a particular scenario.
Linked Data in Use: Schema.org, JSON-LD and hypermedia APIs - Front in Bahia...Ícaro Medeiros
In my talk I walk throgh Semantic Web initiatives, like RDF and SPARQL, linked data principles, discuss some implementation and adoption issues and talk about semantic annotation in HTML. Semantic annotation using the Schema.org vocabulary is demonstrated using both HTML 5 Microdata or JSON-LD input. There is a strong highlight in benefits seen in Google search results with Rich Snippets, Actions in Email, and Google Now with real examples.
Recently a new breed of "multi-model" databases has emerged. They are a document store, a graph database and a key/value store combined in one program. Therefore they are able to cover a lot of use cases which otherwise would need multiple different database systems.
This approach promises a boost to the idea of "polyglot persistence", which has become very popular in recent years although it creates some friction in the form of data conversion and synchronisation between different systems. This is, because with a multi-model database one can enjoy the benefits of polyglot persistence without the disadvantages.
In this talk I will explain the motivation behind the multi-model approach, discuss its advantages and limitations, and will then risk to make some predictions about the NoSQL database market in five years time, which I shall only reveal during the talk.
OrientDB: Unlock the Value of Document Data RelationshipsFabrizio Fortino
a) A general introduction of graph databases and OrientDB,
b) Why connected data has more value than just data,
c)How to "have fun" with OrientDB combining documents with graphs via SQL,
d) A use case on how OrientDB has helped to raise standards in Irish Public Office.
On OrientDB: NOSQL document databases provide an elegant way to deal with data in different shapes enabling developers to create better and faster products quickly. The main goal of these systems is to find the most efficient solution to manage data itself. With the Big Data Explosion we need to deal with a myriad of highly interconnected information. The challenge now is not only on how to store data but on how to manage, analyse, traverse and use your data within the context of relationships. Graph databases shine at maintaining highly connected data and is the fastest growing category in database management systems: 2014 registered an increase of 250% in terms of adoption and Forrester Research predicts that more than a quarter of enterprises will be using graphs by 2017. OrientDB combines more than one NOSQL model offering the unique flexibility of modelling data in the form of either documents, or graphs, while incorporating object oriented programming as a way of encapsulating relationships.
A set of slides that provides a high-level overview of the W3C Linked Data Platform specification presented at the 4th Linked Data in Architecture and Construction Workshop.
For more detailed and technical version of the presentation, please refer to
http://www.slideshare.net/nandana/learning-w3c-linked-data-platform-with-examples
LDAC 2016 programme
http://smartcity.linkeddata.es/LDAC2016/#programme
Comparison with storing data using NoSQL(CouchDB) and a relational database.eross77
This presentation is intended to show those familiar with relational databases how a NoSQL database can make their jobs easier with loosely structured data.
An introduction to Graph databases and in particular Neo4j, including where Neo4j lives on the CAP Scale in relation to other databases, the Graph data model and a very quick introduction to the Cypher Query Language.
CRL: A Rule Language for Table Analysis and InterpretationAlexey Shigarov
Tables presented in spreadsheets can be a source of important information that needs to be loaded into relational databases. However, many of them have complex structures. This does not allow to populate databases with their information directly. The presentation is devoted to the issues of the rule-based information extraction from arbitrary tables presented in spreadsheets and its transformation into structured canonical form that can be loaded into a database by standard ETL tools. We suggest a novel rule language called CRL for table analysis and interpretation. It enables developing a simple program to recover missing relationships describing table semantics. Particular sets of rules can be designed for different types of tables to provide extraction and transformation steps in a process of unstructured tabular data integration.
Aalto University School of Arts, Design and Architecture course Dynamic Visualization Design 1 group work presentation "Visual Impairments" 2012-11-08.
At the Devoxx 2015 conference in Belgium, Guillaume Laforge, Product Ninja & Advocate at Restlet, presented about the never-ending REST API design debate, covering many topics like HTTP status codes, Hypermedia APIs, pagination/searching/filtering, and more.
Building RESTfull Data Services with WebAPIGert Drapers
Data services are a major building block inside a service oriented architecture. Not only do they provide the abstraction and isolation between physical storage systems and the business layer, they can also provide the means for: authentication, authorization, transformation, projection, scale (through for example sharding) and caching. This session will walk you through implementing your RESTfull data service so that you can easily enable and integrate the described capabilities
MuleSoft London Community February 2020 - MuleSoft and ODataPace Integration
Our February Meetup in London took us through MuleSoft and OData. Our guest speaker Martin Gardner (Solution Principal at Slalom), covered how you can use the Mulesoft OData APIKit to wrap a SOAP web service in a Mule app that will present an OData interface for use with the Salesforce connect product. With examples from a recent project, Martin showed us how to avoid the pitfalls he fell into and allow you to be successful.
Flexible metadata schemes for research data repositories - CLARIN Conference'21vty
The development of the Common Framework in Dataverse and the CMDI use case. Building AI/ML based workflow for the prediction and linking concepts from external controlled vocabularies to the CMDI metadata values.
Flexible metadata schemes for research data repositories - Clarin Conference...Vyacheslav Tykhonov
The development of the Common Framework in Dataverse and the CMDI use case. Building AI/ML based workflow for the prediction and linking concepts from external controlled vocabularies to the CMDI metadata values.
The workshop will present how to combine tools to quickly query, transform and model data using command line tools.
The goal is to show that command line tools are efficient at handling reasonable sizes of data and can accelerate the data science
process. We will show that in many instances, command line processing ends up being much faster than ‘big-data’ solutions. The content
of the workshop is derived from the book of the same name (http://datascienceatthecommandline.com/). In addition, we will cover
vowpal-wabbit (https://github.com/JohnLangford/vowpal_wabbit) as a versatile command line tool for modeling large datasets.
The Query Service is the new platform solution for querying a variety of data sources. The goal of Query Service is that administrators can configure a metadata description of the data source that can then be used by end users without detailed knowledge of the underlying data source. This session explains how to configure Query Service data sources and use them with the RESTful API or component collection.
balloon Fusion: SPARQL Rewriting Based on Unified Co-Reference InformationKai Schlegel
Presentation for 5th International Workshop on
Data Engineering meets the Semantic Web (DESWeb)
In conjunction with ICDE 2014, Chicago IL, USA, March 31, 2014 held by Kai Schlegel
Data Con LA 2022 - What's new with MongoDB 6.0 and AtlasData Con LA
Sig Narvaez, Executive Solution Architect at MongoDB
MongoDB is now a Developer Data Platform. Come learn what�s new in the 6.0 release and Atlas following all the recent announcements made at MongoDB World 2022. Topics will include
- Atlas Search which combines 3 systems into one (database, search engine, and sync mechanisms) letting you focus on your product's differentiation.
- Atlas Data Federation to seamlessly query, transform, and aggregate data from one or more MongoDB Atlas databases, Atlas Data Lake and AWS S3 buckets
- Queryable Encryption lets you run expressive queries on fully randomized encrypted data to meet the most stringent security requirements
- Relational Migrator which analyzes your existing relational schemas and helps you design a new MongoDB schema.
- And more!
SAS Online Training Institute in Hyderabad - C-Pointcpointss
C-Point Software Solutions is a Leading Training Institute in Hyderabad. We Provide Training on SAP, SAS, Oracle E Business Suite, Informatica, OBIEE, SQL DBA, Hadoop, Cloud Computing, .Net, Testing Tools, Java, Web Designing, PHP.
Self-Service Data Ingestion Using NiFi, StreamSets & KafkaGuido Schmutz
Many of the Big Data and IoT use cases are based on combining data from multiple data sources and to make them available on a Big Data platform for analysis. The data sources are often very heterogeneous, from simple files, databases to high-volume event streams from sensors (IoT devices). It’s important to retrieve this data in a secure and reliable manner and integrate it with the Big Data platform so that it is available for analysis in real-time (stream processing) as well as in batch (typical big data processing). In past some new tools have emerged, which are especially capable of handling the process of integrating data from outside, often called Data Ingestion. From an outside perspective, they are very similar to a traditional Enterprise Service Bus infrastructures, which in larger organization are often in use to handle message-driven and service-oriented systems. But there are also important differences, they are typically easier to scale in a horizontal fashion, offer a more distributed setup, are capable of handling high-volumes of data/messages, provide a very detailed monitoring on message level and integrate very well with the Hadoop ecosystem. This session will present and compare Apache Flume, Apache NiFi, StreamSets and the Kafka Ecosystem and show how they handle the data ingestion in a Big Data solution architecture.
Multi-cluster Kubernetes Networking- Patterns, Projects and GuidelinesSanjeev Rampal
Talk presented at Kubernetes Community Day, New York, May 2024.
Technical summary of Multi-Cluster Kubernetes Networking architectures with focus on 4 key topics.
1) Key patterns for Multi-cluster architectures
2) Architectural comparison of several OSS/ CNCF projects to address these patterns
3) Evolution trends for the APIs of these projects
4) Some design recommendations & guidelines for adopting/ deploying these solutions.
APNIC Foundation, presented by Ellisha Heppner at the PNG DNS Forum 2024APNIC
Ellisha Heppner, Grant Management Lead, presented an update on APNIC Foundation to the PNG DNS Forum held from 6 to 10 May, 2024 in Port Moresby, Papua New Guinea.
1.Wireless Communication System_Wireless communication is a broad term that i...JeyaPerumal1
Wireless communication involves the transmission of information over a distance without the help of wires, cables or any other forms of electrical conductors.
Wireless communication is a broad term that incorporates all procedures and forms of connecting and communicating between two or more devices using a wireless signal through wireless communication technologies and devices.
Features of Wireless Communication
The evolution of wireless technology has brought many advancements with its effective features.
The transmitted distance can be anywhere between a few meters (for example, a television's remote control) and thousands of kilometers (for example, radio communication).
Wireless communication can be used for cellular telephony, wireless access to the internet, wireless home networking, and so on.
# Internet Security: Safeguarding Your Digital World
In the contemporary digital age, the internet is a cornerstone of our daily lives. It connects us to vast amounts of information, provides platforms for communication, enables commerce, and offers endless entertainment. However, with these conveniences come significant security challenges. Internet security is essential to protect our digital identities, sensitive data, and overall online experience. This comprehensive guide explores the multifaceted world of internet security, providing insights into its importance, common threats, and effective strategies to safeguard your digital world.
## Understanding Internet Security
Internet security encompasses the measures and protocols used to protect information, devices, and networks from unauthorized access, attacks, and damage. It involves a wide range of practices designed to safeguard data confidentiality, integrity, and availability. Effective internet security is crucial for individuals, businesses, and governments alike, as cyber threats continue to evolve in complexity and scale.
### Key Components of Internet Security
1. **Confidentiality**: Ensuring that information is accessible only to those authorized to access it.
2. **Integrity**: Protecting information from being altered or tampered with by unauthorized parties.
3. **Availability**: Ensuring that authorized users have reliable access to information and resources when needed.
## Common Internet Security Threats
Cyber threats are numerous and constantly evolving. Understanding these threats is the first step in protecting against them. Some of the most common internet security threats include:
### Malware
Malware, or malicious software, is designed to harm, exploit, or otherwise compromise a device, network, or service. Common types of malware include:
- **Viruses**: Programs that attach themselves to legitimate software and replicate, spreading to other programs and files.
- **Worms**: Standalone malware that replicates itself to spread to other computers.
- **Trojan Horses**: Malicious software disguised as legitimate software.
- **Ransomware**: Malware that encrypts a user's files and demands a ransom for the decryption key.
- **Spyware**: Software that secretly monitors and collects user information.
### Phishing
Phishing is a social engineering attack that aims to steal sensitive information such as usernames, passwords, and credit card details. Attackers often masquerade as trusted entities in email or other communication channels, tricking victims into providing their information.
### Man-in-the-Middle (MitM) Attacks
MitM attacks occur when an attacker intercepts and potentially alters communication between two parties without their knowledge. This can lead to the unauthorized acquisition of sensitive information.
### Denial-of-Service (DoS) and Distributed Denial-of-Service (DDoS) Attacks
This 7-second Brain Wave Ritual Attracts Money To You.!nirahealhty
Discover the power of a simple 7-second brain wave ritual that can attract wealth and abundance into your life. By tapping into specific brain frequencies, this technique helps you manifest financial success effortlessly. Ready to transform your financial future? Try this powerful ritual and start attracting money today!
Bridging the Digital Gap Brad Spiegel Macon, GA Initiative.pptxBrad Spiegel Macon GA
Brad Spiegel Macon GA’s journey exemplifies the profound impact that one individual can have on their community. Through his unwavering dedication to digital inclusion, he’s not only bridging the gap in Macon but also setting an example for others to follow.
1. Tabular Data on the Web
Intro to W3C CSV on the Web Specifications
Gregg Kellogg
gregg@greggkellogg.net
@gkellogg
1
2. Impact of Tabular Data
• Tabular Data represents a large amount of all
data published on the Web
• According to the Open Data Institute, the vast
majority of published open data is tabular
• “Over 90% of the data on data.gov.uk is
tabular data.”
• data.gov lists 158,631 datasets; largely in CSV
2
3. Sources of Tabular Data
• Easiest way to publish data
• Spreadsheet Dumps
• Database Dumps
• SPARQL results
3
4. CSV data is dumb
• It’s a simple text format, data has no inherent
meaning.
• Cells may be data-typed or have a regular
format: what does “08/09/2015” mean?
• Cells may be related to data in other tables/
columns: Foreign Keys
• Cells may be associated with different entities:
Join results
4
5. Web CSV
• 5-star Linked Data
• CSV URLs
• CSVs link to other CSVs
• CSVs link to other
Resources
• RDF and JSON
conversion
5
6. W3C CSV on the Web
• Working Group chartered to allow applications to provide higher
interoperability with working with CSV, or similar formats.
• Use Cases: http://www.w3.org/TR/csvw-ucr/
• Model for Tabular Data and Metadata on the Web: http://
www.w3.org/TR/tabular-data-model/
• Metadata Vocabulary for Tabular Data: http://www.w3.org/TR/tabular-
metadata/
• Generating JSON from Tabular Data on the Web: http://www.w3.org/
TR/csv2json/
• Generating RDF from Tabular Data on the Web: http://www.w3.org/
TR/csv2rdf/
6
7. Examples
7
countryCode latitude longitude name
AD 42.5 1.6 Andorra
AE 23.4 53.8 United Arab Emirates
AF 33.9 67.7 Afghanistan
countries.csv
countryRef year population
AF 1960 9,616,353
AF 1961 9,799,379
AF 1961 9,989,846
country_slice.csv
8. Model for Tabular Data
id
Table Group
id
Table
notes
transformations
about URL
cells
datatype
default
Column
lang
name
number
ordered
property URL
required
separator
table
text direction
titles
value URL
virtual
cells
number
primary key
titles
Row
referenced rows
source number
table
about URL
column
errors
ordered
Cell
property URL
row
string value
table
text direction
value
value URL
8
notes
foreign keys
other annotations
url
other annotations
tables
columns
rows
table direction
other annotations
rows
table
9. Mapping CSV to Model
• Parse CSV: RFC4180 + dialect metadata.
• delimiter, doubleQuote, headerRowCount,
lineTerminators, quoteChar, …
• Dialect Description comes from Metadata Document.
• Match Headers to Columns.
• Parse Cells using Column metadata/datatype.
• Abstract data model used for viewing, validation, and
conversions.
9
10. Metadata
• Finding Metadata from a CSV
• User-specified, Link Header, well-known
locations
• Matching Metadata to a CSV
• CSV must be compatible with metadata (titles/
names)
• Metadata must reference CSV URL
10
11. foreignKeys
columns
@id
@type
Schema
primaryKey
rowTitles
11
url
targetFormat
scriptFormat
titles
source
@id
@type
Transformation
Definition
name
titles
required
suppressOutput
virtual
@id
@type
Column Description
columnReference
reference
Foreign Key
Definition
resource
schemaReference
columnReference
Foreign Key
Reference
array property
link property
URI template property
column reference property
object property
natural language property
atomic property
Legend:
reference to an array of values of a specific category
reference to a value of a specific category
@language
@base
Top-Level
Properties
tables
transformations
tableDirection
tableSchema
dialect
@context
@id
Table Group
notes
@type
decimalChar
groupChar
pattern
Number Format
url
transformations
tableDirection
tableSchema
dialect
notes
Table
@context
@id
@type
suppressOutput
null
lang
textDirection
separator
ordered
default
datatype
Inherited Properties
aboutUrl
propertyUrl
valueUrl
required
base
format
length
minLength
maxLength
minimum
maximum
Datatype
Description
minInclusive
maxInclusive
minExclusive
maxExclusive
@id
@type
encoding
lineTerminators
quoteChar
doubleQuote
skipRows
commentPrefix
header
Dialect Description
headerRowCount
skipBlankRows
skipInitialSpace
trim
@id
delimiter
skipColumns
13. Embedded Metadata
• Generally Column Titles.
• Formats may define CSV conventions for
embedded metadata.
• Principally used to determine metadata
compatibility.
• Also serves as default metadata if no file
located.
13
15. Other Features
• Split cells into multiple items
• Validate Primary Keys and Foreign Key
references (single and multiple columns)
• Define URL properties for columns
• Multiple subjects per column (may be URLs)
• Values as URLs
15
19. Conversions: RDF (min)
countryCode latitude longitude name
AD 42.5 1.6 Andorra
AE 23.4 53.8 United Arab
Emirates
AF 33.9 67.7 Afghanistan
19
@base <http://example.org/countries.csv> .
_:t1r1
<#countryCode> "AD" ;
<#latitude> "42.5" ;
<#longitude> "1.6" ;
<#name> "Andorra" .
_:t1r2
<#countryCode> "AE" ;
<#latitude> "23.4" ;
<#longitude> "53.8" ;
<#name> "United Arab Emirates" .
_:t1r3
<#countryCode> "AF" ;
<#latitude> "33.9" ;
<#longitude> "67.7" ;
<#name> "Afghanistan" .
countries.csv
countries.json
countries-minimal.ttl
20. Other examples
• Rich Annotations: JSON RDF
• Virtual Columns/Multiple Subjects: JSON RDF
• For more see Specifications and Test Suite
20
21. Tools
• CSVLint
• CKAN – open source data portal platform
• Socrata – cloud-based open data
• Google Fusion Tables – data visualization
• Ruby rdf-tabular – CSVW reference implementation
• RDF Distiller
• Structured Data Linter
21
22. Next Steps
• At-Risk – /.well-known/csvm
• More datatype formats
• Metadata in HTML (embedded JSON-LD)
• Tabular Data in HTML
• More implementations!
• Timeline
• Candidate Recommendation – July 2015
• Proposed Recommendation – Oct 2015
• W3C Recommendation – Dec 2015
22