This is an informal overview of Linked Data and the usage made of it for the project http://res.space (presented on August 11th 2016 during a team meeting)
9. Why Linked Data ?
Slides derived from http://www.slideshare.net/cgueret/linking-knowledge-spaces
10. Dealing with documents until 1989
1. Find a source for the document
2. Find a way to parse the document
3. Create links and index the content
4. Repeat on each update
11. Then came the Web...
Standardized, connected, decentralised, easy
Document
Document
Document
One server Another server
12. Dealing with data until, well, now
1. Find a source for the data
2. Find a way to parse the data
3. Create links and index the data
4. Repeat on each update
13. We deal with data
the way we dealt
with documents
20 years ago
20. Concretely
“Lille is in France and called Rijsel in Dutch”
http://dbpedia.org/resource/Lille
http://dbpedia.org/resource/France“Rijsel”@nl
http://dbpedia.org/ontology/country
http://www.w3.org/2000/01/rdf-schema#label
21. Not rocket science:
● Use IRIs for identifiers
● Bind identifiers to data about them
● Bind identifiers to other identifiers
● Use IRIs for typing the links
22. Vocabularies
● QB : statistics
● PROV-O : provenance
● SIOC : social media
● Schema.org : search engine results
● And many more ...
23. Every data set is a graph part of a
bigger graph
Only need to know the vocabulary
used to meaningfully consume it
24. Vocabularies are part of the graph too!
http://dbpedia.org/ontology/country
http://www.w3.org/2000/01/rdf-schema#comment
“The country where the thing is located.”@en
28. Two major traps to avoid
● Using the same IRI for a thing and
the document describing it
● Applying a license to a thing
instead of applying it to a document
30. LOD + Semantics = Semantic Web
● Document vocabularies with logics => New
data gets derived
● “Lille is in France” + “All cities in France
are in Europe” => “Lille is in Europe”
35. In practice we
● Help GLAMs to publish LOD
● Crawl and index that LOD
● Provide a search interface over the
data crawled
36. Our crawler follows links across data
publishers to hunt for (properly
licensed) LOD
There is a set of rules used to
interpret the data in a specific way
37.
38. All of that is open source ! Both what we use and what we code :-)