Downloads 2016-04

This pages provides all DBpedia datasets as links to files in bzip2 1 compression. The DBpedia datasets are licensed under the terms of the Creative Commons Attribution-ShareAlike License and the GNU Free Documentation License
In addition to the RDF version of the data, we also provide a tabular version of some of the core DBpedia data sets as CSV and JSON files. See DBpediaAsTables.

See also the change log for recent changes and developments.

Contents

 

1. Wikipedia Input Files

 

All xml soure files of Wikipedia are now hosted alongside the extracted data and can be found in the dataset table as 'pages-articles' dataset.

All datasets were extracted from these Wikipedia dump files, generated in March/April 2016. Please refer to to DataIDs of this release for more information about the source datasets.

 

 

2. Ontology

 

The ontology version used while extracting all datasets can be downloaded here:

 

 

3. Datasets

 

The following table provides all datasets extracted by the extraction framework for every wikipedia language with more than 10.000 articles.
Select the languages you are interested in on the top of the table, filter the list of datasets with the search function.
Click on the dataset names to obtain additional information. Click on the question mark next to a download link to preview file contents.

Starting with this release we provide all datasets in two serializations:

  • turtle (ttl): provides data in n-triple format (<subject>  <predicate>  <object> .) as a subset of turtle serialization
  • quad-turtle (tql): the quad turtle serialization (<subject>  <predicate>  <object> <graph/context>.) adds context information to every triple, containing the graph name and provenance information

Some datasets are available in two versions:

  • localized: These datasets contain triples extracted from the respective Wikipedia, including the ones whose URIs do not have an equivalent English article.
  • cannonicalized: These datasets contain triples extracted from the respective Wikipedia whose subject and object resource have an equivalent English article (marked with an *).
 
 

 

4. Links to other datasets

 

A table with all links will appear here in a short while. In the meantime please refer directly to the download directory to get all link-sets of this release:

Links

 

5. NLP Datasets

 

DBpedia also includes a number of NLP Datasets - datasets specifically targeted at supporting Computational Linguistics and Natural Language Processing (NLP) tasks. Among those, we highlight the Lexicalization Dataset, Topic Signatures, Thematic Concepts and Grammatical Genders.

1 Most files were packed with pbzip2, which generates concatenated streams. Some older bzip2 decompressors, for example Apache Commons Compress before version 1.4, cannot handle this format. Please make sure that you use the latest version. Let us know if you experience any problems.


6. Dataset Metadata as DataIDs

 

Starting with the release 2016-04 we provide extensive dataset metadata by adding DataIDs for all extracted languages to the respective language directories. Use these files to gather additional information about the Datasets and the files which represent them. 
A dcat:Catalog file (ttl, json-ld) pointing to all DataIDs (via dcat:record) can be found in the root folder of this release.


7. Previous versions of DBpedia