Tag

encyclopedia

13 datasets
  • Freebase Wikipedia Extraction (WEX)

    Offsite — The Freebase Wikipedia Extraction (WEX) is a processed dump of the English language Wikipedia. The wiki markup for each article is transformed into machine-readable XML, and common relational features such as templates, infoboxes, categories, article sections, and redirects are extracted intabular form. Freebase WEX is provided as a set of database tables in TSV format ...
  • MusicBrainz

    Offsite — MusicBrainz is a user-maintained open music community that collects, and makes available to the public, music metadata, including information about artists, release groups, releases, tracks, labels and the many relationships between them. The database also contains a full history of all the changes that the MusicBrainz community has made to the music metadata. The music ...
  • DBPedia Main

    Offsite — DBpedia is a community effort to extract structured information from Wikipedia and to make this information available on the Web. The DBpedia knowledge base currently describes more than 2.6 million things, including at least 213,000 persons, 328,000 places, 57,000 music albums, 36,000 films, 20,000 companies. The knowledge base consists of 274 million pieces of ...
  • Citizendium

    Offsite — Contributions to Citizendium from Wikipedia are licenced under the GFDL.
  • Open History

    Offsite — Collection of articles – mostly about Japanese history. Started in 2001 and last updated in 2006-09-18.
  • Wikipedia 3

    Offsite — “Wikipedia³ is a conversion of the English Wikipedia into RDF. It’s a monthly updated dataset containing around 47 million triples.” “The Wikipedia³ datasets are of course licensed under the GFDL. Enjoy!”
  • Medpedia

    Offsite — About > The Medpedia Project is a long-term, worldwide project to evolve a new model for sharing and advancing knowledge about health, medicine and the body among medical professionals and the general public. This model is founded on providing a free online technology platform that is collaborative, interdisciplinary and transparent. Read more about the model. > Users ...
  • Freebase Data Dump

    Offsite — Freebase data dumps provide all of the current facts and assertions within the Freebase system. The data dumps are complete, general-purpose extracts of the Freebase data in a variety of formats. Freebase releases a fresh data dump every three months. Freebase is an open database of the world’s information, covering millions of topics across hundreds of categories. ...
  • Jewish Encyclopedia

    Free Download — This dataset consists of a collection of Infoboxes from Wikipedia on the topic of Jewish Encyclopedia.
  • Wikipedia XML Data

    Offsite — This data set contains a complete copy of all Wikimedia wikis, in the form of wikitext source and metadata embedded in XML as provided by the Wikimedia Foundation. The data set will be updated every month and the 3 previous months will always be available for use. We will list previous snapshots in the text of this description.
  • DBPedia

    Offsite — ,DBpedia is a community effort to extract structured information from Wikipedia and to make this information available on the Web. The DBpedia knowledge base currently describes more than 2.6 million things, including at least 213,000 persons, 328,000 places, 57,000 music albums, 36,000 films, 20,000 companies. The knowledge base consists of 274 million pieces of ...
  • Wikipedia Extraction (WEX)

    Offsite — The Freebase Wikipedia Extraction (WEX) is a processed dump of the English language Wikipedia. The wiki markup for each article is transformed into machine-readable XML, and common relational features such as templates, infoboxes, categories, article sections, and redirects are extracted intabular form. Freebase WEX is provided as a set of database tables in TSV format ...
  • Complete and Latest English Wikipedia raw dump with edit history

    Offsite — This is a direct link to the raw wikipedia data dump, roughly 7TB uncompressed. The data is bz2, gz, and 7z compressed and in .xml format. A higher level view of the data is available at this link: http://dumps.wikimedia.org/ As explained on this page: http://en.wikipedia.org/wiki/Wikipedia:Database_download, downloading data of this size uses a lot of bandwidth, which ...