Free Large datasets to experiment with Hadoop

后端 未结 4 1243
鱼传尺愫
鱼传尺愫 2020-12-22 16:46

Do you know any large dataset to experiment with Hadoop which is free/low cost? Any pointers/links related are appreciated.

Preference:

  • At least one

相关标签:
4条回答
  • 2020-12-22 17:12

    Few points about your question regarding crawling and wikipedia.

    You have linked to the wikipedia data dumps and you can use the Cloud9 project from UMD to work with this data in Hadoop.

    They have a page on this: Working with Wikipedia

    Another datasource to add to the list is:

    • ClueWeb09 - 1 billion webpages collected between Jan and Feb 09. 5TB Compressed.

    Using a crawler to generate data should be posted in a separate question to one about Hadoop/MapReduce I would say.

    0 讨论(0)
  • 2020-12-22 17:15

    It's no log file but maybe you could use the planet file from OpenStreetMap: http://wiki.openstreetmap.org/wiki/Planet.osm

    CC licence, about 160 GB (unpacked)

    There are also smaller files for each continent: http://wiki.openstreetmap.org/wiki/World

    0 讨论(0)
  • 2020-12-22 17:29

    One obvious source: the Stack Overflow trilogy data dumps. These are freely available under the Creative Commons license.

    0 讨论(0)
  • 2020-12-22 17:32

    This is a collection of 189 datasets for machine learning (which is one of the nicest applications for hadoop g): http://archive.ics.uci.edu/ml/datasets.html

    0 讨论(0)
提交回复
热议问题