MAHOUT-1826: Fix wikipedia example URLs, this closes apache/mahout#212
Project: http://git-wip-us.apache.org/repos/asf/mahout/repo Commit: http://git-wip-us.apache.org/repos/asf/mahout/commit/c9c89627 Tree: http://git-wip-us.apache.org/repos/asf/mahout/tree/c9c89627 Diff: http://git-wip-us.apache.org/repos/asf/mahout/diff/c9c89627 Branch: refs/heads/master Commit: c9c89627b29156525d1cae714629c59bd011ba53 Parents: a210111 Author: smarthi <[email protected]> Authored: Fri Apr 8 19:38:51 2016 -0400 Committer: smarthi <[email protected]> Committed: Fri Apr 8 19:38:51 2016 -0400 ---------------------------------------------------------------------- examples/bin/classify-wikipedia.sh | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/mahout/blob/c9c89627/examples/bin/classify-wikipedia.sh ---------------------------------------------------------------------- diff --git a/examples/bin/classify-wikipedia.sh b/examples/bin/classify-wikipedia.sh index 8a7889f..88a56c9 100755 --- a/examples/bin/classify-wikipedia.sh +++ b/examples/bin/classify-wikipedia.sh @@ -74,11 +74,11 @@ if [ "x$alg" != "xclean" ]; then # Datasets: uncomment and run "clean" to change dataset ######################################################## ########## partial small 42.5M zipped - # curl http://dumps.wikimedia.org/enwiki/latest/enwiki-latest-pages-articles1.xml-p000000010p000010000.bz2 -o ${WORK_DIR}/wikixml/enwiki-latest-pages-articles.xml.bz2 + # curl https://dumps.wikimedia.org/enwiki/latest/enwiki-latest-pages-articles1.xml-p000000010p000030302.bz2 -o ${WORK_DIR}/wikixml/enwiki-latest-pages-articles.xml.bz2 ########## partial larger 256M zipped - curl http://dumps.wikimedia.org/enwiki/latest/enwiki-latest-pages-articles10.xml-p000925001p001325000.bz2 -o ${WORK_DIR}/wikixml/enwiki-latest-pages-articles.xml.bz2 + curl https://dumps.wikimedia.org/enwiki/latest/enwiki-latest-pages-articles10.xml-p002336425p003046511.bz2 -o ${WORK_DIR}/wikixml/enwiki-latest-pages-articles.xml.bz2 ######### full wikipedia dump: 10G zipped - #curl http://dumps.wikimedia.org/enwiki/latest/enwiki-latest-pages-articles.xml.bz2 -o ${WORK_DIR}/wikixml/enwiki-latest-pages-articles.xml.bz2 + # curl https://dumps.wikimedia.org/enwiki/latest/enwiki-latest-pages-articles.xml.bz2 -o ${WORK_DIR}/wikixml/enwiki-latest-pages-articles.xml.bz2 ######################################################## fi if [ ! -e ${WORK_DIR}/wikixml/enwiki-latest-pages-articles.xml ]; then
