[
https://issues.apache.org/jira/browse/MAHOUT-1451?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Gaurav Misra updated MAHOUT-1451:
---------------------------------
Description:
Cleaning up the following clustering examples:
=====================================
https://mahout.apache.org/users/clustering/clustering-of-synthetic-control-data.html
Introduction
This example will demonstrate clustering of time series data, specifically
control charts. [Control charts : http://en.wikipedia.org/wiki/Control_chart]
are tools used to determine whether a manufacturing or business process is in a
state of statistical control. Such control charts are generated / simulated
repeatedly at equal time intervals. A simulated dataset is available for use in
UCI machine learning repository. The data is described [here :
http://archive.ics.uci.edu/ml/databases/synthetic_control/synthetic_control.data.html].
Problem Description
A time series of control charts needs to be clustered into their close knit
groups. The data set we use is synthetic and is meant to resemble real world
information in an anonymized format. It contains six different classes: Normal,
Cyclic, Increasing trend, Decreasing trend, Upward shift, Downward shift. In
this example we will use Mahout to cluster the data into corresponding class
buckets.
At the end of this example
* You will have clustered data using mahout.
* You will see how to analyse the clusters produced by mahout.
* You will have a starting point for incorporating clustering into your own
software.
Setup
We need to do some initial setup before we are able to run the example.
1. Start out by downloading the input dataset (to be clustered) from the UCI
Machine Learning Repository:
http://archive.ics.uci.edu/ml/databases/synthetic_control/synthetic_control.data
2. Make sure the data consists of 600 rows and 60 columns. The first 100 rows
contains Normal data followed by 100 rows of Cyclic data and so on with a total
of 6 classes.
3. This example assumes that you have already set up Mahout/Hadoop. If you
have not done so yet:
4.
* Hadoop: Follow the instructions on
http://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-common/SingleNodeSetup.html
to set up Hadoop.
* Mahout: Follow the instructions on the [Quickstart:
https://mahout.apache.org/users/basics/quickstart.html] page.
5. Make sure the Hadoop daemons are running if you are running Hadoop in
distributed mode.
6. Create a directory on your local machine called « testdata » and place the
input dataset in this directory.
7. Run the following command to copy the input data into HDFS:
* Create a directory called « testdata » on HDFS:
$HADOOP_HOME/bin/hadoop fs -mkdir testdata
* Copy the directory named « testdata » from your local filesystem to
HDFS:
$HADOOP_HOME/bin/hadoop fs -put testdata
8. The final setup step is to build Mahout by going to the $MAHOUT_HOME
directory and running one of the following commands:
9.
* For a full build: mvn clean install
* For a build without unit tests: mvn -DskipTests clean install
10. You should see a build successful message once the build script has
completed.
11. Finally make sure that the examples have compiled successfully. You
should find the compiled jar in the /examples/target directory under the name
mahout-examples-{version}.job.jar
12. This concludes all the setup required to run the examples.
Clustering Examples
There are examples available for three clustering algorithms:
* Canopy Clustering:
https://mahout.apache.org/users/clustering/canopy-clustering.html
* k-Means Clustering:
https://mahout.apache.org/users/clustering/k-means-clustering.html
* Fuzzy k-Means Clustering:
https://mahout.apache.org/users/clustering/fuzzy-k-means.html
Depending on the example you want to run the following command can be used:
* Canopy Clustering: $MAHOUT_HOME/bin/mahout
org.apache.mahout.clustering.syntheticcontrol.canopy.Job
* k-Means Clustering: $MAHOUT_HOME/bin/mahout
org.apache.mahout.clustering.syntheticcontrol.kmeans.Job
* Fuzzy k-Means Clustering: $MAHOUT_HOME/bin/mahout
org.apache.mahout.clustering.syntheticcontrol.fuzzykmeans.Job
The clustering output will be produced in the « output » directory on HDFS. The
output should be copied to your local filesystem since it is overwritten on
each run.
Use the following command to copy out the data to your local filesystem:
$HADOOP_HOME/bin/hadoop fs -get output $MAHOUT_HOME/examples
This creates an output folder inside examples directory. The output data points
are in vector format. In order to read/analyze the output, you can use
[clusterdump: https://mahout.apache.org/users/clustering/cluster-dumper.html]
utility provided by Mahout.
The source code for these examples is located under the examples project.
=====================================
https://mahout.apache.org/users/clustering/clustering-seinfeld-episodes.html
was:
Cleaning up the following clustering examples:
https://mahout.apache.org/users/clustering/clustering-of-synthetic-control-data.html
https://mahout.apache.org/users/clustering/clustering-seinfeld-episodes.html
> Cleaning up the examples for clustering on the website
> ------------------------------------------------------
>
> Key: MAHOUT-1451
> URL: https://issues.apache.org/jira/browse/MAHOUT-1451
> Project: Mahout
> Issue Type: Improvement
> Reporter: Gaurav Misra
> Original Estimate: 48h
> Remaining Estimate: 48h
>
> Cleaning up the following clustering examples:
> =====================================
> https://mahout.apache.org/users/clustering/clustering-of-synthetic-control-data.html
> Introduction
> This example will demonstrate clustering of time series data, specifically
> control charts. [Control charts : http://en.wikipedia.org/wiki/Control_chart]
> are tools used to determine whether a manufacturing or business process is in
> a state of statistical control. Such control charts are generated / simulated
> repeatedly at equal time intervals. A simulated dataset is available for use
> in UCI machine learning repository. The data is described [here :
> http://archive.ics.uci.edu/ml/databases/synthetic_control/synthetic_control.data.html].
> Problem Description
> A time series of control charts needs to be clustered into their close knit
> groups. The data set we use is synthetic and is meant to resemble real world
> information in an anonymized format. It contains six different classes:
> Normal, Cyclic, Increasing trend, Decreasing trend, Upward shift, Downward
> shift. In this example we will use Mahout to cluster the data into
> corresponding class buckets.
> At the end of this example
> * You will have clustered data using mahout.
> * You will see how to analyse the clusters produced by mahout.
> * You will have a starting point for incorporating clustering into your
> own software.
> Setup
> We need to do some initial setup before we are able to run the example.
> 1. Start out by downloading the input dataset (to be clustered) from the
> UCI Machine Learning Repository:
> http://archive.ics.uci.edu/ml/databases/synthetic_control/synthetic_control.data
> 2. Make sure the data consists of 600 rows and 60 columns. The first 100
> rows contains Normal data followed by 100 rows of Cyclic data and so on with
> a total of 6 classes.
> 3. This example assumes that you have already set up Mahout/Hadoop. If you
> have not done so yet:
> 4.
> * Hadoop: Follow the instructions on
> http://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-common/SingleNodeSetup.html
> to set up Hadoop.
> * Mahout: Follow the instructions on the [Quickstart:
> https://mahout.apache.org/users/basics/quickstart.html] page.
> 5. Make sure the Hadoop daemons are running if you are running Hadoop in
> distributed mode.
> 6. Create a directory on your local machine called « testdata » and place
> the input dataset in this directory.
> 7. Run the following command to copy the input data into HDFS:
> * Create a directory called « testdata » on HDFS:
> $HADOOP_HOME/bin/hadoop fs -mkdir testdata
> * Copy the directory named « testdata » from your local filesystem to
> HDFS:
> $HADOOP_HOME/bin/hadoop fs -put testdata
> 8. The final setup step is to build Mahout by going to the $MAHOUT_HOME
> directory and running one of the following commands:
> 9.
> * For a full build: mvn clean install
> * For a build without unit tests: mvn -DskipTests clean install
> 10. You should see a build successful message once the build script has
> completed.
> 11. Finally make sure that the examples have compiled successfully. You
> should find the compiled jar in the /examples/target directory under the name
> mahout-examples-{version}.job.jar
> 12. This concludes all the setup required to run the examples.
> Clustering Examples
> There are examples available for three clustering algorithms:
> * Canopy Clustering:
> https://mahout.apache.org/users/clustering/canopy-clustering.html
> * k-Means Clustering:
> https://mahout.apache.org/users/clustering/k-means-clustering.html
> * Fuzzy k-Means Clustering:
> https://mahout.apache.org/users/clustering/fuzzy-k-means.html
> Depending on the example you want to run the following command can be used:
> * Canopy Clustering: $MAHOUT_HOME/bin/mahout
> org.apache.mahout.clustering.syntheticcontrol.canopy.Job
> * k-Means Clustering: $MAHOUT_HOME/bin/mahout
> org.apache.mahout.clustering.syntheticcontrol.kmeans.Job
> * Fuzzy k-Means Clustering: $MAHOUT_HOME/bin/mahout
> org.apache.mahout.clustering.syntheticcontrol.fuzzykmeans.Job
> The clustering output will be produced in the « output » directory on HDFS.
> The output should be copied to your local filesystem since it is overwritten
> on each run.
> Use the following command to copy out the data to your local filesystem:
> $HADOOP_HOME/bin/hadoop fs -get output $MAHOUT_HOME/examples
> This creates an output folder inside examples directory. The output data
> points are in vector format. In order to read/analyze the output, you can use
> [clusterdump: https://mahout.apache.org/users/clustering/cluster-dumper.html]
> utility provided by Mahout.
> The source code for these examples is located under the examples project.
> =====================================
> https://mahout.apache.org/users/clustering/clustering-seinfeld-episodes.html
--
This message was sent by Atlassian JIRA
(v6.2#6252)