Author: rvs
Date: Fri Nov 4 00:18:01 2011
New Revision: 1197368
URL: http://svn.apache.org/viewvc?rev=1197368&view=rev
Log:
Making Mahout compile
Modified:
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/do-component-build
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/install_mahout.sh
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/patch
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/deb/mahout/rules
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec
incubator/bigtop/branches/hadoop-0.23/bigtop.mk
Modified:
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/do-component-build
URL:
http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/do-component-build?rev=1197368&r1=1197367&r2=1197368&view=diff
==============================================================================
---
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/do-component-build
(original)
+++
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/do-component-build
Fri Nov 4 00:18:01 2011
@@ -16,5 +16,8 @@
set -ex
-mvn clean install -Dmahout.skip.distribution=false -DskipTests
-Dhadoop.version=0.22.0-SNAPSHOT "$@"
-
+mvn clean install -Dmahout.skip.distribution=false -DskipTests
-Dhadoop.version=0.23.0-SNAPSHOT "$@"
+mkdir build
+for i in distribution/target/mahout*.tar.gz ; do
+ tar -C build --strip-components=1 -xzf $i
+done
Modified:
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/install_mahout.sh
URL:
http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/install_mahout.sh?rev=1197368&r1=1197367&r2=1197368&view=diff
==============================================================================
---
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/install_mahout.sh
(original)
+++
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/install_mahout.sh
Fri Nov 4 00:18:01 2011
@@ -107,7 +107,7 @@ install -d -m 0755 $PREFIX/$DOC_DIR
cp -ra ${BUILD_DIR}/lib/* $PREFIX/${LIB_DIR}/lib/
cp ${BUILD_DIR}/mahout*.jar $PREFIX/$LIB_DIR
-cp ${BUILD_DIR}/mahout*.war $PREFIX/$LIB_DIR
+# cp ${BUILD_DIR}/mahout*.war $PREFIX/$LIB_DIR
cp -a ${BUILD_DIR}/*.txt $PREFIX/$DOC_DIR
cp -a ${BUILD_DIR}/bin/* $PREFIX/${LIB_DIR}/bin
Modified:
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/patch
URL:
http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/patch?rev=1197368&r1=1197367&r2=1197368&view=diff
==============================================================================
---
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/patch
(original)
+++
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/patch
Fri Nov 4 00:18:01 2011
@@ -1,8 +1,357 @@
-Index: apache-mahout-c298f70/core/pom.xml
+Index:
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/Step1MapperTest.java
===================================================================
---- apache-mahout-c298f70.orig/core/pom.xml 2011-09-29 15:22:25.369554983
-0700
-+++ apache-mahout-c298f70/core/pom.xml 2011-09-29 15:22:34.417110581 -0700
-@@ -140,10 +140,6 @@
+---
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/Step1MapperTest.java
(revision 1197192)
++++
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/Step1MapperTest.java
(working copy)
+@@ -22,6 +22,7 @@
+ import org.apache.hadoop.conf.Configuration;
+ import org.apache.hadoop.io.LongWritable;
+ import org.apache.hadoop.io.Text;
++import org.apache.hadoop.mapreduce.Mapper;
+ import org.apache.hadoop.mapreduce.TaskAttemptID;
+ import org.apache.mahout.common.MahoutTestCase;
+ import org.apache.mahout.common.RandomUtils;
+@@ -30,6 +31,7 @@
+ import org.apache.mahout.classifier.df.data.DataLoader;
+ import org.apache.mahout.classifier.df.data.Dataset;
+ import org.apache.mahout.classifier.df.data.Utils;
++import org.apache.mahout.classifier.df.mapreduce.MapredOutput;
+ import org.apache.mahout.classifier.df.node.Leaf;
+ import org.apache.mahout.classifier.df.node.Node;
+ import org.junit.Test;
+@@ -109,8 +111,11 @@
+ // expected number of trees that this mapper will build
+ int mapNbTrees = Step1Mapper.nbTrees(NUM_MAPPERS, NUM_TREES, partition);
+
+- MockContext context = new MockContext(new Step1Mapper(),
+- new Configuration(), new TaskAttemptID(), mapNbTrees);
++ MockMapContextWrapper<LongWritable,Text,TreeID,MapredOutput>
mockMapContextWrapper =
++ new MockMapContextWrapper<LongWritable,Text,TreeID,MapredOutput>();
++ Mapper<LongWritable,Text,TreeID,MapredOutput>.Context context =
++ mockMapContextWrapper.getMockContext(new Configuration(),
++ new TaskAttemptID(), mapNbTrees);
+
+ MockStep1Mapper mapper = new MockStep1Mapper(treeBuilder, dataset, seed,
+ partition, NUM_MAPPERS, NUM_TREES);
+@@ -127,10 +132,10 @@
+ mapper.cleanup(context);
+
+ // make sure the mapper built all its trees
+- assertEquals(mapNbTrees, context.nbOutputs());
++ assertEquals(mapNbTrees, mockMapContextWrapper.nbOutputs());
+
+ // check the returned keys
+- for (TreeID k : context.getKeys()) {
++ for (TreeID k : mockMapContextWrapper.getKeys()) {
+ assertEquals(partition, k.partition());
+ assertEquals(treeIndex, k.treeId());
+
+Index:
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockMapContextWrapper.java
+===================================================================
+---
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockMapContextWrapper.java
(revision 0)
++++
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockMapContextWrapper.java
(revision 0)
+@@ -0,0 +1,94 @@
++/**
++ * Licensed to the Apache Software Foundation (ASF) under one
++ * or more contributor license agreements. See the NOTICE file
++ * distributed with this work for additional information
++ * regarding copyright ownership. The ASF licenses this file
++ * to you under the Apache License, Version 2.0 (the
++ * "License"); you may not use this file except in compliance
++ * with the License. You may obtain a copy of the License at
++ *
++ * http://www.apache.org/licenses/LICENSE-2.0
++ *
++ * Unless required by applicable law or agreed to in writing, software
++ * distributed under the License is distributed on an "AS IS" BASIS,
++ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
++ * See the License for the specific language governing permissions and
++ * limitations under the License.
++ */
++
++package org.apache.mahout.classifier.df.mapreduce.partial;
++
++import java.io.IOException;
++
++import org.apache.hadoop.conf.Configuration;
++import org.apache.hadoop.mapreduce.Mapper;
++import org.apache.hadoop.mapreduce.TaskAttemptID;
++import org.apache.hadoop.mapreduce.lib.map.WrappedMapper;
++import org.apache.hadoop.mapreduce.task.MapContextImpl;
++import org.apache.mahout.classifier.df.mapreduce.MapredOutput;
++
++public class MockMapContextWrapper<K1, V1, K2, V2> extends Mapper<K1, V1, K2,
V2> {
++
++ private MockMapContext mockMapContext;
++
++ public class MockMapContext extends MapContextImpl<K1, V1, K2, V2> {
++
++ private final TreeID[] keys;
++ private final MapredOutput[] values;
++ private int index;
++
++ public MockMapContext(Configuration conf, TaskAttemptID taskAttemptID,
int nbTrees) {
++ super(conf, taskAttemptID, null, null, null, null, null);
++
++ keys = new TreeID[nbTrees];
++ values = new MapredOutput[nbTrees];
++ }
++
++ @Override
++ public void write(Object key, Object value) throws IOException {
++ if (index == keys.length) {
++ throw new IOException("Received more output than expected : " +
index);
++ }
++
++ keys[index] = ((TreeID) key).clone();
++ values[index] = ((MapredOutput) value).clone();
++
++ index++;
++ }
++
++ /**
++ * @return number of outputs collected
++ */
++ public int nbOutputs() {
++ return index;
++ }
++
++ public TreeID[] getKeys() {
++ return keys;
++ }
++
++ public MapredOutput[] getValues() {
++ return values;
++ }
++ }
++
++ public Mapper<K1, V1, K2, V2>.Context getMockContext(Configuration conf,
TaskAttemptID taskAttemptID, int nbTrees) {
++ mockMapContext = new MockMapContext(conf, taskAttemptID, nbTrees);
++ return new WrappedMapper<K1, V1, K2, V2>().getMapContext(mockMapContext);
++ }
++
++ /**
++ * @return number of outputs collected
++ */
++ public int nbOutputs() {
++ return mockMapContext.nbOutputs();
++ }
++
++ public TreeID[] getKeys() {
++ return mockMapContext.getKeys();
++ }
++
++ public MapredOutput[] getValues() {
++ return mockMapContext.getValues();
++ }
++}
+
+Property changes on:
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockMapContextWrapper.java
+___________________________________________________________________
+Added: svn:eol-style
+ + native
+
+Index:
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockContext.java
+===================================================================
+---
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockContext.java
(revision 1197192)
++++
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockContext.java
(working copy)
+@@ -1,70 +0,0 @@
+-/**
+- * Licensed to the Apache Software Foundation (ASF) under one or more
+- * contributor license agreements. See the NOTICE file distributed with
+- * this work for additional information regarding copyright ownership.
+- * The ASF licenses this file to You under the Apache License, Version 2.0
+- * (the "License"); you may not use this file except in compliance with
+- * the License. You may obtain a copy of the License at
+- *
+- * http://www.apache.org/licenses/LICENSE-2.0
+- *
+- * Unless required by applicable law or agreed to in writing, software
+- * distributed under the License is distributed on an "AS IS" BASIS,
+- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+- * See the License for the specific language governing permissions and
+- * limitations under the License.
+- */
+-package org.apache.mahout.classifier.df.mapreduce.partial;
+-
+-import java.io.IOException;
+-
+-import org.apache.hadoop.conf.Configuration;
+-import org.apache.hadoop.mapreduce.Mapper;
+-import org.apache.hadoop.mapreduce.TaskAttemptID;
+-import org.apache.hadoop.mapreduce.Mapper.Context;
+-import org.apache.mahout.classifier.df.mapreduce.MapredOutput;
+-
+-/**
+- * Special implementation that collects the output of the mappers
+- */
+-final class MockContext extends Context {
+-
+- private final TreeID[] keys;
+- private final MapredOutput[] values;
+- private int index;
+-
+- MockContext(Mapper<?,?,?,?> mapper, Configuration conf, TaskAttemptID
taskid, int nbTrees)
+- throws IOException, InterruptedException {
+- mapper.super(conf, taskid, null, null, null, null, null);
+-
+- keys = new TreeID[nbTrees];
+- values = new MapredOutput[nbTrees];
+- }
+-
+- @Override
+- public void write(Object key, Object value) throws IOException {
+- if (index == keys.length) {
+- throw new IOException("Received more output than expected : " + index);
+- }
+-
+- keys[index] = ((TreeID) key).clone();
+- values[index] = ((MapredOutput) value).clone();
+-
+- index++;
+- }
+-
+- /**
+- * @return number of outputs collected
+- */
+- public int nbOutputs() {
+- return index;
+- }
+-
+- public TreeID[] getKeys() {
+- return keys;
+- }
+-
+- public MapredOutput[] getValues() {
+- return values;
+- }
+-}
+Index:
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/PartialSequentialBuilder.java
+===================================================================
+---
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/PartialSequentialBuilder.java
(revision 1197192)
++++
core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/PartialSequentialBuilder.java
(working copy)
+@@ -27,10 +27,12 @@
+ import org.apache.hadoop.io.Text;
+ import org.apache.hadoop.mapreduce.InputSplit;
+ import org.apache.hadoop.mapreduce.Job;
++import org.apache.hadoop.mapreduce.Mapper;
+ import org.apache.hadoop.mapreduce.RecordReader;
+ import org.apache.hadoop.mapreduce.TaskAttemptContext;
+ import org.apache.hadoop.mapreduce.TaskAttemptID;
+ import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
++import org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl;
+ import org.apache.mahout.classifier.df.DFUtils;
+ import org.apache.mahout.classifier.df.DecisionForest;
+ import org.apache.mahout.classifier.df.builder.TreeBuilder;
+@@ -51,7 +53,7 @@
+
+ private static final Logger log =
LoggerFactory.getLogger(PartialSequentialBuilder.class);
+
+- private MockContext firstOutput;
++ private MockMapContextWrapper<LongWritable,Text,TreeID,MapredOutput>
mockMapContextWrapper;
+
+ private final Dataset dataset;
+
+@@ -96,9 +98,11 @@
+
+ int numTrees = Builder.getNbTrees(conf); // total number of trees
+
+- TaskAttemptContext task = new TaskAttemptContext(conf, new
TaskAttemptID());
++ TaskAttemptContext task = new TaskAttemptContextImpl(conf, new
TaskAttemptID());
+
+- firstOutput = new MockContext(new Step1Mapper(), conf,
task.getTaskAttemptID(), numTrees);
++ mockMapContextWrapper = new
MockMapContextWrapper<LongWritable,Text,TreeID,MapredOutput>();
++ Mapper<LongWritable,Text,TreeID,MapredOutput>.Context firstOutput =
++ mockMapContextWrapper.getMockContext(conf, task.getTaskAttemptID(),
numTrees);
+
+ /* first instance id in hadoop's order */
+ int[] firstIds = new int[nbSplits];
+@@ -143,7 +147,7 @@
+
+ @Override
+ protected DecisionForest parseOutput(Job job) throws IOException {
+- return processOutput(firstOutput.getKeys(), firstOutput.getValues());
++ return processOutput(mockMapContextWrapper.getKeys(),
mockMapContextWrapper.getValues());
+ }
+
+ /**
+Index: core/src/test/java/org/apache/mahout/common/DummyStatusReporter.java
+===================================================================
+--- core/src/test/java/org/apache/mahout/common/DummyStatusReporter.java
(revision 1197192)
++++ core/src/test/java/org/apache/mahout/common/DummyStatusReporter.java
(working copy)
+@@ -55,4 +55,9 @@
+ public void setStatus(String status) {
+ }
+
++ @Override
++ public float getProgress() {
++ return 0;
++ }
++
+ }
+Index: core/src/test/java/org/apache/mahout/common/DummyRecordWriter.java
+===================================================================
+--- core/src/test/java/org/apache/mahout/common/DummyRecordWriter.java
(revision 1197192)
++++ core/src/test/java/org/apache/mahout/common/DummyRecordWriter.java
(working copy)
+@@ -30,6 +30,10 @@
+ import org.apache.hadoop.mapreduce.Reducer;
+ import org.apache.hadoop.mapreduce.TaskAttemptContext;
+ import org.apache.hadoop.mapreduce.TaskAttemptID;
++import org.apache.hadoop.mapreduce.lib.map.WrappedMapper;
++import org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer;
++import org.apache.hadoop.mapreduce.task.MapContextImpl;
++import org.apache.hadoop.mapreduce.task.ReduceContextImpl;
+
+ public final class DummyRecordWriter<K, V> extends RecordWriter<K, V> {
+
+@@ -65,7 +69,8 @@
+
Configuration configuration,
+
RecordWriter<K2, V2> output)
+ throws IOException, InterruptedException {
+- return mapper.new Context(configuration, new TaskAttemptID(), null,
output, null, new DummyStatusReporter(), null);
++ return new WrappedMapper<K1, V1, K2, V2>().getMapContext(
++ new MapContextImpl<K1, V1, K2, V2>(configuration, new TaskAttemptID(),
null, output, null, new DummyStatusReporter(), null));
+ }
+
+ public static <K1, V1, K2, V2> Reducer<K1, V1, K2, V2>.Context
build(Reducer<K1, V1, K2, V2> reducer,
+@@ -74,7 +79,8 @@
+
Class<K1> keyClass,
+
Class<V1> valueClass)
+ throws IOException, InterruptedException {
+- return reducer.new Context(configuration,
++ return new WrappedReducer<K1, V1, K2, V2>().getReducerContext(
++ new ReduceContextImpl<K1, V1, K2, V2>(configuration,
+ new TaskAttemptID(),
+ new MockIterator(),
+ null,
+@@ -84,7 +90,7 @@
+ new DummyStatusReporter(),
+ null,
+ keyClass,
+- valueClass);
++ valueClass));
+ }
+
+ }
+Index: core/src/test/java/org/apache/mahout/common/DummyCounter.java
+===================================================================
+--- core/src/test/java/org/apache/mahout/common/DummyCounter.java
(revision 1197192)
++++ core/src/test/java/org/apache/mahout/common/DummyCounter.java
(working copy)
+@@ -19,8 +19,8 @@
+
+ package org.apache.mahout.common;
+
+-import org.apache.hadoop.mapreduce.Counter;
++import org.apache.hadoop.mapreduce.counters.GenericCounter;
+
+-final class DummyCounter extends Counter {
++final class DummyCounter extends GenericCounter {
+
+ }
+Index: core/pom.xml
+===================================================================
+--- core/pom.xml (revision 1197192)
++++ core/pom.xml (working copy)
+@@ -140,19 +140,6 @@
<!-- Third Party -->
<dependency>
@@ -10,10 +359,19 @@ Index: apache-mahout-c298f70/core/pom.xm
- <artifactId>hadoop-core</artifactId>
- </dependency>
- <dependency>
- <groupId>org.codehaus.jackson</groupId>
- <artifactId>jackson-core-asl</artifactId>
+- <groupId>org.codehaus.jackson</groupId>
+- <artifactId>jackson-core-asl</artifactId>
+- </dependency>
+- <dependency>
+- <groupId>org.codehaus.jackson</groupId>
+- <artifactId>jackson-mapper-asl</artifactId>
+- </dependency>
+-
+- <dependency>
+ <groupId>org.slf4j</groupId>
+ <artifactId>slf4j-api</artifactId>
</dependency>
-@@ -211,4 +207,43 @@
+@@ -211,4 +198,51 @@
</dependency>
</dependencies>
@@ -47,25 +405,51 @@ Index: apache-mahout-c298f70/core/pom.xm
+ </dependency>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
-+ <artifactId>hadoop-hdfs</artifactId>
++ <artifactId>hadoop-mapreduce-client-common</artifactId>
+ </dependency>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
-+ <artifactId>hadoop-mapred</artifactId>
++ <artifactId>hadoop-mapreduce-client-core</artifactId>
++ </dependency>
++ <dependency>
++ <groupId>org.codehaus.jackson</groupId>
++ <artifactId>jackson-core-asl</artifactId>
++ </dependency>
++ <dependency>
++ <groupId>org.codehaus.jackson</groupId>
++ <artifactId>jackson-mapper-asl</artifactId>
+ </dependency>
+ </dependencies>
+ </profile>
+ </profiles>
</project>
-Index: apache-mahout-c298f70/pom.xml
+Index: pom.xml
===================================================================
---- apache-mahout-c298f70.orig/pom.xml 2011-09-29 15:22:25.347055295 -0700
-+++ apache-mahout-c298f70/pom.xml 2011-09-29 15:22:29.555805208 -0700
-@@ -189,6 +189,220 @@
+--- pom.xml (revision 1197192)
++++ pom.xml (working copy)
+@@ -103,6 +103,17 @@
+ <url>https://issues.apache.org/jira/browse/MAHOUT</url>
+ </issueManagement>
+
++ <repositories>
++ <repository>
++ <id>apache.snapshots</id>
++ <name>Apache Snapshot Repository</name>
++ <url>http://repository.apache.org/snapshots</url>
++ <releases>
++ <enabled>false</enabled>
++ </releases>
++ </repository>
++ </repositories>
++
+ <dependencyManagement>
+ <dependencies>
+
+@@ -260,6 +271,100 @@
+ </exclusions>
</dependency>
-
<dependency>
-+ <groupId>org.apache.hadoop</groupId>
++ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-common</artifactId>
+ <version>${hadoop.version}</version>
+ <exclusions>
@@ -133,152 +517,32 @@ Index: apache-mahout-c298f70/pom.xml
+ <groupId>org.eclipse.jdt</groupId>
+ <artifactId>core</artifactId>
+ </exclusion>
-+ </exclusions>
-+ </dependency>
-+ <dependency>
-+ <groupId>org.apache.hadoop</groupId>
-+ <artifactId>hadoop-hdfs</artifactId>
-+ <version>${hadoop.version}</version>
-+ <exclusions>
-+ <exclusion>
-+ <groupId>net.sf.kosmosfs</groupId>
-+ <artifactId>kfs</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>org.mortbay.jetty</groupId>
-+ <artifactId>jetty</artifactId>
-+ </exclusion>
+ <exclusion>
-+ <groupId>org.mortbay.jetty</groupId>
-+ <artifactId>jetty-util</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>hsqldb</groupId>
-+ <artifactId>hsqldb</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>commons-el</groupId>
-+ <artifactId>commons-el</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>junit</groupId>
-+ <artifactId>junit</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>oro</groupId>
-+ <artifactId>oro</artifactId>
++ <groupId>org.slf4j</groupId>
++ <artifactId>slf4j-api</artifactId>
+ </exclusion>
+ <exclusion>
-+ <groupId>org.mortbay.jetty</groupId>
-+ <artifactId>jsp-2.1</artifactId>
++ <groupId>org.slf4j</groupId>
++ <artifactId>slf4j-jcl</artifactId>
+ </exclusion>
+ <exclusion>
-+ <groupId>org.mortbay.jetty</groupId>
-+ <artifactId>jsp-api-2.1</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>org.mortbay.jetty</groupId>
-+ <artifactId>servlet-api-2.5</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>commons-net</groupId>
-+ <artifactId>commons-net</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>tomcat</groupId>
-+ <artifactId>jasper-runtime</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>tomcat</groupId>
-+ <artifactId>jasper-compiler</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>xmlenc</groupId>
-+ <artifactId>xmlenc</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>net.java.dev.jets3t</groupId>
-+ <artifactId>jets3t</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>org.eclipse.jdt</groupId>
-+ <artifactId>core</artifactId>
++ <groupId>org.slf4j</groupId>
++ <artifactId>slf4j-log4j12</artifactId>
+ </exclusion>
+ </exclusions>
+ </dependency>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
-+ <artifactId>hadoop-mapred</artifactId>
++ <artifactId>hadoop-mapreduce-client-core</artifactId>
++ <version>${hadoop.version}</version>
++ </dependency>
++ <dependency>
++ <groupId>org.apache.hadoop</groupId>
++ <artifactId>hadoop-mapreduce-client-common</artifactId>
+ <version>${hadoop.version}</version>
-+ <exclusions>
-+ <exclusion>
-+ <groupId>net.sf.kosmosfs</groupId>
-+ <artifactId>kfs</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>org.mortbay.jetty</groupId>
-+ <artifactId>jetty</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>org.mortbay.jetty</groupId>
-+ <artifactId>jetty-util</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>hsqldb</groupId>
-+ <artifactId>hsqldb</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>commons-el</groupId>
-+ <artifactId>commons-el</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>junit</groupId>
-+ <artifactId>junit</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>oro</groupId>
-+ <artifactId>oro</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>org.mortbay.jetty</groupId>
-+ <artifactId>jsp-2.1</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>org.mortbay.jetty</groupId>
-+ <artifactId>jsp-api-2.1</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>org.mortbay.jetty</groupId>
-+ <artifactId>servlet-api-2.5</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>commons-net</groupId>
-+ <artifactId>commons-net</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>tomcat</groupId>
-+ <artifactId>jasper-runtime</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>tomcat</groupId>
-+ <artifactId>jasper-compiler</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>xmlenc</groupId>
-+ <artifactId>xmlenc</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>net.java.dev.jets3t</groupId>
-+ <artifactId>jets3t</artifactId>
-+ </exclusion>
-+ <exclusion>
-+ <groupId>org.eclipse.jdt</groupId>
-+ <artifactId>core</artifactId>
-+ </exclusion>
-+ </exclusions>
+ </dependency>
+
+ <dependency>
- <groupId>org.apache.hadoop</groupId>
- <artifactId>hadoop-core</artifactId>
- <version>0.20.204.0</version>
+ <groupId>org.codehaus.jackson</groupId>
+ <artifactId>jackson-core-asl</artifactId>
+ <version>1.8.2</version>
Modified:
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/deb/mahout/rules
URL:
http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/deb/mahout/rules?rev=1197368&r1=1197367&r2=1197368&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/deb/mahout/rules
(original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/deb/mahout/rules
Fri Nov 4 00:18:01 2011
@@ -25,7 +25,7 @@ export DH_OPTIONS
patch: patch-stamp
patch-stamp:
- patch -p1 < debian/patch
+ patch -p0 < debian/patch
touch $@
clean:
@@ -40,9 +40,6 @@ build-indep: build-indep-stamp
build-indep-stamp: patch-stamp
# we'll just use the build from the tarball.
bash debian/do-component-build -Dmaven.repo.local=${HOME}/.m2/repository
- mkdir -p debian/tmp
- (cd
distribution/target/mahout-distribution-${MAHOUT_BASE_VERSION}/mahout-distribution-${MAHOUT_BASE_VERSION}
&& tar cf - --exclude=debian/\* .) | (cd debian/tmp && tar xf -)
-# (cd debian/tmp && tar xvzf
../../distribution/target/mahout-distribution-${MAHOUT_BASE_VERSION}.tar.gz)
touch $@
install: install-indep
@@ -50,7 +47,7 @@ install-indep:
dh_testdir
dh_testroot
sh -x debian/install_mahout.sh \
- --build-dir=debian/tmp \
+ --build-dir=build \
--doc-dir=/usr/share/doc/mahout \
--prefix=debian/mahout
rm debian/mahout/usr/lib/mahout/lib/hadoop-*.jar
Modified:
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec
URL:
http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec?rev=1197368&r1=1197367&r2=1197368&view=diff
==============================================================================
---
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec
(original)
+++
incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec
Fri Nov 4 00:18:01 2011
@@ -66,8 +66,8 @@ diverse community to facilitate discussi
also on potential use cases. Come to the mailing lists to find out more.
%prep
-%setup -n apache-mahout-c298f70
-%patch0 -p1
+%setup -n apache-mahout-464be41
+%patch0 -p0
%build
bash $RPM_SOURCE_DIR/do-component-build
@@ -75,7 +75,7 @@ bash $RPM_SOURCE_DIR/do-component-build
%install
%__rm -rf $RPM_BUILD_ROOT
sh $RPM_SOURCE_DIR/install_mahout.sh \
-
--build-dir=distribution/target/mahout-distribution-%{mahout_base_version}/mahout-distribution-%{mahout_base_version}
\
+ --build-dir=build \
--prefix=$RPM_BUILD_ROOT \
--doc-dir=%{doc_mahout}
rm -f $RPM_BUILD_ROOT/usr/lib/mahout/lib/hadoop*.jar
Modified: incubator/bigtop/branches/hadoop-0.23/bigtop.mk
URL:
http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop.mk?rev=1197368&r1=1197367&r2=1197368&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop.mk (original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop.mk Fri Nov 4 00:18:01 2011
@@ -128,7 +128,7 @@ MAHOUT_RELEASE_VERSION=1
MAHOUT_TARBALL_DST=mahout-distribution-$(MAHOUT_BASE_VERSION)-src.tar.gz
#MAHOUT_TARBALL_SRC=$(MAHOUT_TARBALL_DST)
#MAHOUT_SITE=$(APACHE_MIRROR)/mahout/0.5/
-MAHOUT_TARBALL_SRC=c298f70
+MAHOUT_TARBALL_SRC=464be41
MAHOUT_SITE=https://github.com/apache/mahout/tarball
$(eval $(call PACKAGE,mahout,MAHOUT))