Author: szetszwo Date: Thu Apr 5 20:16:15 2012 New Revision: 1310048 URL: http://svn.apache.org/viewvc?rev=1310048&view=rev Log: Merge r1308633 through r1310044 from trunk.
Modified: hadoop/common/branches/HDFS-3092/ (props changed) hadoop/common/branches/HDFS-3092/dev-support/test-patch.sh hadoop/common/branches/HDFS-3092/hadoop-assemblies/src/main/resources/assemblies/hadoop-mapreduce-dist.xml hadoop/common/branches/HDFS-3092/hadoop-project/pom.xml hadoop/common/branches/HDFS-3092/hadoop-tools/hadoop-archives/src/main/java/org/apache/hadoop/tools/HadoopArchives.java hadoop/common/branches/HDFS-3092/hadoop-tools/hadoop-archives/src/test/java/org/apache/hadoop/tools/TestHadoopArchives.java Propchange: hadoop/common/branches/HDFS-3092/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk:r1308633-1310044 Modified: hadoop/common/branches/HDFS-3092/dev-support/test-patch.sh URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-3092/dev-support/test-patch.sh?rev=1310048&r1=1310047&r2=1310048&view=diff ============================================================================== --- hadoop/common/branches/HDFS-3092/dev-support/test-patch.sh (original) +++ hadoop/common/branches/HDFS-3092/dev-support/test-patch.sh Thu Apr 5 20:16:15 2012 @@ -311,7 +311,7 @@ checkTests () { echo "======================================================================" echo "" echo "" - testReferences=`$GREP -c -i '/test' $PATCH_DIR/patch` + testReferences=`$GREP -c -i -e '^+++.*/test' $PATCH_DIR/patch` echo "There appear to be $testReferences test files referenced in the patch." if [[ $testReferences == 0 ]] ; then if [[ $JENKINS == "true" ]] ; then @@ -333,7 +333,7 @@ checkTests () { fi JIRA_COMMENT="$JIRA_COMMENT - +1 tests included. The patch appears to include $testReferences new or modified tests." + +1 tests included. The patch appears to include $testReferences new or modified test files." return 0 } Modified: hadoop/common/branches/HDFS-3092/hadoop-assemblies/src/main/resources/assemblies/hadoop-mapreduce-dist.xml URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-3092/hadoop-assemblies/src/main/resources/assemblies/hadoop-mapreduce-dist.xml?rev=1310048&r1=1310047&r2=1310048&view=diff ============================================================================== --- hadoop/common/branches/HDFS-3092/hadoop-assemblies/src/main/resources/assemblies/hadoop-mapreduce-dist.xml (original) +++ hadoop/common/branches/HDFS-3092/hadoop-assemblies/src/main/resources/assemblies/hadoop-mapreduce-dist.xml Thu Apr 5 20:16:15 2012 @@ -155,6 +155,9 @@ <excludes> <exclude>org.apache.hadoop:hadoop-common</exclude> <exclude>org.apache.hadoop:hadoop-hdfs</exclude> + <!-- use slf4j from common to avoid multiple binding warnings --> + <exclude>org.slf4j:slf4j-api</exclude> + <exclude>org.slf4j:slf4j-log4j12</exclude> </excludes> </dependencySet> </dependencySets> Modified: hadoop/common/branches/HDFS-3092/hadoop-project/pom.xml URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-3092/hadoop-project/pom.xml?rev=1310048&r1=1310047&r2=1310048&view=diff ============================================================================== --- hadoop/common/branches/HDFS-3092/hadoop-project/pom.xml (original) +++ hadoop/common/branches/HDFS-3092/hadoop-project/pom.xml Thu Apr 5 20:16:15 2012 @@ -253,7 +253,7 @@ <dependency> <groupId>com.google.guava</groupId> <artifactId>guava</artifactId> - <version>r09</version> + <version>11.0.2</version> </dependency> <dependency> <groupId>commons-cli</groupId> @@ -508,11 +508,6 @@ <version>1.6</version> </dependency> <dependency> - <groupId>hsqldb</groupId> - <artifactId>hsqldb</artifactId> - <version>1.8.0.7</version> - </dependency> - <dependency> <groupId>org.slf4j</groupId> <artifactId>slf4j-api</artifactId> <version>1.6.1</version> @@ -684,6 +679,11 @@ <groupId>org.apache.maven.plugins</groupId> <artifactId>maven-jar-plugin</artifactId> <version>2.3.1</version> + <configuration> + <excludes> + <exclude>mrapp-generated-classpath</exclude> + </excludes> + </configuration> </plugin> <plugin> <groupId>org.apache.maven.plugins</groupId> @@ -789,6 +789,21 @@ </executions> </plugin> <plugin> + <artifactId>maven-dependency-plugin</artifactId> + <executions> + <execution> + <id>build-classpath</id> + <phase>generate-sources</phase> + <goals> + <goal>build-classpath</goal> + </goals> + <configuration> + <outputFile>target/classes/mrapp-generated-classpath</outputFile> + </configuration> + </execution> + </executions> + </plugin> + <plugin> <groupId>org.apache.maven.plugins</groupId> <artifactId>maven-surefire-plugin</artifactId> <configuration> Modified: hadoop/common/branches/HDFS-3092/hadoop-tools/hadoop-archives/src/main/java/org/apache/hadoop/tools/HadoopArchives.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-3092/hadoop-tools/hadoop-archives/src/main/java/org/apache/hadoop/tools/HadoopArchives.java?rev=1310048&r1=1310047&r2=1310048&view=diff ============================================================================== --- hadoop/common/branches/HDFS-3092/hadoop-tools/hadoop-archives/src/main/java/org/apache/hadoop/tools/HadoopArchives.java (original) +++ hadoop/common/branches/HDFS-3092/hadoop-tools/hadoop-archives/src/main/java/org/apache/hadoop/tools/HadoopArchives.java Thu Apr 5 20:16:15 2012 @@ -830,11 +830,18 @@ public class HadoopArchives implements T throw new IOException("Parent path not specified."); } parentPath = new Path(args[i+1]); + if (!parentPath.isAbsolute()) { + parentPath= parentPath.getFileSystem(getConf()).makeQualified(parentPath); + } + i+=2; //read the rest of the paths for (; i < args.length; i++) { if (i == (args.length - 1)) { destPath = new Path(args[i]); + if (!destPath.isAbsolute()) { + destPath = destPath.getFileSystem(getConf()).makeQualified(destPath); + } } else { Path argPath = new Path(args[i]); Modified: hadoop/common/branches/HDFS-3092/hadoop-tools/hadoop-archives/src/test/java/org/apache/hadoop/tools/TestHadoopArchives.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-3092/hadoop-tools/hadoop-archives/src/test/java/org/apache/hadoop/tools/TestHadoopArchives.java?rev=1310048&r1=1310047&r2=1310048&view=diff ============================================================================== --- hadoop/common/branches/HDFS-3092/hadoop-tools/hadoop-archives/src/test/java/org/apache/hadoop/tools/TestHadoopArchives.java (original) +++ hadoop/common/branches/HDFS-3092/hadoop-tools/hadoop-archives/src/test/java/org/apache/hadoop/tools/TestHadoopArchives.java Thu Apr 5 20:16:15 2012 @@ -103,7 +103,42 @@ public class TestHadoopArchives extends super.tearDown(); } - + + public void testRelativePath() throws Exception { + fs.delete(archivePath, true); + + final Path sub1 = new Path(inputPath, "dir1"); + fs.mkdirs(sub1); + createFile(sub1, "a", fs); + final Configuration conf = mapred.createJobConf(); + final FsShell shell = new FsShell(conf); + + final List<String> originalPaths = lsr(shell, "input"); + System.out.println("originalPath: " + originalPaths); + final URI uri = fs.getUri(); + final String prefix = "har://hdfs-" + uri.getHost() +":" + uri.getPort() + + archivePath.toUri().getPath() + Path.SEPARATOR; + + { + final String harName = "foo.har"; + final String[] args = { + "-archiveName", + harName, + "-p", + "input", + "*", + "archive" + }; + System.setProperty(HadoopArchives.TEST_HADOOP_ARCHIVES_JAR_PATH, HADOOP_ARCHIVES_JAR); + final HadoopArchives har = new HadoopArchives(mapred.createJobConf()); + assertEquals(0, ToolRunner.run(har, args)); + + //compare results + final List<String> harPaths = lsr(shell, prefix + harName); + assertEquals(originalPaths, harPaths); + } + } + public void testPathWithSpaces() throws Exception { fs.delete(archivePath, true); @@ -170,8 +205,11 @@ public class TestHadoopArchives extends System.setErr(oldErr); } System.out.println("lsr results:\n" + results); + String dirname = dir; + if (dir.lastIndexOf(Path.SEPARATOR) != -1 ) { + dirname = dir.substring(dir.lastIndexOf(Path.SEPARATOR)); + } - final String dirname = dir.substring(dir.lastIndexOf(Path.SEPARATOR)); final List<String> paths = new ArrayList<String>(); for(StringTokenizer t = new StringTokenizer(results, "\n"); t.hasMoreTokens(); ) {