Repository: hadoop Updated Branches: refs/heads/branch-2 bb828565f -> 77297017d
HADOOP-11720. [JDK8] Fix javadoc errors caused by incorrect or illegal tags in hadoop-tools. Contributed by Akira AJISAKA. (cherry picked from commit ef9946cd52d54200c658987c1dbc3e6fce133f77) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/77297017 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/77297017 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/77297017 Branch: refs/heads/branch-2 Commit: 77297017d88ff795622619375445b8aa9ee0c95d Parents: bb82856 Author: Tsuyoshi Ozawa <[email protected]> Authored: Tue Mar 17 16:09:21 2015 +0900 Committer: Tsuyoshi Ozawa <[email protected]> Committed: Tue Mar 17 16:09:38 2015 +0900 ---------------------------------------------------------------------- hadoop-common-project/hadoop-common/CHANGES.txt | 3 +++ .../java/org/apache/hadoop/record/Buffer.java | 8 +++---- .../java/org/apache/hadoop/record/Utils.java | 8 +++---- .../java/org/apache/hadoop/ant/DfsTask.java | 6 ++--- .../org/apache/hadoop/fs/s3/S3FileSystem.java | 4 +--- .../hadoop/fs/s3a/S3AFastOutputStream.java | 4 ++-- .../hadoop/fs/s3native/NativeS3FileSystem.java | 24 ++++++++++++-------- .../fs/azure/AzureNativeFileSystemStore.java | 22 ++++++------------ .../hadoop/fs/azure/NativeAzureFileSystem.java | 16 +++++-------- .../hadoop/tools/CopyListingFileStatus.java | 8 +++---- .../apache/hadoop/tools/SimpleCopyListing.java | 2 +- .../apache/hadoop/tools/util/DistCpUtils.java | 4 ++-- 12 files changed, 51 insertions(+), 58 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hadoop/blob/77297017/hadoop-common-project/hadoop-common/CHANGES.txt ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index a1d2125..7f47197 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -697,6 +697,9 @@ Release 2.7.0 - UNRELEASED HADOOP-11638. OpensslSecureRandom.c pthreads_thread_id should support FreeBSD and Solaris in addition to Linux. (Kiran Kumar M R via cnauroth) + HADOOP-11720. [JDK8] Fix javadoc errors caused by incorrect or illegal + tags in hadoop-tools. (Akira AJISAKA via ozawa) + Release 2.6.1 - UNRELEASED INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/77297017/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/Buffer.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/Buffer.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/Buffer.java index 50cc1a1..737d63d 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/Buffer.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/Buffer.java @@ -25,10 +25,10 @@ import org.apache.hadoop.classification.InterfaceStability; /** * A byte sequence that is used as a Java native type for buffer. - * It is resizable and distinguishes between the count of the seqeunce and + * It is resizable and distinguishes between the count of the sequence and * the current capacity. * - * @deprecated Replaced by <a href="http://hadoop.apache.org/avro/">Avro</a>. + * @deprecated Replaced by <a href="http://avro.apache.org/">Avro</a>. */ @Deprecated @InterfaceAudience.Public @@ -124,7 +124,7 @@ public class Buffer implements Comparable, Cloneable { /** * Change the capacity of the backing storage. - * The data is preserved if newCapacity >= getCount(). + * The data is preserved if newCapacity {@literal >=} getCount(). * @param newCapacity The new capacity in bytes. */ public void setCapacity(int newCapacity) { @@ -162,7 +162,7 @@ public class Buffer implements Comparable, Cloneable { public void truncate() { setCapacity(count); } - + /** * Append specified bytes to the buffer. * http://git-wip-us.apache.org/repos/asf/hadoop/blob/77297017/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/Utils.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/Utils.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/Utils.java index d5be59c..59e2080 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/Utils.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/Utils.java @@ -28,9 +28,9 @@ import org.apache.hadoop.io.WritableComparator; import org.apache.hadoop.io.WritableUtils; /** - * Various utility functions for Hadooop record I/O runtime. + * Various utility functions for Hadoop record I/O runtime. * - * @deprecated Replaced by <a href="http://hadoop.apache.org/avro/">Avro</a>. + * @deprecated Replaced by <a href="http://avro.apache.org/">Avro</a>. */ @Deprecated @InterfaceAudience.Public @@ -462,8 +462,8 @@ public class Utils { /** * Serializes a long to a binary stream with zero-compressed encoding. - * For -112 <= i <= 127, only one byte is used with the actual value. - * For other values of i, the first byte value indicates whether the + * For {@literal -112 <= i <= 127}, only one byte is used with the actual + * value. For other values of i, the first byte value indicates whether the * long is positive or negative, and the number of bytes that follow. * If the first byte value v is between -113 and -120, the following long * is positive, with number of bytes that follow are -(v+112). http://git-wip-us.apache.org/repos/asf/hadoop/blob/77297017/hadoop-tools/hadoop-ant/src/main/java/org/apache/hadoop/ant/DfsTask.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-ant/src/main/java/org/apache/hadoop/ant/DfsTask.java b/hadoop-tools/hadoop-ant/src/main/java/org/apache/hadoop/ant/DfsTask.java index 78cb360..9d0b3a4 100644 --- a/hadoop-tools/hadoop-ant/src/main/java/org/apache/hadoop/ant/DfsTask.java +++ b/hadoop-tools/hadoop-ant/src/main/java/org/apache/hadoop/ant/DfsTask.java @@ -41,8 +41,8 @@ import org.apache.hadoop.hdfs.HdfsConfiguration; public class DfsTask extends Task { /** - * Default sink for {@link java.lang.System.out System.out} - * and {@link java.lang.System.err System.err}. + * Default sink for {@link java.lang.System#out} + * and {@link java.lang.System#err}. */ private static final OutputStream nullOut = new OutputStream() { public void write(int b) { /* ignore */ } @@ -171,7 +171,7 @@ public class DfsTask extends Task { } /** - * Invoke {@link org.apache.hadoop.fs.FsShell#doMain FsShell.doMain} after a + * Invoke {@link org.apache.hadoop.fs.FsShell#main} after a * few cursory checks of the configuration. */ public void execute() throws BuildException { http://git-wip-us.apache.org/repos/asf/hadoop/blob/77297017/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3/S3FileSystem.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3/S3FileSystem.java b/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3/S3FileSystem.java index 5afc900..5f0ad83 100644 --- a/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3/S3FileSystem.java +++ b/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3/S3FileSystem.java @@ -44,10 +44,9 @@ import org.apache.hadoop.io.retry.RetryProxy; import org.apache.hadoop.util.Progressable; /** - * <p> * A block-based {@link FileSystem} backed by * <a href="http://aws.amazon.com/s3">Amazon S3</a>. - * </p> + * * @see NativeS3FileSystem */ @InterfaceAudience.Public @@ -70,7 +69,6 @@ public class S3FileSystem extends FileSystem { /** * Return the protocol scheme for the FileSystem. - * <p/> * * @return <code>s3</code> */ http://git-wip-us.apache.org/repos/asf/hadoop/blob/77297017/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFastOutputStream.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFastOutputStream.java b/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFastOutputStream.java index a29c47b..6819581 100644 --- a/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFastOutputStream.java +++ b/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFastOutputStream.java @@ -57,11 +57,11 @@ import java.util.concurrent.ThreadPoolExecutor; /** * Upload files/parts asap directly from a memory buffer (instead of buffering * to a file). - * <p/> + * <p> * Uploads are managed low-level rather than through the AWS TransferManager. * This allows for uploading each part of a multi-part upload as soon as * the bytes are in memory, rather than waiting until the file is closed. - * <p/> + * <p> * Unstable: statistics and error handling might evolve */ @InterfaceStability.Unstable http://git-wip-us.apache.org/repos/asf/hadoop/blob/77297017/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3native/NativeS3FileSystem.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3native/NativeS3FileSystem.java b/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3native/NativeS3FileSystem.java index e490edf..c34d53e 100644 --- a/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3native/NativeS3FileSystem.java +++ b/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3native/NativeS3FileSystem.java @@ -62,24 +62,29 @@ import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** - * <p> * A {@link FileSystem} for reading and writing files stored on * <a href="http://aws.amazon.com/s3">Amazon S3</a>. * Unlike {@link org.apache.hadoop.fs.s3.S3FileSystem} this implementation * stores files on S3 in their * native form so they can be read by other S3 tools. - * + * <p> * A note about directories. S3 of course has no "native" support for them. * The idiom we choose then is: for any directory created by this class, * we use an empty object "#{dirpath}_$folder$" as a marker. * Further, to interoperate with other S3 tools, we also accept the following: - * - an object "#{dirpath}/' denoting a directory marker - * - if there exists any objects with the prefix "#{dirpath}/", then the - * directory is said to exist - * - if both a file with the name of a directory and a marker for that - * directory exists, then the *file masks the directory*, and the directory - * is never returned. - * </p> + * <ul> + * <li>an object "#{dirpath}/' denoting a directory marker</li> + * <li> + * if there exists any objects with the prefix "#{dirpath}/", then the + * directory is said to exist + * </li> + * <li> + * if both a file with the name of a directory and a marker for that + * directory exists, then the *file masks the directory*, and the directory + * is never returned. + * </li> + * </ul> + * * @see org.apache.hadoop.fs.s3.S3FileSystem */ @InterfaceAudience.Public @@ -308,7 +313,6 @@ public class NativeS3FileSystem extends FileSystem { /** * Return the protocol scheme for the FileSystem. - * <p/> * * @return <code>s3n</code> */ http://git-wip-us.apache.org/repos/asf/hadoop/blob/77297017/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java index c6ba84f..5dc0963 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java @@ -600,8 +600,6 @@ public class AzureNativeFileSystemStore implements NativeFileSystemStore { * Azure. * * @throws AzureException - * @throws ConfigurationException - * */ private void configureAzureStorageSession() throws AzureException { @@ -705,7 +703,7 @@ public class AzureNativeFileSystemStore implements NativeFileSystemStore { * raised on errors communicating with Azure storage. * @throws IOException * raised on errors performing I/O or setting up the session. - * @throws URISyntaxExceptions + * @throws URISyntaxException * raised on creating mal-formed URI's. */ private void connectUsingAnonymousCredentials(final URI uri) @@ -1036,7 +1034,6 @@ public class AzureNativeFileSystemStore implements NativeFileSystemStore { /** * Checks if the given key in Azure Storage should be stored as a page * blob instead of block blob. - * @throws URISyntaxException */ public boolean isPageBlobKey(String key) { return isKeyForDirectorySet(key, pageBlobDirs); @@ -1755,7 +1752,7 @@ public class AzureNativeFileSystemStore implements NativeFileSystemStore { * the path and returns a path relative to the root directory of the * container. * - * @param blob + * @param directory * - adjust the key to this directory to a path relative to the root * directory * @@ -2142,14 +2139,10 @@ public class AzureNativeFileSystemStore implements NativeFileSystemStore { * uses a in-order first traversal of blob directory structures to maintain * the sorted order of the blob names. * - * @param dir - * -- Azure blob directory - * - * @param list - * -- a list of file metadata objects for each non-directory blob. - * - * @param maxListingLength - * -- maximum length of the built up list. + * @param aCloudBlobDirectory Azure blob directory + * @param aFileMetadataList a list of file metadata objects for each + * non-directory blob. + * @param maxListingCount maximum length of the built up list. */ private void buildUpList(CloudBlobDirectoryWrapper aCloudBlobDirectory, ArrayList<FileMetadata> aFileMetadataList, final int maxListingCount, @@ -2320,8 +2313,7 @@ public class AzureNativeFileSystemStore implements NativeFileSystemStore { * swallow the error since what most probably happened is that * the first operation succeeded on the server. * @param blob The blob to delete. - * @param leaseID A string identifying the lease, or null if no - * lease is to be used. + * @param lease Azure blob lease, or null if no lease is to be used. * @throws StorageException */ private void safeDelete(CloudBlobWrapper blob, SelfRenewingLease lease) throws StorageException { http://git-wip-us.apache.org/repos/asf/hadoop/blob/77297017/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/NativeAzureFileSystem.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/NativeAzureFileSystem.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/NativeAzureFileSystem.java index e39b37d..623645a 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/NativeAzureFileSystem.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/NativeAzureFileSystem.java @@ -81,12 +81,10 @@ import com.microsoft.azure.storage.blob.CloudBlob; import com.microsoft.azure.storage.core.*; /** - * <p> * A {@link FileSystem} for reading and writing files stored on <a * href="http://store.azure.com/">Windows Azure</a>. This implementation is * blob-based and stores files on Azure in their native form so they can be read * by other Azure tools. - * </p> */ @InterfaceAudience.Public @InterfaceStability.Stable @@ -218,9 +216,11 @@ public class NativeAzureFileSystem extends FileSystem { } /** - * Write to disk the information needed to redo folder rename, in JSON format. - * The file name will be wasb://<sourceFolderPrefix>/folderName-RenamePending.json + * Write to disk the information needed to redo folder rename, + * in JSON format. The file name will be + * {@code wasb://<sourceFolderPrefix>/folderName-RenamePending.json} * The file format will be: + * <pre>{@code * { * FormatVersion: "1.0", * OperationTime: "<YYYY-MM-DD HH:MM:SS.MMM>", @@ -239,7 +239,7 @@ public class NativeAzureFileSystem extends FileSystem { * "innerFile", * "innerFile2" * ] - * } + * } }</pre> * @throws IOException */ public void writeFile(FileSystem fs) throws IOException { @@ -913,9 +913,6 @@ public class NativeAzureFileSystem extends FileSystem { * The create also includes the name of the original key value which is * stored in the m_key member variable. This method should only be called * when the stream is closed. - * - * @param anEncodedKey - * Encoding of the original key stored in m_key member. */ private void restoreKey() throws IOException { store.rename(getEncodedKey(), getKey()); @@ -1796,7 +1793,7 @@ public class NativeAzureFileSystem extends FileSystem { * * @param permission * The permission to mask. - * @param applyDefaultUmask + * @param applyMode * Whether to also apply the default umask. * @return The masked persmission. */ @@ -2409,7 +2406,6 @@ public class NativeAzureFileSystem extends FileSystem { * recover the original key. * * @param aKey - * @param numBuckets * @return Encoded version of the original key. */ private static String encodeKey(String aKey) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/77297017/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/CopyListingFileStatus.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/CopyListingFileStatus.java b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/CopyListingFileStatus.java index 9b31dfb..8af799a 100644 --- a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/CopyListingFileStatus.java +++ b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/CopyListingFileStatus.java @@ -79,7 +79,7 @@ public final class CopyListingFileStatus extends FileStatus { /** * Returns the full logical ACL. * - * @return List<AclEntry> containing full logical ACL + * @return List containing full logical ACL */ public List<AclEntry> getAclEntries() { return AclUtil.getAclFromPermAndEntries(getPermission(), @@ -89,7 +89,7 @@ public final class CopyListingFileStatus extends FileStatus { /** * Sets optional ACL entries. * - * @param aclEntries List<AclEntry> containing all ACL entries + * @param aclEntries List containing all ACL entries */ public void setAclEntries(List<AclEntry> aclEntries) { this.aclEntries = aclEntries; @@ -98,7 +98,7 @@ public final class CopyListingFileStatus extends FileStatus { /** * Returns all xAttrs. * - * @return Map<String, byte[]> containing all xAttrs + * @return Map containing all xAttrs */ public Map<String, byte[]> getXAttrs() { return xAttrs != null ? xAttrs : Collections.<String, byte[]>emptyMap(); @@ -107,7 +107,7 @@ public final class CopyListingFileStatus extends FileStatus { /** * Sets optional xAttrs. * - * @param xAttrs Map<String, byte[]> containing all xAttrs + * @param xAttrs Map containing all xAttrs */ public void setXAttrs(Map<String, byte[]> xAttrs) { this.xAttrs = xAttrs; http://git-wip-us.apache.org/repos/asf/hadoop/blob/77297017/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/SimpleCopyListing.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/SimpleCopyListing.java b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/SimpleCopyListing.java index f9cfc86..6dc827a 100644 --- a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/SimpleCopyListing.java +++ b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/SimpleCopyListing.java @@ -141,7 +141,7 @@ public class SimpleCopyListing extends CopyListing { } /** * Collect the list of - * <sourceRelativePath, sourceFileStatus> + * {@literal <sourceRelativePath, sourceFileStatus>} * to be copied and write to the sequence file. In essence, any file or * directory that need to be copied or sync-ed is written as an entry to the * sequence file, with the possible exception of the source root: http://git-wip-us.apache.org/repos/asf/hadoop/blob/77297017/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java index 20fdf11..d34faaf 100644 --- a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java +++ b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java @@ -270,7 +270,7 @@ public class DistCpUtils { * * @param fileSystem FileSystem containing the file * @param fileStatus FileStatus of file - * @return List<AclEntry> containing full logical ACL + * @return List containing full logical ACL * @throws IOException if there is an I/O error */ public static List<AclEntry> getAcl(FileSystem fileSystem, @@ -285,7 +285,7 @@ public class DistCpUtils { * * @param fileSystem FileSystem containing the file * @param path file path - * @return Map<String, byte[]> containing all xAttrs + * @return Map containing all xAttrs * @throws IOException if there is an I/O error */ public static Map<String, byte[]> getXAttrs(FileSystem fileSystem,
