HDFS-5578. [JDK8] Fix Javadoc errors caused by incorrect or illegal tags in doc comments. Contributed by Andrew Purtell.
Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/92916ae4 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/92916ae4 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/92916ae4 Branch: refs/heads/YARN-2139 Commit: 92916ae4876e4e24f85b22ccf9418f347d8d9666 Parents: 84d5000 Author: Haohui Mai <whe...@apache.org> Authored: Wed Dec 10 14:54:08 2014 -0800 Committer: Haohui Mai <whe...@apache.org> Committed: Wed Dec 10 14:54:37 2014 -0800 ---------------------------------------------------------------------- .../hadoop/fs/http/client/HttpFSFileSystem.java | 24 +++++------ .../fs/http/client/HttpsFSFileSystem.java | 8 ++-- .../server/CheckUploadContentTypeFilter.java | 4 +- .../http/server/HttpFSAuthenticationFilter.java | 2 +- .../fs/http/server/HttpFSExceptionProvider.java | 1 - .../hadoop/fs/http/server/HttpFSServer.java | 4 +- .../fs/http/server/HttpFSServerWebApp.java | 4 +- .../org/apache/hadoop/lib/lang/XException.java | 4 +- .../apache/hadoop/lib/server/BaseService.java | 16 ++++---- .../org/apache/hadoop/lib/server/Server.java | 42 ++++++++++---------- .../lib/servlet/FileSystemReleaseFilter.java | 6 +-- .../hadoop/lib/servlet/HostnameFilter.java | 6 +-- .../apache/hadoop/lib/servlet/MDCFilter.java | 6 +-- .../apache/hadoop/lib/servlet/ServerWebApp.java | 20 +++++----- .../java/org/apache/hadoop/lib/util/Check.java | 2 +- .../hadoop/lib/util/ConfigurationUtils.java | 2 +- .../org/apache/hadoop/lib/wsrs/Parameters.java | 4 +- .../nfs/nfs3/PrivilegedNfsGatewayStarter.java | 2 +- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 3 ++ 19 files changed, 81 insertions(+), 79 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSFileSystem.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSFileSystem.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSFileSystem.java index 367308d..5b079e9 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSFileSystem.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSFileSystem.java @@ -78,7 +78,7 @@ import java.util.Map; /** * HttpFSServer implementation of the FileSystemAccess FileSystem. - * <p/> + * <p> * This implementation allows a user to access HDFS over HTTP via a HttpFSServer server. */ @InterfaceAudience.Private @@ -223,7 +223,7 @@ public class HttpFSFileSystem extends FileSystem /** * Convenience method that creates a <code>HttpURLConnection</code> for the * HttpFSServer file system operations. - * <p/> + * <p> * This methods performs and injects any needed authentication credentials * via the {@link #getConnection(URL, String)} method * @@ -289,7 +289,7 @@ public class HttpFSFileSystem extends FileSystem /** * Convenience method that creates a <code>HttpURLConnection</code> for the specified URL. - * <p/> + * <p> * This methods performs and injects any needed authentication credentials. * * @param url url to connect to. @@ -371,7 +371,7 @@ public class HttpFSFileSystem extends FileSystem /** * HttpFSServer subclass of the <code>FSDataInputStream</code>. - * <p/> + * <p> * This implementation does not support the * <code>PositionReadable</code> and <code>Seekable</code> methods. */ @@ -414,8 +414,8 @@ public class HttpFSFileSystem extends FileSystem /** * Opens an FSDataInputStream at the indicated Path. - * </p> - * IMPORTANT: the returned <code><FSDataInputStream/code> does not support the + * <p> + * IMPORTANT: the returned <code>FSDataInputStream</code> does not support the * <code>PositionReadable</code> and <code>Seekable</code> methods. * * @param f the file name to open @@ -434,7 +434,7 @@ public class HttpFSFileSystem extends FileSystem /** * HttpFSServer subclass of the <code>FSDataOutputStream</code>. - * <p/> + * <p> * This implementation closes the underlying HTTP connection validating the Http connection status * at closing time. */ @@ -516,7 +516,7 @@ public class HttpFSFileSystem extends FileSystem /** * Opens an FSDataOutputStream at the indicated Path with write-progress * reporting. - * <p/> + * <p> * IMPORTANT: The <code>Progressable</code> parameter is not used. * * @param f the file name to open. @@ -549,7 +549,7 @@ public class HttpFSFileSystem extends FileSystem /** * Append to an existing file (optional operation). - * <p/> + * <p> * IMPORTANT: The <code>Progressable</code> parameter is not used. * * @param f the existing file to be appended. @@ -838,7 +838,7 @@ public class HttpFSFileSystem extends FileSystem * Modify the ACL entries for a file. * * @param path Path to modify - * @param aclSpec List<AclEntry> describing modifications + * @param aclSpec describing modifications * @throws IOException */ @Override @@ -855,7 +855,7 @@ public class HttpFSFileSystem extends FileSystem /** * Remove the specified ACL entries from a file * @param path Path to modify - * @param aclSpec List<AclEntry> describing entries to remove + * @param aclSpec describing entries to remove * @throws IOException */ @Override @@ -900,7 +900,7 @@ public class HttpFSFileSystem extends FileSystem /** * Set the ACLs for the given file * @param path Path to modify - * @param aclSpec List<AclEntry> describing modifications, must include + * @param aclSpec describing modifications, must include * entries for user, group, and others for compatibility * with permission bits. * @throws IOException http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpsFSFileSystem.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpsFSFileSystem.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpsFSFileSystem.java index a696cd8..0a2e08d 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpsFSFileSystem.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpsFSFileSystem.java @@ -18,10 +18,10 @@ package org.apache.hadoop.fs.http.client; /** - * HttpFSServer implementation of the FileSystemAccess FileSystem for SSL. - * <p/> - * This implementation allows a user to access HDFS over HTTPS via a - * HttpFSServer server. + * <p>HttpFSServer implementation of the FileSystemAccess FileSystem for SSL. + * </p> + * <p>This implementation allows a user to access HDFS over HTTPS via a + * HttpFSServer server.</p> */ public class HttpsFSFileSystem extends HttpFSFileSystem { http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/CheckUploadContentTypeFilter.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/CheckUploadContentTypeFilter.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/CheckUploadContentTypeFilter.java index 67df9a8..836b4ce 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/CheckUploadContentTypeFilter.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/CheckUploadContentTypeFilter.java @@ -50,7 +50,7 @@ public class CheckUploadContentTypeFilter implements Filter { /** * Initializes the filter. - * <p/> + * <p> * This implementation is a NOP. * * @param config filter configuration. @@ -103,7 +103,7 @@ public class CheckUploadContentTypeFilter implements Filter { /** * Destroys the filter. - * <p/> + * <p> * This implementation is a NOP. */ @Override http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSAuthenticationFilter.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSAuthenticationFilter.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSAuthenticationFilter.java index 8b332fc..f0fe4c5 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSAuthenticationFilter.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSAuthenticationFilter.java @@ -44,7 +44,7 @@ public class HttpFSAuthenticationFilter /** * Returns the hadoop-auth configuration from HttpFSServer's configuration. - * <p/> + * <p> * It returns all HttpFSServer's configuration properties prefixed with * <code>httpfs.authentication</code>. The <code>httpfs.authentication</code> * prefix is removed from the returned property names. http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSExceptionProvider.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSExceptionProvider.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSExceptionProvider.java index 3a8d9ad..aed6343 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSExceptionProvider.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSExceptionProvider.java @@ -43,7 +43,6 @@ public class HttpFSExceptionProvider extends ExceptionProvider { /** * Maps different exceptions thrown by HttpFSServer to HTTP status codes. - * <p/> * <ul> * <li>SecurityException : HTTP UNAUTHORIZED</li> * <li>FileNotFoundException : HTTP NOT_FOUND</li> http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java index f9eb454..9103718 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java @@ -85,7 +85,7 @@ import java.util.Map; /** * Main class of HttpFSServer server. - * <p/> + * <p> * The <code>HttpFSServer</code> class uses Jersey JAX-RS to binds HTTP requests to the * different operations. */ @@ -117,7 +117,7 @@ public class HttpFSServer { /** * Returns a filesystem instance. The fileystem instance is wired for release at the completion of * the current Servlet request via the {@link FileSystemReleaseFilter}. - * <p/> + * <p> * If a do-as user is specified, the current user must be a valid proxyuser, otherwise an * <code>AccessControlException</code> will be thrown. * http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServerWebApp.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServerWebApp.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServerWebApp.java index b7ae301..66438b5 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServerWebApp.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServerWebApp.java @@ -34,9 +34,9 @@ import java.io.IOException; * HttpFSServer server, it is a <code>javax.servlet.ServletContextListener * </code> implementation that is wired in HttpFSServer's WAR * <code>WEB-INF/web.xml</code>. - * <p/> + * <p> * It provides acces to the server context via the singleton {@link #get}. - * <p/> + * <p> * All the configuration is loaded from configuration properties prefixed * with <code>httpfs.</code>. */ http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/lang/XException.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/lang/XException.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/lang/XException.java index f974159..467ca23 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/lang/XException.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/lang/XException.java @@ -61,7 +61,7 @@ public class XException extends Exception { /** * Creates an XException using another XException as cause. - * <p/> + * <p> * The error code and error message are extracted from the cause. * * @param cause exception cause. @@ -95,7 +95,7 @@ public class XException extends Exception { /** * Creates a message using a error message template and arguments. - * <p/> + * <p> * The template must be in JDK <code>MessageFormat</code> syntax * (using {#} positional parameters). * http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/BaseService.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/BaseService.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/BaseService.java index 088f900..9d9ce7a 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/BaseService.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/BaseService.java @@ -44,11 +44,11 @@ public abstract class BaseService implements Service { /** * Initializes the service. - * <p/> + * <p> * It collects all service properties (properties having the * <code>#SERVER#.#SERVICE#.</code> prefix). The property names are then * trimmed from the <code>#SERVER#.#SERVICE#.</code> prefix. - * <p/> + * <p> * After collecting the service properties it delegates to the * {@link #init()} method. * @@ -75,7 +75,7 @@ public abstract class BaseService implements Service { /** * Post initializes the service. This method is called by the * {@link Server} after all services of the server have been initialized. - * <p/> + * <p> * This method does a NOP. * * @throws ServiceException thrown if the service could not be @@ -88,7 +88,7 @@ public abstract class BaseService implements Service { /** * Destroy the services. This method is called once, when the * {@link Server} owning the service is being destroyed. - * <p/> + * <p> * This method does a NOP. */ @Override @@ -98,7 +98,7 @@ public abstract class BaseService implements Service { /** * Returns the service dependencies of this service. The service will be * instantiated only if all the service dependencies are already initialized. - * <p/> + * <p> * This method returns an empty array (size 0) * * @return an empty array (size 0). @@ -110,7 +110,7 @@ public abstract class BaseService implements Service { /** * Notification callback when the server changes its status. - * <p/> + * <p> * This method returns an empty array (size 0) * * @param oldStatus old server status. @@ -154,7 +154,7 @@ public abstract class BaseService implements Service { /** * Returns the service configuration properties. Property * names are trimmed off from its prefix. - * <p/> + * <p> * The sevice configuration properties are all properties * with names starting with <code>#SERVER#.#SERVICE#.</code> * in the server configuration. @@ -169,7 +169,7 @@ public abstract class BaseService implements Service { /** * Initializes the server. - * <p/> + * <p> * This method is called by {@link #init(Server)} after all service properties * (properties prefixed with * http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java index d083831..5c1bb4f 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java @@ -42,40 +42,39 @@ import java.util.Properties; /** * A Server class provides standard configuration, logging and {@link Service} * lifecyle management. - * <p/> + * <p> * A Server normally has a home directory, a configuration directory, a temp * directory and logs directory. - * <p/> + * <p> * The Server configuration is loaded from 2 overlapped files, * <code>#SERVER#-default.xml</code> and <code>#SERVER#-site.xml</code>. The * default file is loaded from the classpath, the site file is laoded from the * configuration directory. - * <p/> + * <p> * The Server collects all configuration properties prefixed with * <code>#SERVER#</code>. The property names are then trimmed from the * <code>#SERVER#</code> prefix. - * <p/> + * <p> * The Server log configuration is loaded from the * <code>#SERVICE#-log4j.properties</code> file in the configuration directory. - * <p/> + * <p> * The lifecycle of server is defined in by {@link Server.Status} enum. * When a server is create, its status is UNDEF, when being initialized it is * BOOTING, once initialization is complete by default transitions to NORMAL. * The <code>#SERVER#.startup.status</code> configuration property can be used * to specify a different startup status (NORMAL, ADMIN or HALTED). - * <p/> + * <p> * Services classes are defined in the <code>#SERVER#.services</code> and * <code>#SERVER#.services.ext</code> properties. They are loaded in order * (services first, then services.ext). - * <p/> + * <p> * Before initializing the services, they are traversed and duplicate service * interface are removed from the service list. The last service using a given * interface wins (this enables a simple override mechanism). - * <p/> + * <p> * After the services have been resoloved by interface de-duplication they are * initialized in order. Once all services are initialized they are * post-initialized (this enables late/conditional service bindings). - * <p/> */ @InterfaceAudience.Private public class Server { @@ -152,7 +151,7 @@ public class Server { /** * Creates a server instance. - * <p/> + * <p> * The config, log and temp directories are all under the specified home directory. * * @param name server name. @@ -177,9 +176,9 @@ public class Server { /** * Creates a server instance. - * <p/> + * <p> * The config, log and temp directories are all under the specified home directory. - * <p/> + * <p> * It uses the provided configuration instead loading it from the config dir. * * @param name server name. @@ -192,7 +191,7 @@ public class Server { /** * Creates a server instance. - * <p/> + * <p> * It uses the provided configuration instead loading it from the config dir. * * @param name server name. @@ -250,9 +249,9 @@ public class Server { /** * Sets a new server status. - * <p/> + * <p> * The status must be settable. - * <p/> + * <p> * All services will be notified o the status change via the * {@link Service#serverStatusChange(Server.Status, Server.Status)} method. If a service * throws an exception during the notification, the server will be destroyed. @@ -299,7 +298,7 @@ public class Server { /** * Convenience method that returns a resource as inputstream from the * classpath. - * <p/> + * <p> * It first attempts to use the Thread's context classloader and if not * set it uses the <code>ClassUtils</code> classloader. * @@ -319,7 +318,7 @@ public class Server { /** * Initializes the Server. - * <p/> + * <p> * The initialization steps are: * <ul> * <li>It verifies the service home and temp directories exist</li> @@ -335,6 +334,7 @@ public class Server { * <li>Initializes the services</li> * <li>Post-initializes the services</li> * <li>Sets the server startup status</li> + * </ul> * * @throws ServerException thrown if the server could not be initialized. */ @@ -625,7 +625,7 @@ public class Server { /** * Destroys the server. - * <p/> + * <p> * All services are destroyed in reverse order of initialization, then the * Log4j framework is shutdown. */ @@ -651,7 +651,7 @@ public class Server { /** * Returns the server prefix for server configuration properties. - * <p/> + * <p> * By default it is the server name. * * @return the prefix for server configuration properties. @@ -733,10 +733,10 @@ public class Server { /** * Adds a service programmatically. - * <p/> + * <p> * If a service with the same interface exists, it will be destroyed and * removed before the given one is initialized and added. - * <p/> + * <p> * If an exception is thrown the server is destroyed. * * @param klass service class to add. http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/FileSystemReleaseFilter.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/FileSystemReleaseFilter.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/FileSystemReleaseFilter.java index 827bcff..cf73979 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/FileSystemReleaseFilter.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/FileSystemReleaseFilter.java @@ -33,7 +33,7 @@ import java.io.IOException; /** * The <code>FileSystemReleaseFilter</code> releases back to the * {@link FileSystemAccess} service a <code>FileSystem</code> instance. - * <p/> + * <p> * This filter is useful in situations where a servlet request * is streaming out HDFS data and the corresponding filesystem * instance have to be closed after the streaming completes. @@ -44,7 +44,7 @@ public abstract class FileSystemReleaseFilter implements Filter { /** * Initializes the filter. - * <p/> + * <p> * This implementation is a NOP. * * @param filterConfig filter configuration. @@ -83,7 +83,7 @@ public abstract class FileSystemReleaseFilter implements Filter { /** * Destroys the filter. - * <p/> + * <p> * This implementation is a NOP. */ @Override http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/HostnameFilter.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/HostnameFilter.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/HostnameFilter.java index dd395f6..64f4926 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/HostnameFilter.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/HostnameFilter.java @@ -43,7 +43,7 @@ public class HostnameFilter implements Filter { /** * Initializes the filter. - * <p/> + * <p> * This implementation is a NOP. * * @param config filter configuration. @@ -56,7 +56,7 @@ public class HostnameFilter implements Filter { /** * Resolves the requester hostname and delegates the request to the chain. - * <p/> + * <p> * The requester hostname is available via the {@link #get} method. * * @param request servlet request. @@ -101,7 +101,7 @@ public class HostnameFilter implements Filter { /** * Destroys the filter. - * <p/> + * <p> * This implementation is a NOP. */ @Override http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/MDCFilter.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/MDCFilter.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/MDCFilter.java index 07b552d..156cf64 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/MDCFilter.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/MDCFilter.java @@ -33,7 +33,7 @@ import java.security.Principal; /** * Filter that sets request contextual information for the slf4j MDC. - * <p/> + * <p> * It sets the following values: * <ul> * <li>hostname: if the {@link HostnameFilter} is present and configured @@ -48,7 +48,7 @@ public class MDCFilter implements Filter { /** * Initializes the filter. - * <p/> + * <p> * This implementation is a NOP. * * @param config filter configuration. @@ -93,7 +93,7 @@ public class MDCFilter implements Filter { /** * Destroys the filter. - * <p/> + * <p> * This implementation is a NOP. */ @Override http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java index 9b0ea2a..cd16593 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java @@ -75,21 +75,21 @@ public abstract class ServerWebApp extends Server implements ServletContextListe /** * Constructor. Subclasses must have a default constructor specifying * the server name. - * <p/> + * <p> * The server name is used to resolve the Java System properties that define * the server home, config, log and temp directories. - * <p/> + * <p> * The home directory is looked in the Java System property * <code>#SERVER_NAME#.home.dir</code>. - * <p/> + * <p> * The config directory is looked in the Java System property * <code>#SERVER_NAME#.config.dir</code>, if not defined it resolves to * the <code>#SERVER_HOME_DIR#/conf</code> directory. - * <p/> + * <p> * The log directory is looked in the Java System property * <code>#SERVER_NAME#.log.dir</code>, if not defined it resolves to * the <code>#SERVER_HOME_DIR#/log</code> directory. - * <p/> + * <p> * The temp directory is looked in the Java System property * <code>#SERVER_NAME#.temp.dir</code>, if not defined it resolves to * the <code>#SERVER_HOME_DIR#/temp</code> directory. @@ -105,7 +105,7 @@ public abstract class ServerWebApp extends Server implements ServletContextListe /** * Returns the server home directory. - * <p/> + * <p> * It is looked up in the Java System property * <code>#SERVER_NAME#.home.dir</code>. * @@ -159,15 +159,15 @@ public abstract class ServerWebApp extends Server implements ServletContextListe } /** - * Resolves the host & port InetSocketAddress the web server is listening to. - * <p/> + * Resolves the host and port InetSocketAddress the web server is listening to. + * <p> * This implementation looks for the following 2 properties: * <ul> * <li>#SERVER_NAME#.http.hostname</li> * <li>#SERVER_NAME#.http.port</li> * </ul> * - * @return the host & port InetSocketAddress the web server is listening to. + * @return the host and port InetSocketAddress the web server is listening to. * @throws ServerException thrown if any of the above 2 properties is not defined. */ protected InetSocketAddress resolveAuthority() throws ServerException { @@ -217,7 +217,7 @@ public abstract class ServerWebApp extends Server implements ServletContextListe /** * Sets an alternate hostname:port InetSocketAddress to use. - * <p/> + * <p> * For testing purposes. * * @param authority alterante authority. http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/util/Check.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/util/Check.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/util/Check.java index a398e75..31666e8 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/util/Check.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/util/Check.java @@ -26,7 +26,7 @@ import java.util.regex.Pattern; /** * Utility methods to check preconditions. - * <p/> + * <p> * Commonly used for method arguments preconditions. */ @InterfaceAudience.Private http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/util/ConfigurationUtils.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/util/ConfigurationUtils.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/util/ConfigurationUtils.java index 660eae0..6611dd2 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/util/ConfigurationUtils.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/util/ConfigurationUtils.java @@ -90,7 +90,7 @@ public abstract class ConfigurationUtils { /** * Create a configuration from an InputStream. - * <p/> + * <p> * ERROR canibalized from <code>Configuration.loadResource()</code>. * * @param is inputstream to read the configuration from. http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/Parameters.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/Parameters.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/Parameters.java index 0f16a9b..e0f6200 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/Parameters.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/wsrs/Parameters.java @@ -26,7 +26,7 @@ import java.util.Map; /** * Class that contains all parsed JAX-RS parameters. - * <p/> + * <p> * Instances are created by the {@link ParametersProvider} class. */ @InterfaceAudience.Private @@ -63,7 +63,7 @@ public class Parameters { * * @param name parameter name. * @param klass class of the parameter, used for value casting. - * @return List<V> the values of the parameter. + * @return the values of the parameter. */ @SuppressWarnings("unchecked") public <V, T extends Param<V>> List<V> getValues(String name, Class<T> klass) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/PrivilegedNfsGatewayStarter.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/PrivilegedNfsGatewayStarter.java b/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/PrivilegedNfsGatewayStarter.java index 98862ed..3934d7c 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/PrivilegedNfsGatewayStarter.java +++ b/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/PrivilegedNfsGatewayStarter.java @@ -26,7 +26,7 @@ import org.apache.hadoop.hdfs.nfs.conf.NfsConfiguration; /** * This class is used to allow the initial registration of the NFS gateway with - * the system portmap daemon to come from a privileged (< 1024) port. This is + * the system portmap daemon to come from a privileged (< 1024) port. This is * necessary on certain operating systems to work around this bug in rpcbind: * * Red Hat: https://bugzilla.redhat.com/show_bug.cgi?id=731542 http://git-wip-us.apache.org/repos/asf/hadoop/blob/92916ae4/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 9f3f9ee..1db358f 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -564,6 +564,9 @@ Release 2.7.0 - UNRELEASED HDFS-7502. Fix findbugs warning in hdfs-nfs project. (Brandon Li via wheat9) + HDFS-5578. [JDK8] Fix Javadoc errors caused by incorrect or illegal tags + in doc comments. (Andrew Purtell via wheat9) + Release 2.6.1 - UNRELEASED INCOMPATIBLE CHANGES