[FLINK-2559] Clean up JavaDocs

- Remove broken HTML tags like <br/>, <p/>, ...
- close unclosed HTML tags
- replaces special chars by HTML escaping, e.g., '<' by &lt;
- wrap code examples by {@code}
- fix incorrect @see and @link references
- fix incorrect @throws
- fix typos

This closes #1298


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/680b5a90
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/680b5a90
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/680b5a90

Branch: refs/heads/master
Commit: 680b5a90de2428b4b6b986f7b40409c2d1d181a7
Parents: ec7bf50
Author: Hubert Czerpak <hubert.czer...@gmail.com>
Authored: Thu Oct 22 22:51:25 2015 +0100
Committer: Fabian Hueske <fhue...@apache.org>
Committed: Thu Oct 29 10:52:22 2015 +0100

----------------------------------------------------------------------
 .../org/apache/flink/client/CliFrontend.java    |   3 +-
 .../storm/excamation/ExclamationLocal.java      |  13 +-
 .../storm/excamation/ExclamationTopology.java   |  13 +-
 .../storm/excamation/ExclamationWithBolt.java   |  13 +-
 .../storm/excamation/ExclamationWithSpout.java  |  13 +-
 .../flink/storm/split/SpoutSplitExample.java    |   4 +-
 .../storm/wordcount/BoltTokenizerWordCount.java |   9 +-
 .../wordcount/BoltTokenizerWordCountPojo.java   |   9 +-
 .../BoltTokenizerWordCountWithNames.java        |   9 +-
 .../storm/wordcount/SpoutSourceWordCount.java   |  11 +-
 .../flink/storm/wordcount/WordCountLocal.java   |  11 +-
 .../storm/wordcount/WordCountLocalByName.java   |  11 +-
 .../wordcount/WordCountRemoteByClient.java      |  11 +-
 .../wordcount/WordCountRemoteBySubmitter.java   |  11 +-
 .../storm/wordcount/WordCountTopology.java      |   9 +-
 .../org/apache/flink/storm/api/FlinkClient.java |   2 +-
 .../flink/storm/api/FlinkTopologyBuilder.java   |   4 +-
 .../flink/storm/util/SplitStreamType.java       |   3 +-
 .../flink/storm/wrappers/BoltWrapper.java       |   4 +-
 .../flink/storm/wrappers/SpoutWrapper.java      |  10 +-
 .../model/tweet/entities/Entities.java          |   2 +-
 .../model/tweet/entities/HashTags.java          |   2 +-
 .../model/tweet/entities/Media.java             |   2 +-
 .../model/tweet/entities/Symbol.java            |   2 +-
 .../model/tweet/entities/URL.java               |   2 +-
 .../model/tweet/entities/UserMention.java       |   2 +-
 .../flink/api/common/ExecutionConfig.java       |   2 +-
 .../apache/flink/api/common/ExecutionMode.java  |   8 +-
 .../api/common/accumulators/Histogram.java      |   2 +-
 .../api/common/aggregators/Aggregator.java      |   2 +-
 .../common/distributions/DataDistribution.java  |   2 +-
 .../api/common/functions/CoGroupFunction.java   |   4 +-
 .../api/common/functions/CrossFunction.java     |   4 +-
 .../api/common/functions/FilterFunction.java    |   4 +-
 .../api/common/functions/FlatJoinFunction.java  |   4 +-
 .../api/common/functions/FlatMapFunction.java   |   4 +-
 .../api/common/functions/FoldFunction.java      |   4 +-
 .../common/functions/GroupReduceFunction.java   |   4 +-
 .../api/common/functions/JoinFunction.java      |   4 +-
 .../flink/api/common/functions/MapFunction.java |   4 +-
 .../common/functions/MapPartitionFunction.java  |   4 +-
 .../api/common/functions/ReduceFunction.java    |   4 +-
 .../functions/RichGroupReduceFunction.java      |   2 +-
 .../flink/api/common/io/FileOutputFormat.java   |   2 +-
 .../common/operators/AbstractUdfOperator.java   |   2 +-
 .../api/common/typeutils/TypeComparator.java    |   8 +-
 .../flink/configuration/ConfigConstants.java    |   2 +-
 .../org/apache/flink/core/fs/FileSystem.java    |  18 ++-
 .../apache/flink/core/memory/MemorySegment.java |   4 +-
 .../java/org/apache/flink/types/Record.java     |   2 +-
 .../org/apache/flink/types/StringValue.java     |   4 +-
 .../java/org/apache/flink/util/Visitable.java   |   4 +-
 .../flink/examples/java/clustering/KMeans.java  |   4 +-
 .../clustering/util/KMeansDataGenerator.java    |   2 +
 .../flink/examples/java/distcp/DistCp.java      |   2 +-
 .../examples/java/graph/PageRankBasic.java      |   2 +-
 .../examples/java/ml/LinearRegression.java      |   4 +-
 .../examples/java/relational/TPCHQuery10.java   |   4 +-
 .../examples/java/relational/TPCHQuery3.java    |   4 +-
 .../java/relational/WebLogAnalysis.java         |   8 +-
 .../examples/java/wordcount/PojoExample.java    |   2 +-
 .../examples/java/wordcount/WordCount.java      |   2 +-
 .../examples/java/wordcount/WordCountMeta.java  |   4 +-
 .../java/org/apache/flink/api/java/DataSet.java | 124 +++++++++----------
 .../java/org/apache/flink/api/java/Utils.java   |   1 -
 .../api/java/functions/FunctionAnnotation.java  |  30 ++---
 .../java/hadoop/mapred/HadoopInputFormat.java   |   2 +-
 .../java/hadoop/mapred/HadoopOutputFormat.java  |   2 +-
 .../flink/api/java/io/SplitDataProperties.java  |  16 +--
 .../api/java/operators/CoGroupOperator.java     |  42 +++----
 .../flink/api/java/operators/CrossOperator.java |  30 ++---
 .../flink/api/java/operators/DataSink.java      |   8 +-
 .../flink/api/java/operators/Grouping.java      |   2 +-
 .../flink/api/java/operators/JoinOperator.java  |  52 ++++----
 .../java/operators/SingleInputUdfOperator.java  |   6 +-
 .../api/java/operators/SortedGrouping.java      |  14 +--
 .../api/java/operators/TwoInputUdfOperator.java |  12 +-
 .../api/java/operators/UnsortedGrouping.java    |  22 ++--
 .../operators/join/JoinOperatorSetsBase.java    |  26 ++--
 .../translation/Tuple3WrappingCollector.java    |   2 +-
 .../translation/TupleWrappingCollector.java     |   2 +-
 .../record/functions/FunctionAnnotation.java    |   6 +-
 .../record/io/ExternalProcessInputFormat.java   |   2 +-
 .../flink/api/java/typeutils/AvroTypeInfo.java  |   4 +-
 .../typeutils/runtime/kryo/Serializers.java     |   4 +-
 .../flink/api/java/utils/DataSetUtils.java      |   4 +-
 .../examples/java8/relational/TPCHQuery10.java  |   4 +-
 .../main/java/org/apache/flink/graph/Graph.java |  30 ++---
 .../graph/example/EuclideanGraphWeighing.java   |   3 +-
 .../flink/graph/example/GraphMetrics.java       |   2 +-
 .../flink/graph/example/IncrementalSSSP.java    |   2 +-
 .../flink/graph/example/MusicProfiles.java      |   8 +-
 .../apache/flink/graph/gsa/ApplyFunction.java   |   6 +-
 .../apache/flink/graph/gsa/GatherFunction.java  |   6 +-
 .../org/apache/flink/graph/gsa/Neighbor.java    |   4 +-
 .../org/apache/flink/graph/gsa/SumFunction.java |   6 +-
 .../graph/library/ConnectedComponents.java      |   2 +-
 .../graph/library/GSAConnectedComponents.java   |   2 +-
 .../apache/flink/graph/library/GSAPageRank.java |   2 +-
 .../apache/flink/graph/library/PageRank.java    |   2 +-
 .../flink/graph/spargel/MessagingFunction.java  |   8 +-
 .../graph/spargel/VertexCentricIteration.java   |   2 +-
 .../graph/spargel/VertexUpdateFunction.java     |  16 ++-
 .../flink/optimizer/dag/OptimizerNode.java      |   2 +-
 .../operators/OperatorDescriptorSingle.java     |   2 +-
 .../apache/flink/optimizer/plan/Channel.java    |   1 -
 .../optimizer/postpass/SparseKeySchema.java     |   2 +-
 .../apache/flink/runtime/blob/BlobService.java  |   1 -
 .../runtime/checkpoint/PendingCheckpoint.java   |   1 -
 .../flink/runtime/execution/ExecutionState.java |   4 +-
 .../flink/runtime/filecache/FileCache.java      |   2 +-
 .../instance/SlotSharingGroupAssignment.java    |   4 +-
 .../io/disk/iomanager/BlockChannelReader.java   |   2 +-
 .../io/disk/iomanager/BlockChannelWriter.java   |   2 +-
 .../network/partition/consumer/InputGate.java   |   8 +-
 .../partition/consumer/SingleInputGate.java     |   8 +-
 .../task/IterationIntermediateTask.java         |   2 +-
 .../runtime/jobmanager/scheduler/Scheduler.java |   2 +-
 .../sort/AbstractMergeInnerJoinIterator.java    |   2 +-
 .../runtime/operators/util/BloomFilter.java     |   8 +-
 .../runtime/webmonitor/WebMonitorUtils.java     |   2 +-
 .../runtime/yarn/AbstractFlinkYarnCluster.java  |   4 +-
 .../runtime/zookeeper/StateStorageHelper.java   |   2 +-
 .../zookeeper/ZooKeeperStateHandleStore.java    |   8 +-
 .../mapred/HadoopMapFunction.java               |   2 +-
 .../mapred/HadoopReduceCombineFunction.java     |   2 +-
 .../mapred/HadoopReduceFunction.java            |   4 +-
 .../org/apache/flink/api/java/JarHelper.java    |   2 +-
 .../flink/api/java/table/package-info.java      |   8 +-
 .../connectors/kafka/FlinkKafkaConsumer.java    |   2 +-
 .../kafka/partitioner/FixedPartitioner.java     |  16 +--
 .../examples/iteration/IterateExample.java      |   7 +-
 .../ml/IncrementalLearningSkeleton.java         |   3 +-
 .../socket/SocketTextStreamWordCount.java       |  13 +-
 .../examples/twitter/TwitterStream.java         |  17 ++-
 .../examples/wordcount/PojoExample.java         |   4 +-
 .../streaming/examples/wordcount/WordCount.java |   4 +-
 .../streaming/api/datastream/DataStream.java    |   2 +-
 .../datastream/SingleOutputStreamOperator.java  |   9 +-
 .../environment/StreamExecutionEnvironment.java |  18 +--
 .../functions/source/RichSourceFunction.java    |   1 -
 .../api/functions/source/SourceFunction.java    |  12 +-
 .../api/graph/StreamGraphGenerator.java         |   2 +-
 .../transformations/StreamTransformation.java   |  14 +--
 .../streaming/api/windowing/time/Time.java      |   3 +-
 .../streaming/runtime/tasks/StreamTask.java     |   4 +-
 .../tasks/OneInputStreamTaskTestHarness.java    |   2 +-
 .../runtime/tasks/StreamTaskTestHarness.java    |   2 +-
 .../tasks/TwoInputStreamTaskTestHarness.java    |   2 +-
 .../test/util/MultipleProgramsTestBase.java     |   4 +-
 .../test/recordJobs/wordcount/WordCount.java    |   6 +
 151 files changed, 513 insertions(+), 570 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-clients/src/main/java/org/apache/flink/client/CliFrontend.java
----------------------------------------------------------------------
diff --git 
a/flink-clients/src/main/java/org/apache/flink/client/CliFrontend.java 
b/flink-clients/src/main/java/org/apache/flink/client/CliFrontend.java
index c638894..5485030 100644
--- a/flink-clients/src/main/java/org/apache/flink/client/CliFrontend.java
+++ b/flink-clients/src/main/java/org/apache/flink/client/CliFrontend.java
@@ -696,7 +696,8 @@ public class CliFrontend {
         * Creates a Packaged program from the given command line options.
         *
         * @return A PackagedProgram (upon success)
-        * @throws java.io.FileNotFoundException, 
org.apache.flink.client.program.ProgramInvocationException, java.lang.Throwable
+        * @throws java.io.FileNotFoundException
+        * @throws org.apache.flink.client.program.ProgramInvocationException
         */
        protected PackagedProgram buildProgram(ProgramOptions options)
                        throws FileNotFoundException, ProgramInvocationException

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationLocal.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationLocal.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationLocal.java
index 985cd68..56a0125 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationLocal.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationLocal.java
@@ -28,19 +28,16 @@ import 
org.apache.flink.storm.excamation.operators.ExclamationBolt;
  * Implements the "Exclamation" program that attaches five exclamation mark to 
every line of a text files in a streaming
  * fashion. The program is constructed as a regular {@link 
backtype.storm.generated.StormTopology} and submitted to
  * Flink for execution in the same way as to a Storm {@link 
backtype.storm.LocalCluster}.
- * <p/>
+ * <p>
  * This example shows how to run program directly within Java, thus it cannot 
be used to submit a
  * {@link backtype.storm.generated.StormTopology} via Flink command line 
clients (ie, bin/flink).
- * <p/>
- * <p/>
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
- * Usage: <code>ExclamationLocal &lt;text path&gt; &lt;result 
path&gt;</code><br/>
+ * <p>
+ * Usage: <code>ExclamationLocal &lt;text path&gt; &lt;result 
path&gt;</code><br>
  * If no parameters are provided, the program is run with default data from
  * {@link org.apache.flink.examples.java.wordcount.util.WordCountData}.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>run a regular Storm program locally on Flink</li>

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationTopology.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationTopology.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationTopology.java
index 70d25a2..9d94f5c 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationTopology.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationTopology.java
@@ -29,17 +29,14 @@ import org.apache.flink.storm.util.BoltPrintSink;
 
 /**
  * Implements the "Exclamation" program that attaches two exclamation marks to 
every line of a text files in a streaming
- * fashion. The program is constructed as a regular {@link StormTopology}.
- * <p/>
- * <p/>
+ * fashion. The program is constructed as a regular {@link 
backtype.storm.generated.StormTopology}.
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
+ * <p>
  * Usage: <code>Exclamation[Local|RemoteByClient|RemoteBySubmitter] &lt;text 
path&gt;
- * &lt;result path&gt;</code><br/>
+ * &lt;result path&gt;</code><br>
  * If no parameters are provided, the program is run with default data from 
{@link WordCountData}.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>construct a regular Storm topology as Flink program</li>

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationWithBolt.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationWithBolt.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationWithBolt.java
index 01ab907..19fe977 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationWithBolt.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationWithBolt.java
@@ -31,17 +31,14 @@ import backtype.storm.utils.Utils;
 
 /**
  * Implements the "Exclamation" program that attaches 3+x exclamation marks to 
every line of a text files in a streaming
- * fashion. The program is constructed as a regular {@link StormTopology}.
- * <p/>
- * <p/>
+ * fashion. The program is constructed as a regular {@link 
backtype.storm.generated.StormTopology}.
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
+ * <p>
  * Usage:
- * <code>ExclamationWithmBolt &lt;text path&gt; &lt;result path&gt; &lt;number 
of exclamation marks&gt;</code><br/>
+ * <code>ExclamationWithmBolt &lt;text path&gt; &lt;result path&gt; &lt;number 
of exclamation marks&gt;</code><br>
  * If no parameters are provided, the program is run with default data from 
{@link WordCountData} with x=2.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>use a Bolt within a Flink Streaming program</li>

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationWithSpout.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationWithSpout.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationWithSpout.java
index 22938e5..a196995 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationWithSpout.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/excamation/ExclamationWithSpout.java
@@ -32,16 +32,13 @@ import backtype.storm.utils.Utils;
 
 /**
  * Implements the "Exclamation" program that attaches six exclamation marks to 
every line of a text files in a streaming
- * fashion. The program is constructed as a regular {@link StormTopology}.
- * <p/>
- * <p/>
+ * fashion. The program is constructed as a regular {@link 
backtype.storm.generated.StormTopology}.
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
- * Usage: <code>ExclamationWithSpout &lt;text path&gt; &lt;result 
path&gt;</code><br/>
+ * <p>
+ * Usage: <code>ExclamationWithSpout &lt;text path&gt; &lt;result 
path&gt;</code><br>
  * If no parameters are provided, the program is run with default data from 
{@link WordCountData}.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>use a Storm spout within a Flink Streaming program</li>

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/split/SpoutSplitExample.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/split/SpoutSplitExample.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/split/SpoutSplitExample.java
index 560fe51..03c87f4 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/split/SpoutSplitExample.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/split/SpoutSplitExample.java
@@ -33,14 +33,14 @@ import 
org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
 
 /**
  * Implements a simple example with two declared output streams for the 
embedded spout.
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>handle multiple output stream of a spout</li>
  * <li>accessing each stream by .split(...) and .select(...)</li>
  * <li>strip wrapper data type SplitStreamType for further processing in 
Flink</li>
  * </ul>
- * <p/>
+ * <p>
  * This example would work the same way for multiple bolt output streams.
  */
 public class SpoutSplitExample {

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCount.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCount.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCount.java
index aa3a075..6ee7fd9 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCount.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCount.java
@@ -30,15 +30,12 @@ import 
org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
 /**
  * Implements the "WordCount" program that computes a simple word occurrence 
histogram over text files in a streaming
  * fashion. The tokenizer step is performed by a {@link IRichBolt Bolt}.
- * <p/>
- * <p/>
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
+ * <p>
  * Usage: <code>WordCount &lt;text path&gt; &lt;result path&gt;</code><br>
  * If no parameters are provided, the program is run with default data from 
{@link WordCountData}.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>use a Bolt within a Flink Streaming program.</li>

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCountPojo.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCountPojo.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCountPojo.java
index f72acb3..9bdcead 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCountPojo.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCountPojo.java
@@ -36,15 +36,12 @@ import 
org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
  * Implements the "WordCount" program that computes a simple word occurrence 
histogram over text files in a streaming
  * fashion. The tokenizer step is performed by a {@link IRichBolt Bolt}. In 
contrast to {@link BoltTokenizerWordCount}
  * the tokenizer's input is a POJO type and the single field is accessed by 
name.
- * <p/>
- * <p/>
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
+ * <p>
  * Usage: <code>WordCount &lt;text path&gt; &lt;result path&gt;</code><br>
  * If no parameters are provided, the program is run with default data from 
{@link WordCountData}.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>how to access attributes by name within a Bolt for POJO type input 
streams

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCountWithNames.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCountWithNames.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCountWithNames.java
index 7617e95..019f1bc 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCountWithNames.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/BoltTokenizerWordCountWithNames.java
@@ -38,15 +38,12 @@ import 
org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
  * Implements the "WordCount" program that computes a simple word occurrence 
histogram over text files in a streaming
  * fashion. The tokenizer step is performed by a {@link IRichBolt Bolt}. In 
contrast to {@link BoltTokenizerWordCount}
  * the tokenizer's input is a {@link Tuple} type and the single field is 
accessed by name.
- * <p/>
- * <p/>
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
+ * <p>
  * Usage: <code>WordCount &lt;text path&gt; &lt;result path&gt;</code><br>
  * If no parameters are provided, the program is run with default data from 
{@link WordCountData}.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>how to access attributes by name within a Bolt for {@link Tuple} type 
input streams

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/SpoutSourceWordCount.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/SpoutSourceWordCount.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/SpoutSourceWordCount.java
index bb451fe..281780e 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/SpoutSourceWordCount.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/SpoutSourceWordCount.java
@@ -34,15 +34,12 @@ import org.apache.flink.util.Collector;
 /**
  * Implements the "WordCount" program that computes a simple word occurrence 
histogram over text files in a streaming
  * fashion. The used data source is a {@link IRichSpout Spout}.
- * <p/>
- * <p/>
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
+ * <p>
  * Usage: <code>WordCount &lt;text path&gt; &lt;result path&gt;</code><br>
  * If no parameters are provided, the program is run with default data from 
{@link WordCountData}.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>use a Spout within a Flink Streaming program.</li>
@@ -89,7 +86,7 @@ public class SpoutSourceWordCount {
 
        /**
         * Implements the string tokenizer that splits sentences into words as 
a user-defined FlatMapFunction. The function
-        * takes a line (String) and splits it into multiple pairs in the form 
of "(word,1)" (Tuple2<String, Integer>).
+        * takes a line (String) and splits it into multiple pairs in the form 
of "(word,1)" ({@code Tuple2<String, Integer>}).
         */
        public static final class Tokenizer implements FlatMapFunction<String, 
Tuple2<String, Integer>> {
                private static final long serialVersionUID = 1L;

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountLocal.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountLocal.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountLocal.java
index 18f49c1..0a7dfa0 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountLocal.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountLocal.java
@@ -29,18 +29,15 @@ import org.apache.flink.storm.api.FlinkTopologyBuilder;
  * Implements the "WordCount" program that computes a simple word occurrence 
histogram over text files in a streaming
  * fashion. The program is constructed as a regular {@link StormTopology} and 
submitted to Flink for execution in the
  * same way as to a Storm {@link LocalCluster}.
- * <p/>
+ * <p>
  * This example shows how to run program directly within Java, thus it cannot 
be used to submit a {@link StormTopology}
  * via Flink command line clients (ie, bin/flink).
- * <p/>
- * <p/>
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
+ * <p>
  * Usage: <code>WordCountLocal &lt;text path&gt; &lt;result path&gt;</code><br>
  * If no parameters are provided, the program is run with default data from 
{@link WordCountData}.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>run a regular Storm program locally on Flink</li>

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountLocalByName.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountLocalByName.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountLocalByName.java
index 71a5e8d..67f4bbe 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountLocalByName.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountLocalByName.java
@@ -30,18 +30,15 @@ import org.apache.flink.storm.api.FlinkTopologyBuilder;
  * fashion. The program is constructed as a regular {@link StormTopology} and 
submitted to Flink for execution in the
  * same way as to a Storm {@link LocalCluster}. In contrast to {@link 
WordCountLocal} all bolts access the field of
  * input tuples by name instead of index.
- * <p/>
+ * <p>
  * This example shows how to run program directly within Java, thus it cannot 
be used to submit a {@link StormTopology}
  * via Flink command line clients (ie, bin/flink).
- * <p/>
- * <p/>
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
+ * <p>
  * Usage: <code>WordCountLocalByName &lt;text path&gt; &lt;result 
path&gt;</code><br>
  * If no parameters are provided, the program is run with default data from 
{@link WordCountData}.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>run a regular Storm program locally on Flink

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountRemoteByClient.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountRemoteByClient.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountRemoteByClient.java
index 2e4fb03..8cb7cdd 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountRemoteByClient.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountRemoteByClient.java
@@ -33,18 +33,15 @@ import org.apache.flink.storm.api.FlinkTopologyBuilder;
  * Implements the "WordCount" program that computes a simple word occurrence 
histogram over text files in a streaming
  * fashion. The program is constructed as a regular {@link StormTopology} and 
submitted to Flink for execution in the
  * same way as to a Storm cluster similar to {@link NimbusClient}. The Flink 
cluster can be local or remote.
- * <p/>
+ * <p>
  * This example shows how to submit the program via Java, thus it cannot be 
used to submit a {@link StormTopology} via
  * Flink command line clients (ie, bin/flink).
- * <p/>
- * <p/>
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
+ * <p>
  * Usage: <code>WordCountRemoteByClient &lt;text path&gt; &lt;result 
path&gt;</code><br>
  * If no parameters are provided, the program is run with default data from 
{@link WordCountData}.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>submit a regular Storm program to a local or remote Flink cluster.</li>

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountRemoteBySubmitter.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountRemoteBySubmitter.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountRemoteBySubmitter.java
index 173074c..a20843e 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountRemoteBySubmitter.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountRemoteBySubmitter.java
@@ -30,17 +30,14 @@ import org.apache.flink.storm.api.FlinkTopologyBuilder;
  * Implements the "WordCount" program that computes a simple word occurrence 
histogram over text files in a streaming
  * fashion. The program is constructed as a regular {@link StormTopology} and 
submitted to Flink for execution in the
  * same way as to a Storm cluster similar to {@link StormSubmitter}. The Flink 
cluster can be local or remote.
- * <p/>
+ * <p>
  * This example shows how to submit the program via Java as well as Flink's 
command line client (ie, bin/flink).
- * <p/>
- * <p/>
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
+ * <p>
  * Usage: <code>WordCountRemoteBySubmitter &lt;text path&gt; &lt;result 
path&gt;</code><br>
  * If no parameters are provided, the program is run with default data from 
{@link WordCountData}.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>submit a regular Storm program to a local or remote Flink cluster.</li>

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountTopology.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountTopology.java
 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountTopology.java
index 8ee374d..138df65 100644
--- 
a/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountTopology.java
+++ 
b/flink-contrib/flink-storm-examples/src/main/java/org/apache/flink/storm/wordcount/WordCountTopology.java
@@ -36,16 +36,13 @@ import 
org.apache.flink.storm.wordcount.operators.WordCountInMemorySpout;
 /**
  * Implements the "WordCount" program that computes a simple word occurrence 
histogram over text files in a streaming
  * fashion. The program is constructed as a regular {@link StormTopology}.
- * <p/>
- * <p/>
+ * <p>
  * The input is a plain text file with lines separated by newline characters.
- * <p/>
- * <p/>
+ * <p>
  * Usage:
  * <code>WordCount[Local|LocalByName|RemoteByClient|RemoteBySubmitter] 
&lt;text path&gt; &lt;result path&gt;</code><br>
  * If no parameters are provided, the program is run with default data from 
{@link WordCountData}.
- * <p/>
- * <p/>
+ * <p>
  * This example shows how to:
  * <ul>
  * <li>how to construct a regular Storm topology as Flink program</li>

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/api/FlinkClient.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/api/FlinkClient.java
 
b/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/api/FlinkClient.java
index 2be7599..3607fad 100644
--- 
a/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/api/FlinkClient.java
+++ 
b/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/api/FlinkClient.java
@@ -138,7 +138,7 @@ public class FlinkClient {
 
        /**
         * Return a reference to itself.
-        * <p/>
+        * <p>
         * {@link FlinkClient} mimics both, {@link NimbusClient} and {@link 
Nimbus}{@code .Client}, at once.
         *
         * @return A reference to itself.

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/api/FlinkTopologyBuilder.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/api/FlinkTopologyBuilder.java
 
b/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/api/FlinkTopologyBuilder.java
index 8a88eac..47aa68e 100644
--- 
a/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/api/FlinkTopologyBuilder.java
+++ 
b/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/api/FlinkTopologyBuilder.java
@@ -55,8 +55,8 @@ import java.util.Set;
 /**
  * {@link FlinkTopologyBuilder} mimics a {@link TopologyBuilder}, but builds a 
Flink program instead of a Storm
  * topology. Most methods (except {@link #createTopology()} are copied from 
the original {@link TopologyBuilder}
- * implementation to ensure equal behavior.<br />
- * <br />
+ * implementation to ensure equal behavior.<br>
+ * <br>
  * <strong>CAUTION: {@link IRichStateSpout StateSpout}s are currently not 
supported.</strong>
  */
 public class FlinkTopologyBuilder {

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/util/SplitStreamType.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/util/SplitStreamType.java
 
b/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/util/SplitStreamType.java
index a4b5f8e..5056795 100644
--- 
a/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/util/SplitStreamType.java
+++ 
b/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/util/SplitStreamType.java
@@ -20,10 +20,11 @@ package org.apache.flink.storm.util;
 import org.apache.flink.streaming.api.datastream.DataStream;
 
 /**
- * Used by {@link org.apache.flink.storm.wrappers.AbstractStormCollector 
AbstractStormCollector} to wrap
+ * Used by org.apache.flink.storm.wrappers.AbstractStormCollector to wrap
  * output tuples if multiple output streams are declared. For this case, the 
Flink output data stream must be split via
  * {@link 
DataStream#split(org.apache.flink.streaming.api.collector.selector.OutputSelector)
 .split(...)} using
  * {@link StormStreamSelector}.
+ *
  */
 public class SplitStreamType<T> {
 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/wrappers/BoltWrapper.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/wrappers/BoltWrapper.java
 
b/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/wrappers/BoltWrapper.java
index 12d967a..d9a2f91 100644
--- 
a/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/wrappers/BoltWrapper.java
+++ 
b/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/wrappers/BoltWrapper.java
@@ -42,8 +42,8 @@ import com.google.common.collect.Sets;
  * A {@link BoltWrapper} wraps an {@link IRichBolt} in order to execute the 
Storm bolt within a Flink Streaming
  * program. It takes the Flink input tuples of type {@code IN} and transforms 
them into {@link StormTuple}s that the
  * bolt can process. Furthermore, it takes the bolt's output tuples and 
transforms them into Flink tuples of type
- * {@code OUT} (see {@link AbstractStormCollector} for supported types).<br />
- * <br />
+ * {@code OUT} (see {@link AbstractStormCollector} for supported types).<br>
+ * <br>
  * <strong>CAUTION: currently, only simple bolts are supported! (ie, bolts 
that do not use the Storm configuration
  * <code>Map</code> or <code>TopologyContext</code> that is provided by the 
bolt's <code>open(..)</code> method.
  * Furthermore, acking and failing of tuples as well as accessing tuple 
attributes by field names is not supported so

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/wrappers/SpoutWrapper.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/wrappers/SpoutWrapper.java
 
b/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/wrappers/SpoutWrapper.java
index e78dd5c..62b36be 100644
--- 
a/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/wrappers/SpoutWrapper.java
+++ 
b/flink-contrib/flink-storm/src/main/java/org/apache/flink/storm/wrappers/SpoutWrapper.java
@@ -38,15 +38,15 @@ import com.google.common.collect.Sets;
 /**
  * A {@link SpoutWrapper} wraps an {@link IRichSpout} in order to execute it 
within a Flink Streaming program. It
  * takes the spout's output tuples and transforms them into Flink tuples of 
type {@code OUT} (see
- * {@link SpoutCollector} for supported types).<br />
- * <br />
+ * {@link SpoutCollector} for supported types).<br>
+ * <br>
  * Per default, {@link SpoutWrapper} calls the wrapped spout's {@link 
IRichSpout#nextTuple() nextTuple()} method in
- * an infinite loop.<br />
+ * an infinite loop.<br>
  * Alternatively, {@link SpoutWrapper} can call {@link IRichSpout#nextTuple() 
nextTuple()} for a finite number of
  * times and terminate automatically afterwards (for finite input streams). 
The number of {@code nextTuple()} calls can
  * be specified as a certain number of invocations or can be undefined. In the 
undefined case, {@link SpoutWrapper}
  * terminates if no record was emitted to the output collector for the first 
time during a call to
- * {@link IRichSpout#nextTuple() nextTuple()}.<br />
+ * {@link IRichSpout#nextTuple() nextTuple()}.<br>
  * If the given spout implements {@link FiniteSpout} interface and {@link 
#numberOfInvocations} is not provided or
  * is {@code null}, {@link SpoutWrapper} calls {@link IRichSpout#nextTuple() 
nextTuple()} method until
  * {@link FiniteSpout#reachedEnd()} returns true.
@@ -258,7 +258,7 @@ public final class SpoutWrapper<OUT> extends 
RichParallelSourceFunction<OUT> {
 
        /**
         * {@inheritDoc}
-        * <p/>
+        * <p>
         * Sets the {@link #isRunning} flag to {@code false}.
         */
        @Override

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Entities.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Entities.java
 
b/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Entities.java
index d88ea34..8a8f91e 100755
--- 
a/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Entities.java
+++ 
b/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Entities.java
@@ -22,7 +22,7 @@ import java.util.List;
 
 /**
  * Entities which have been parsed out of the text of the
- * {@link package org.apache.flink.contrib.tweetinputformat.model.tweet.Tweet}.
+ * {@link org.apache.flink.contrib.tweetinputformat.model.tweet.Tweet}.
  */
 public class Entities {
 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/HashTags.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/HashTags.java
 
b/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/HashTags.java
index 1900859..ba6f5c0 100755
--- 
a/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/HashTags.java
+++ 
b/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/HashTags.java
@@ -19,7 +19,7 @@ package 
org.apache.flink.contrib.tweetinputformat.model.tweet.entities;
 
 /**
  * Represents hashtags which have been parsed out of the
- * {@link package org.apache.flink.contrib.tweetinputformat.model.tweet.Tweet} 
text.
+ * {@link org.apache.flink.contrib.tweetinputformat.model.tweet.Tweet} text.
  */
 
 public class HashTags {

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Media.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Media.java
 
b/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Media.java
index f006aac..37d2ab8 100755
--- 
a/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Media.java
+++ 
b/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Media.java
@@ -21,7 +21,7 @@ import java.util.HashMap;
 import java.util.Map;
 
 /**
- * Represents media elements uploaded with the {@link package 
org.apache.flink.contrib.tweetinputformat.model.tweet.Tweet}.
+ * Represents media elements uploaded with the {@link 
org.apache.flink.contrib.tweetinputformat.model.tweet.Tweet}.
  */
 public class Media {
 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Symbol.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Symbol.java
 
b/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Symbol.java
index db562e8..c2a6dd0 100755
--- 
a/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Symbol.java
+++ 
b/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/Symbol.java
@@ -19,7 +19,7 @@ package 
org.apache.flink.contrib.tweetinputformat.model.tweet.entities;
 
 /**
  * An array of financial symbols starting with the dollar sign extracted from 
the
- * {@link package org.apache.flink.contrib.tweetinputformat.model.tweet.Tweet} 
text.
+ * {@link org.apache.flink.contrib.tweetinputformat.model.tweet.Tweet} text.
  */
 
 public class Symbol {

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/URL.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/URL.java
 
b/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/URL.java
index 6d0f184..93ddebd 100755
--- 
a/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/URL.java
+++ 
b/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/URL.java
@@ -19,7 +19,7 @@ package 
org.apache.flink.contrib.tweetinputformat.model.tweet.entities;
 
 /**
  * Represents URLs included in the text of a Tweet or within textual fields of 
a
- * {@link package 
org.apache.flink.contrib.tweetinputformat.model.tweet.User.Users} object.
+ * {@link 
org.apache.flink.contrib.tweetinputformat.model.tweet.entities.UserMention} 
object.
  */
 public class URL {
 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/UserMention.java
----------------------------------------------------------------------
diff --git 
a/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/UserMention.java
 
b/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/UserMention.java
index a56f7c7..2a3cf1a 100755
--- 
a/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/UserMention.java
+++ 
b/flink-contrib/flink-tweet-inputformat/src/main/java/org/apache/flink/contrib/tweetinputformat/model/tweet/entities/UserMention.java
@@ -19,7 +19,7 @@ package 
org.apache.flink.contrib.tweetinputformat.model.tweet.entities;
 
 /**
  * Represents other Twitter users mentioned in the text of the
- * {@link package org.apache.flink.contrib.tweetinputformat.model.tweet.Tweet}.
+ * {@link org.apache.flink.contrib.tweetinputformat.model.tweet.Tweet}.
  */
 public class UserMention {
 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java 
b/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java
index f38ecb0..b9ebaf7 100644
--- a/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java
+++ b/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java
@@ -677,7 +677,7 @@ public class ExecutionConfig implements Serializable {
                private static final long serialVersionUID = 1L;
 
                /**
-                * Convert UserConfig into a Map<String, String> representation.
+                * Convert UserConfig into a {@code Map<String, String>} 
representation.
                 * This can be used by the runtime, for example for presenting 
the user config in the web frontend.
                 *
                 * @return Key/Value representation of the UserConfig, or null.

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/ExecutionMode.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/ExecutionMode.java 
b/flink-core/src/main/java/org/apache/flink/api/common/ExecutionMode.java
index f3e958e..b1463cc 100644
--- a/flink-core/src/main/java/org/apache/flink/api/common/ExecutionMode.java
+++ b/flink-core/src/main/java/org/apache/flink/api/common/ExecutionMode.java
@@ -33,10 +33,10 @@ public enum ExecutionMode {
         * pipelined manner) are data flows that branch (one data set consumed 
by multiple
         * operations) and re-join later:
         * <pre>{@code
-        *    DataSet data = ...;
-        *    DataSet mapped1 = data.map(new MyMapper());
-        *    DataSet mapped2 = data.map(new AnotherMapper());
-        *    mapped1.join(mapped2).where(...).equalTo(...);
+        * DataSet data = ...;
+        * DataSet mapped1 = data.map(new MyMapper());
+        * DataSet mapped2 = data.map(new AnotherMapper());
+        * mapped1.join(mapped2).where(...).equalTo(...);
         * }</pre>
         */
        PIPELINED,

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/accumulators/Histogram.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/accumulators/Histogram.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/accumulators/Histogram.java
index f5e959a..f587fee 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/accumulators/Histogram.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/accumulators/Histogram.java
@@ -23,7 +23,7 @@ import java.util.TreeMap;
 
 /**
  * Histogram accumulator, which builds a histogram in a distributed manner.
- * Implemented as a Integer->Integer TreeMap, so that the entries are sorted
+ * Implemented as a Integer-&gt;Integer TreeMap, so that the entries are sorted
  * according to the values.
  * 
  * This class does not extend to continuous values later, because it makes no

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/aggregators/Aggregator.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/aggregators/Aggregator.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/aggregators/Aggregator.java
index 035b591..aabb82f 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/aggregators/Aggregator.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/aggregators/Aggregator.java
@@ -47,7 +47,7 @@ import org.apache.flink.types.Value;
  *     }
  *     
  *     public boolean filter (Double value) {
- *         if (value > 1000000.0) {
+ *         if (value &gt; 1000000.0) {
  *             agg.aggregate(1);
  *             return false
  *         }

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/distributions/DataDistribution.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/distributions/DataDistribution.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/distributions/DataDistribution.java
index 9bf1abf..ebf1319 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/distributions/DataDistribution.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/distributions/DataDistribution.java
@@ -39,7 +39,7 @@ public interface DataDistribution extends IOReadableWritable, 
Serializable {
         * <p>
         * Note: The last bucket's upper bound is actually discarded by many 
algorithms.
         * The last bucket is assumed to hold all values <i>v</i> such that
-        * {@code v &gt; getBucketBoundary(n-1, n)}, where <i>n</i> is the 
number of buckets.
+        * {@code v > getBucketBoundary(n-1, n)}, where <i>n</i> is the number 
of buckets.
         * 
         * @param bucketNum The number of the bucket for which to get the upper 
bound.
         * @param totalNumBuckets The number of buckets to split the data into.

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/functions/CoGroupFunction.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/CoGroupFunction.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/CoGroupFunction.java
index 2d7a1d7..1098341 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/CoGroupFunction.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/CoGroupFunction.java
@@ -28,12 +28,12 @@ import org.apache.flink.util.Collector;
  * If a key is present in only one of the two inputs, it may be that one of 
the groups is empty.
  * <p>
  * The basic syntax for using CoGoup on two data sets is as follows:
- * <pre><blockquote>
+ * <pre>{@code
  * DataSet<X> set1 = ...;
  * DataSet<Y> set2 = ...;
  * 
  * 
set1.coGroup(set2).where(<key-definition>).equalTo(<key-definition>).with(new 
MyCoGroupFunction());
- * </blockquote></pre>
+ * }</pre>
  * <p>
  * {@code set1} is here considered the first input, {@code set2} the second 
input.
  * <p>

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/functions/CrossFunction.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/CrossFunction.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/CrossFunction.java
index 0e9d2a5..3660e63 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/CrossFunction.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/CrossFunction.java
@@ -28,12 +28,12 @@ import java.io.Serializable;
  * pair of elements, instead of processing 2-tuples that contain the pairs.
  * <p>
  * The basic syntax for using Cross on two data sets is as follows:
- * <pre><blockquote>
+ * <pre>{@code
  * DataSet<X> set1 = ...;
  * DataSet<Y> set2 = ...;
  * 
  * set1.cross(set2).with(new MyCrossFunction());
- * </blockquote></pre>
+ * }</pre>
  * <p>
  * {@code set1} is here considered the first input, {@code set2} the second 
input.
  * 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/functions/FilterFunction.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/FilterFunction.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/FilterFunction.java
index 1f21a96..1a80cb8 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/FilterFunction.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/FilterFunction.java
@@ -25,11 +25,11 @@ import java.io.Serializable;
  * The predicate decides whether to keep the element, or to discard it.
  * <p>
  * The basic syntax for using a FilterFunction is as follows:
- * <pre><blockquote>
+ * <pre>{@code
  * DataSet<X> input = ...;
  * 
  * DataSet<X> result = input.filter(new MyFilterFunction());
- * </blockquote></pre>
+ * }</pre>
  * <p>
  * <strong>IMPORTANT:</strong> The system assumes that the function does not
  * modify the elements on which the predicate is applied. Violating this 
assumption

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/functions/FlatJoinFunction.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/FlatJoinFunction.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/FlatJoinFunction.java
index f99aec4..c79046b 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/FlatJoinFunction.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/FlatJoinFunction.java
@@ -34,12 +34,12 @@ import java.io.Serializable;
  * if their key is not contained in the other data set.
  * <p>
  * The basic syntax for using Join on two data sets is as follows:
- * <pre><blockquote>
+ * <pre>{@code
  * DataSet<X> set1 = ...;
  * DataSet<Y> set2 = ...;
  * 
  * set1.join(set2).where(<key-definition>).equalTo(<key-definition>).with(new 
MyJoinFunction());
- * </blockquote></pre>
+ * }</pre>
  * <p>
  * {@code set1} is here considered the first input, {@code set2} the second 
input.
  * <p>

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/functions/FlatMapFunction.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/FlatMapFunction.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/FlatMapFunction.java
index 37c6e83..5b955f7 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/FlatMapFunction.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/FlatMapFunction.java
@@ -29,11 +29,11 @@ import java.io.Serializable;
  * use the {@link MapFunction}.
  * <p>
  * The basic syntax for using a FlatMapFunction is as follows:
- * <pre><blockquote>
+ * <pre>{@code
  * DataSet<X> input = ...;
  * 
  * DataSet<Y> result = input.flatMap(new MyFlatMapFunction());
- * </blockquote></pre>
+ * }</pre>
  * 
  * @param <T> Type of the input elements.
  * @param <O> Type of the returned elements.

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/functions/FoldFunction.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/FoldFunction.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/FoldFunction.java
index a9c5b2b..c4be00b 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/FoldFunction.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/FoldFunction.java
@@ -25,12 +25,12 @@ import java.io.Serializable;
  * a single value, by applying a binary operation to an initial accumulator 
element every element from a group elements.
  * <p>
  * The basic syntax for using a FoldFunction is as follows:
- * <pre><blockquote>
+ * <pre>{@code
  * DataSet<X> input = ...;
  *
  * X initialValue = ...;
  * DataSet<X> result = input.fold(new MyFoldFunction(), initialValue);
- * </blockquote></pre>
+ * }</pre>
  * <p>
  * Like all functions, the FoldFunction needs to be serializable, as defined 
in {@link java.io.Serializable}.
  *

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/functions/GroupReduceFunction.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/GroupReduceFunction.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/GroupReduceFunction.java
index befbee3..36500f6 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/GroupReduceFunction.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/GroupReduceFunction.java
@@ -32,11 +32,11 @@ import org.apache.flink.util.Collector;
  * {@link ReduceFunction}.
  * <p>
  * The basic syntax for using a grouped GroupReduceFunction is as follows:
- * <pre><blockquote>
+ * <pre>{@code
  * DataSet<X> input = ...;
  * 
  * DataSet<X> result = input.groupBy(<key-definition>).reduceGroup(new 
MyGroupReduceFunction());
- * </blockquote></pre>
+ * }</pre>
  * 
  * @param <T> Type of the elements that this function processes.
  * @param <O> The type of the elements returned by the user-defined function.

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/functions/JoinFunction.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/JoinFunction.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/JoinFunction.java
index 085bae9..0032a0f 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/JoinFunction.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/JoinFunction.java
@@ -29,12 +29,12 @@ import java.io.Serializable;
  * if their key is not contained in the other data set.
  * <p>
  * The basic syntax for using Join on two data sets is as follows:
- * <pre><blockquote>
+ * <pre>{@code
  * DataSet<X> set1 = ...;
  * DataSet<Y> set2 = ...;
  * 
  * set1.join(set2).where(<key-definition>).equalTo(<key-definition>).with(new 
MyJoinFunction());
- * </blockquote></pre>
+ * }</pre>
  * <p>
  * {@code set1} is here considered the first input, {@code set2} the second 
input.
  * <p>

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/functions/MapFunction.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/MapFunction.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/MapFunction.java
index bbeacc6..45f3d1f 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/MapFunction.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/MapFunction.java
@@ -28,11 +28,11 @@ import java.io.Serializable;
  * using the {@link FlatMapFunction}.
  * <p>
  * The basic syntax for using a MapFunction is as follows:
- * <pre><blockquote>
+ * <pre>{@code
  * DataSet<X> input = ...;
  * 
  * DataSet<Y> result = input.map(new MyMapFunction());
- * </blockquote></pre>
+ * }</pre>
  * 
  * @param <T> Type of the input elements.
  * @param <O> Type of the returned elements.

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/functions/MapPartitionFunction.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/MapPartitionFunction.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/MapPartitionFunction.java
index 9011d25..d2c0077 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/MapPartitionFunction.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/MapPartitionFunction.java
@@ -31,11 +31,11 @@ import java.io.Serializable;
  * For most of the simple use cases, consider using the {@link MapFunction} or 
{@link FlatMapFunction}.
  * <p>
  * The basic syntax for a MapPartitionFunction is as follows:
- * <pre><blockquote>
+ * <pre>{@code
  * DataSet<X> input = ...;
  * 
  * DataSet<Y> result = input.mapPartition(new MyMapPartitionFunction());
- * </blockquote></pre>
+ * }</pre>
  * 
  * @param <T> Type of the input elements.
  * @param <O> Type of the returned elements.

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/functions/ReduceFunction.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/ReduceFunction.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/ReduceFunction.java
index 8de7861..1b34047 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/ReduceFunction.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/ReduceFunction.java
@@ -32,11 +32,11 @@ import java.io.Serializable;
  * execution strategies.
  * <p>
  * The basic syntax for using a grouped ReduceFunction is as follows:
- * <pre><blockquote>
+ * <pre>{@code
  * DataSet<X> input = ...;
  * 
  * DataSet<X> result = input.groupBy(<key-definition>).reduce(new 
MyReduceFunction());
- * </blockquote></pre>
+ * }</pre>
  * <p>
  * Like all functions, the ReduceFunction needs to be serializable, as defined 
in {@link java.io.Serializable}.
  * 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/functions/RichGroupReduceFunction.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/RichGroupReduceFunction.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/RichGroupReduceFunction.java
index 48e27d3..b5edc64 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/functions/RichGroupReduceFunction.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/functions/RichGroupReduceFunction.java
@@ -49,7 +49,7 @@ public abstract class RichGroupReduceFunction<IN, OUT> 
extends AbstractRichFunct
         * <p>
         * This method is only ever invoked when the subclass of {@link 
RichGroupReduceFunction}
         * adds the {@link Combinable} annotation, or if the <i>combinable</i> 
flag is set when defining
-        * the <i>reduceGroup<i> operation via
+        * the <i>reduceGroup</i> operation via
         * 
org.apache.flink.api.java.operators.GroupReduceOperator#setCombinable(boolean).
         * <p>
         * Since the reduce function will be called on the result of this 
method, it is important that this

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/io/FileOutputFormat.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/io/FileOutputFormat.java 
b/flink-core/src/main/java/org/apache/flink/api/common/io/FileOutputFormat.java
index 6854268..219877d 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/io/FileOutputFormat.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/io/FileOutputFormat.java
@@ -51,7 +51,7 @@ public abstract class FileOutputFormat<IT> extends 
RichOutputFormat<IT> implemen
                /** A directory is always created, regardless of number of 
write tasks. */
                ALWAYS, 
                
-               /** A directory is only created for parallel output tasks, 
i.e., number of output tasks > 1.
+               /** A directory is only created for parallel output tasks, 
i.e., number of output tasks &gt; 1.
                 * If number of output tasks = 1, the output is written to a 
single file. */
                PARONLY
        }

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/operators/AbstractUdfOperator.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/operators/AbstractUdfOperator.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/operators/AbstractUdfOperator.java
index 1c79a37..74b0d01 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/operators/AbstractUdfOperator.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/operators/AbstractUdfOperator.java
@@ -105,7 +105,7 @@ public abstract class AbstractUdfOperator<OUT, FT extends 
Function> extends Oper
         * Clears all previous broadcast inputs and binds the given inputs as
         * broadcast variables of this operator.
         * 
-        * @param inputs The <name, root> pairs to be set as broadcast inputs.
+        * @param inputs The {@code<name, root>} pairs to be set as broadcast 
inputs.
         */
        public <T> void setBroadcastVariables(Map<String, Operator<T>> inputs) {
                this.broadcastInputs.clear();

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/api/common/typeutils/TypeComparator.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/typeutils/TypeComparator.java
 
b/flink-core/src/main/java/org/apache/flink/api/common/typeutils/TypeComparator.java
index 3a545e4..cd5da84 100644
--- 
a/flink-core/src/main/java/org/apache/flink/api/common/typeutils/TypeComparator.java
+++ 
b/flink-core/src/main/java/org/apache/flink/api/common/typeutils/TypeComparator.java
@@ -76,7 +76,7 @@ public abstract class TypeComparator<T> implements 
Serializable {
         * of the fields from the record, this method may extract those fields.
         * <p>
         * A typical example for checking the equality of two elements is the 
following:
-        * <pre>
+        * <pre>{@code
         * E e1 = ...;
         * E e2 = ...;
         * 
@@ -84,7 +84,7 @@ public abstract class TypeComparator<T> implements 
Serializable {
         * 
         * acc.setReference(e1);
         * boolean equal = acc.equalToReference(e2);
-        * </pre>
+        * }</pre>
         * 
         * The rational behind this method is that elements are typically 
compared using certain features that
         * are extracted from them, (such de-serializing as a subset of 
fields). When setting the
@@ -113,7 +113,7 @@ public abstract class TypeComparator<T> implements 
Serializable {
         * elements {@code e1} and {@code e2} via a comparator, this method can 
be used the
         * following way.
         * 
-        * <pre>
+        * <pre>{@code
         * E e1 = ...;
         * E e2 = ...;
         * 
@@ -124,7 +124,7 @@ public abstract class TypeComparator<T> implements 
Serializable {
         * acc2.setReference(e2);
         * 
         * int comp = acc1.compareToReference(acc2);
-        * </pre>
+        * }</pre>
         * 
         * The rational behind this method is that elements are typically 
compared using certain features that
         * are extracted from them, (such de-serializing as a subset of 
fields). When setting the

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java 
b/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java
index b64939e..d331548 100644
--- 
a/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java
+++ 
b/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java
@@ -600,7 +600,7 @@ public final class ConfigConstants {
        public static final boolean DEFAULT_FILESYSTEM_OVERWRITE = false;
 
        /**
-        * The default behavior for output directory creating (create only 
directory when parallelism > 1).
+        * The default behavior for output directory creating (create only 
directory when parallelism &gt; 1).
         */
        public static final boolean DEFAULT_FILESYSTEM_ALWAYS_CREATE_DIRECTORY 
= false;
        

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java
----------------------------------------------------------------------
diff --git a/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java 
b/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java
index 185b5f2..7dc92c6 100644
--- a/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java
+++ b/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java
@@ -162,8 +162,6 @@ public abstract class FileSystem {
         * 
         * @return a reference to the {@link FileSystem} instance for accessing 
the
         *         local file system.
-        * @throws IOException
-        *         thrown if a reference to the file system instance could not 
be obtained
         */
        public static FileSystem getLocalFileSystem() {
                // this should really never fail.
@@ -485,20 +483,20 @@ public abstract class FileSystem {
        /**
         * Initializes output directories on local file systems according to 
the given write mode.
         * 
-        * WriteMode.CREATE & parallel output:
+        * WriteMode.CREATE &amp; parallel output:
         *  - A directory is created if the output path does not exist.
         *  - An existing directory is reused, files contained in the directory 
are NOT deleted.
         *  - An existing file raises an exception.
         *    
-        * WriteMode.CREATE & NONE parallel output:
+        * WriteMode.CREATE &amp; NONE parallel output:
         *  - An existing file or directory raises an exception.
         *  
-        * WriteMode.OVERWRITE & parallel output:
+        * WriteMode.OVERWRITE &amp; parallel output:
         *  - A directory is created if the output path does not exist.
         *  - An existing directory is reused, files contained in the directory 
are NOT deleted.
         *  - An existing file is deleted and replaced by a new directory.
         *  
-        * WriteMode.OVERWRITE & NONE parallel output:
+        * WriteMode.OVERWRITE &amp; NONE parallel output:
         *  - An existing file or directory (and all its content) is deleted
         * 
         * Files contained in an existing directory are not deleted, because 
multiple instances of a 
@@ -646,19 +644,19 @@ public abstract class FileSystem {
        /**
         * Initializes output directories on distributed file systems according 
to the given write mode.
         * 
-        * WriteMode.CREATE & parallel output:
+        * WriteMode.CREATE &amp; parallel output:
         *  - A directory is created if the output path does not exist.
         *  - An existing file or directory raises an exception.
         * 
-        * WriteMode.CREATE & NONE parallel output:
+        * WriteMode.CREATE &amp; NONE parallel output:
         *  - An existing file or directory raises an exception. 
         *    
-        * WriteMode.OVERWRITE & parallel output:
+        * WriteMode.OVERWRITE &amp; parallel output:
         *  - A directory is created if the output path does not exist.
         *  - An existing directory and its content is deleted and a new 
directory is created.
         *  - An existing file is deleted and replaced by a new directory.
         *  
-        *  WriteMode.OVERWRITE & NONE parallel output:
+        *  WriteMode.OVERWRITE &amp; NONE parallel output:
         *  - An existing file or directory is deleted and replaced by a new 
directory.
         * 
         * @param outPath Output path that should be prepared.

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/core/memory/MemorySegment.java
----------------------------------------------------------------------
diff --git 
a/flink-core/src/main/java/org/apache/flink/core/memory/MemorySegment.java 
b/flink-core/src/main/java/org/apache/flink/core/memory/MemorySegment.java
index 31d5563..8d13921 100644
--- a/flink-core/src/main/java/org/apache/flink/core/memory/MemorySegment.java
+++ b/flink-core/src/main/java/org/apache/flink/core/memory/MemorySegment.java
@@ -210,7 +210,7 @@ public abstract class MemorySegment {
        }
 
        /**
-        * Wraps the chunk of the underlying memory located between 
<tt>offset<tt> and 
+        * Wraps the chunk of the underlying memory located between 
<tt>offset</tt> and
         * <tt>length</tt> in a NIO ByteBuffer.
         *
         * @param offset The offset in the memory segment.
@@ -1220,7 +1220,7 @@ public abstract class MemorySegment {
         * @param offset2 Offset of seg2 to start comparing
         * @param len Length of the compared memory region
         *
-        * @return 0 if equal, -1 if seg1 < seg2, 1 otherwise
+        * @return 0 if equal, -1 if seg1 &lt; seg2, 1 otherwise
         */
        public final int compare(MemorySegment seg2, int offset1, int offset2, 
int len) {
                while (len >= 8) {

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/types/Record.java
----------------------------------------------------------------------
diff --git a/flink-core/src/main/java/org/apache/flink/types/Record.java 
b/flink-core/src/main/java/org/apache/flink/types/Record.java
index 24ff979..8ef972f 100644
--- a/flink-core/src/main/java/org/apache/flink/types/Record.java
+++ b/flink-core/src/main/java/org/apache/flink/types/Record.java
@@ -803,7 +803,7 @@ public final class Record implements Value, 
CopyableValue<Record> {
         * Bin-copies fields from a source record to this record. The following 
caveats apply:
         * 
         * If the source field is in a modified state, no binary representation 
will exist yet.
-        * In that case, this method is equivalent to setField(..., 
source.getField(..., <class>)). 
+        * In that case, this method is equivalent to {@code setField(..., 
source.getField(..., <class>))}.
         * In particular, if setValue is called on the source field Value 
instance, that change 
         * will propagate to this record.
         * 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/types/StringValue.java
----------------------------------------------------------------------
diff --git a/flink-core/src/main/java/org/apache/flink/types/StringValue.java 
b/flink-core/src/main/java/org/apache/flink/types/StringValue.java
index 2249019..873e1dd 100644
--- a/flink-core/src/main/java/org/apache/flink/types/StringValue.java
+++ b/flink-core/src/main/java/org/apache/flink/types/StringValue.java
@@ -377,7 +377,7 @@ public class StringValue implements 
NormalizableKey<StringValue>, CharSequence,
         * @param prefix The prefix character sequence.
         * @param startIndex The position to start checking for the prefix.
         * 
-        * @return True, if this StringValue substring, starting at position 
<code>startIndex</code> has </code>prefix</code>
+        * @return True, if this StringValue substring, starting at position 
<code>startIndex</code> has <code>prefix</code>
         *         as its prefix.
         */
        public boolean startsWith(CharSequence prefix, int startIndex) {
@@ -403,7 +403,7 @@ public class StringValue implements 
NormalizableKey<StringValue>, CharSequence,
         * 
         * @param prefix The prefix character sequence.
         * 
-        * @return True, if this StringValue has </code>prefix</code> as its 
prefix.
+        * @return True, if this StringValue has <code>prefix</code> as its 
prefix.
         */
        public boolean startsWith(CharSequence prefix) {
                return startsWith(prefix, 0);

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-core/src/main/java/org/apache/flink/util/Visitable.java
----------------------------------------------------------------------
diff --git a/flink-core/src/main/java/org/apache/flink/util/Visitable.java 
b/flink-core/src/main/java/org/apache/flink/util/Visitable.java
index 49a7866..559cb96 100644
--- a/flink-core/src/main/java/org/apache/flink/util/Visitable.java
+++ b/flink-core/src/main/java/org/apache/flink/util/Visitable.java
@@ -38,7 +38,7 @@ public interface Visitable<T extends Visitable<T>> {
         * and then invokes the post-visit method.
         * <p>
         * A typical code example is the following:
-        * <code>
+        * <pre>{@code
         * public void accept(Visitor<Operator> visitor) {
         *     boolean descend = visitor.preVisit(this);        
         *     if (descend) {
@@ -48,7 +48,7 @@ public interface Visitable<T extends Visitable<T>> {
         *         visitor.postVisit(this);
         *     }
         * }
-        * </code>
+        * }</pre>
         * 
         * @param visitor The visitor to be called with this object as the 
parameter.
         * 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/clustering/KMeans.java
----------------------------------------------------------------------
diff --git 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/clustering/KMeans.java
 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/clustering/KMeans.java
index 73f90ca..2db6f65 100644
--- 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/clustering/KMeans.java
+++ 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/clustering/KMeans.java
@@ -198,7 +198,7 @@ public class KMeans {
        //     USER FUNCTIONS
        // 
*************************************************************************
        
-       /** Converts a Tuple2<Double,Double> into a Point. */
+       /** Converts a {@code Tuple2<Double,Double>} into a Point. */
        @ForwardedFields("0->x; 1->y")
        public static final class TuplePointConverter implements 
MapFunction<Tuple2<Double, Double>, Point> {
 
@@ -208,7 +208,7 @@ public class KMeans {
                }
        }
        
-       /** Converts a Tuple3<Integer, Double,Double> into a Centroid. */
+       /** Converts a {@code Tuple3<Integer, Double,Double>} into a Centroid. 
*/
        @ForwardedFields("0->id; 1->x; 2->y")
        public static final class TupleCentroidConverter implements 
MapFunction<Tuple3<Integer, Double, Double>, Centroid> {
 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/clustering/util/KMeansDataGenerator.java
----------------------------------------------------------------------
diff --git 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/clustering/util/KMeansDataGenerator.java
 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/clustering/util/KMeansDataGenerator.java
index 8f44034..8f48d0a 100644
--- 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/clustering/util/KMeansDataGenerator.java
+++ 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/clustering/util/KMeansDataGenerator.java
@@ -66,6 +66,8 @@ public class KMeansDataGenerator {
         * <li><b>Optional</b> Double: Value range of cluster centers
         * <li><b>Optional</b> Long: Random seed
         * </ol>
+        *
+        * @throws IOException
         */
        public static void main(String[] args) throws IOException {
 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/distcp/DistCp.java
----------------------------------------------------------------------
diff --git 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/distcp/DistCp.java
 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/distcp/DistCp.java
index 08f90a6..8e87892 100644
--- 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/distcp/DistCp.java
+++ 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/distcp/DistCp.java
@@ -51,7 +51,7 @@ import java.util.Map;
  * (see <a 
href="http://hadoop.apache.org/docs/r1.2.1/distcp.html";>http://hadoop.apache.org/docs/r1.2.1/distcp.html</a>)
  * with a dynamic input format
  * Note that this tool does not deal with retriability. Additionally, empty 
directories are not copied over.
- * <p/>
+ * <p>
  * When running locally, local file systems paths can be used.
  * However, in a distributed environment HDFS paths must be provided both as 
input and output.
  */

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/graph/PageRankBasic.java
----------------------------------------------------------------------
diff --git 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/graph/PageRankBasic.java
 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/graph/PageRankBasic.java
index f05a15d..7b05158 100644
--- 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/graph/PageRankBasic.java
+++ 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/graph/PageRankBasic.java
@@ -53,7 +53,7 @@ import org.apache.flink.examples.java.graph.util.PageRankData;
  * For example <code>"1\n2\n12\n42\n63"</code> gives five pages with IDs 1, 2, 
12, 42, and 63.
  * <li>Links are represented as pairs of page IDs which are separated by space 
  * characters. Links are separated by new-line characters.<br>
- * For example <code>"1 2\n2 12\n1 12\n42 63"</code> gives four (directed) 
links (1)->(2), (2)->(12), (1)->(12), and (42)->(63).<br>
+ * For example <code>"1 2\n2 12\n1 12\n42 63"</code> gives four (directed) 
links (1)-&gt;(2), (2)-&gt;(12), (1)-&gt;(12), and (42)-&gt;(63).<br>
  * For this simple implementation it is required that each page has at least 
one incoming and one outgoing link (a page can point to itself).
  * </ul>
  * 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/ml/LinearRegression.java
----------------------------------------------------------------------
diff --git 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/ml/LinearRegression.java
 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/ml/LinearRegression.java
index 341daa6..9c3356c 100644
--- 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/ml/LinearRegression.java
+++ 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/ml/LinearRegression.java
@@ -182,7 +182,7 @@ public class LinearRegression {
        //     USER FUNCTIONS
        // 
*************************************************************************
 
-       /** Converts a Tuple2<Double,Double> into a Data. */
+       /** Converts a {@code Tuple2<Double,Double>} into a Data. */
        @ForwardedFields("0->x; 1->y")
        public static final class TupleDataConverter implements 
MapFunction<Tuple2<Double, Double>, Data> {
 
@@ -192,7 +192,7 @@ public class LinearRegression {
                }
        }
 
-       /** Converts a Tuple2<Double,Double> into a Params. */
+       /** Converts a {@code Tuple2<Double,Double>} into a Params. */
        @ForwardedFields("0->theta0; 1->theta1")
        public static final class TupleParamsConverter implements 
MapFunction<Tuple2<Double, Double>,Params> {
 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/relational/TPCHQuery10.java
----------------------------------------------------------------------
diff --git 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/relational/TPCHQuery10.java
 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/relational/TPCHQuery10.java
index 3dc0472..495ff26 100644
--- 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/relational/TPCHQuery10.java
+++ 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/relational/TPCHQuery10.java
@@ -39,7 +39,7 @@ import org.apache.flink.api.java.ExecutionEnvironment;
  * This program implements the following SQL equivalent:
  * 
  * <p>
- * <code><pre>
+ * <pre>{@code
  * SELECT 
  *        c_custkey,
  *        c_name, 
@@ -64,7 +64,7 @@ import org.apache.flink.api.java.ExecutionEnvironment;
  *        c_acctbal, 
  *        n_name, 
  *        c_address
- * </pre></code>
+ * }</pre>
  *        
  * <p>
  * Compared to the original TPC-H query this version does not print 

http://git-wip-us.apache.org/repos/asf/flink/blob/680b5a90/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/relational/TPCHQuery3.java
----------------------------------------------------------------------
diff --git 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/relational/TPCHQuery3.java
 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/relational/TPCHQuery3.java
index 9a6e58c..a9e47f6 100644
--- 
a/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/relational/TPCHQuery3.java
+++ 
b/flink-examples/flink-java-examples/src/main/java/org/apache/flink/examples/java/relational/TPCHQuery3.java
@@ -42,7 +42,7 @@ import org.apache.flink.api.java.tuple.Tuple4;
  * This program implements the following SQL equivalent:
  *
  * <p>
- * <code><pre>
+ * <pre>{@code
  * SELECT 
  *      l_orderkey, 
  *      SUM(l_extendedprice*(1-l_discount)) AS revenue,
@@ -61,7 +61,7 @@ import org.apache.flink.api.java.tuple.Tuple4;
  *      l_orderkey, 
  *      o_orderdate, 
  *      o_shippriority;
- * </pre></code>
+ * }</pre>
  *
  * <p>
  * Compared to the original TPC-H query this version does not sort the result 
by revenue

Reply via email to