flink git commit: [hotfix][kinesis] Fix build instructions

2016-08-08 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/release-1.1 e4f62d3b3 -> b054da46e


[hotfix][kinesis] Fix build instructions


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/b054da46
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/b054da46
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/b054da46

Branch: refs/heads/release-1.1
Commit: b054da46ea00e958b1c055f6e4f8b5dd51d3a390
Parents: e4f62d3
Author: Robert Metzger 
Authored: Mon Aug 8 12:19:03 2016 +0200
Committer: Robert Metzger 
Committed: Mon Aug 8 12:23:51 2016 +0200

--
 docs/apis/streaming/connectors/kinesis.md | 3 +++
 1 file changed, 3 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/b054da46/docs/apis/streaming/connectors/kinesis.md
--
diff --git a/docs/apis/streaming/connectors/kinesis.md 
b/docs/apis/streaming/connectors/kinesis.md
index e4ded62..63683e4 100644
--- a/docs/apis/streaming/connectors/kinesis.md
+++ b/docs/apis/streaming/connectors/kinesis.md
@@ -47,6 +47,9 @@ Flink releases because of the licensing issue. Therefore, you 
need to build the
 Download the Flink source or check it out from the git repository. Then, use 
the following Maven command to build the module:
 {% highlight bash %}
 mvn clean install -Pinclude-kinesis -DskipTests
+# In Maven 3.3 the shading of flink-dist doesn't work properly in one run, so 
we need to run mvn for flink-dist again. 
+cd flink-dist
+mvn clean install -Pinclude-kinesis -DskipTests
 {% endhighlight %}
 
 



flink git commit: [hotfix][kinesis] Fix build instructions

2016-08-08 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master 27232fe32 -> 6a883776d


[hotfix][kinesis] Fix build instructions


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/6a883776
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/6a883776
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/6a883776

Branch: refs/heads/master
Commit: 6a883776da244a7dcc37ee5c24c5705cc4f55a4d
Parents: 27232fe
Author: Robert Metzger 
Authored: Mon Aug 8 12:19:03 2016 +0200
Committer: Robert Metzger 
Committed: Mon Aug 8 12:19:03 2016 +0200

--
 docs/apis/streaming/connectors/kinesis.md | 3 +++
 1 file changed, 3 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/6a883776/docs/apis/streaming/connectors/kinesis.md
--
diff --git a/docs/apis/streaming/connectors/kinesis.md 
b/docs/apis/streaming/connectors/kinesis.md
index e4ded62..63683e4 100644
--- a/docs/apis/streaming/connectors/kinesis.md
+++ b/docs/apis/streaming/connectors/kinesis.md
@@ -47,6 +47,9 @@ Flink releases because of the licensing issue. Therefore, you 
need to build the
 Download the Flink source or check it out from the git repository. Then, use 
the following Maven command to build the module:
 {% highlight bash %}
 mvn clean install -Pinclude-kinesis -DskipTests
+# In Maven 3.3 the shading of flink-dist doesn't work properly in one run, so 
we need to run mvn for flink-dist again. 
+cd flink-dist
+mvn clean install -Pinclude-kinesis -DskipTests
 {% endhighlight %}
 
 



flink git commit: [FLINK-4080] Guarantee exactly-once for Kinesis consumer for failures in the middle of aggregated records

2016-06-29 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master dbe41f487 -> fa42cdabf


[FLINK-4080] Guarantee exactly-once for Kinesis consumer for failures in the 
middle of aggregated records

This closes #2180


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/fa42cdab
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/fa42cdab
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/fa42cdab

Branch: refs/heads/master
Commit: fa42cdabfdd37fa3c2f86198ca04d37113ddda00
Parents: dbe41f4
Author: Gordon Tai 
Authored: Wed Jun 29 15:46:35 2016 +0800
Committer: Robert Metzger 
Committed: Wed Jun 29 13:36:54 2016 +0200

--
 .../kinesis/FlinkKinesisConsumer.java   |  23 ++--
 .../kinesis/internals/KinesisDataFetcher.java   |  11 +-
 .../kinesis/internals/ShardConsumerThread.java  | 110 +--
 .../kinesis/model/SentinelSequenceNumber.java   |  18 ++-
 .../kinesis/model/SequenceNumber.java   | 104 ++
 .../kinesis/FlinkKinesisConsumerTest.java   |  15 +--
 .../internals/KinesisDataFetcherTest.java   |   3 +-
 .../internals/ShardConsumerThreadTest.java  |  47 +---
 8 files changed, 254 insertions(+), 77 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/fa42cdab/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java
--
diff --git 
a/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java
 
b/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java
index 9cbc9d9..19b36e4 100644
--- 
a/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java
+++ 
b/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java
@@ -28,6 +28,7 @@ import 
org.apache.flink.streaming.connectors.kinesis.config.KinesisConfigConstan
 import 
org.apache.flink.streaming.connectors.kinesis.internals.KinesisDataFetcher;
 import org.apache.flink.streaming.connectors.kinesis.model.KinesisStreamShard;
 import 
org.apache.flink.streaming.connectors.kinesis.model.SentinelSequenceNumber;
+import org.apache.flink.streaming.connectors.kinesis.model.SequenceNumber;
 import org.apache.flink.streaming.connectors.kinesis.proxy.KinesisProxy;
 import 
org.apache.flink.streaming.connectors.kinesis.serialization.KinesisDeserializationSchema;
 import 
org.apache.flink.streaming.connectors.kinesis.serialization.KinesisDeserializationSchemaWrapper;
@@ -64,7 +65,7 @@ import static 
org.apache.flink.util.Preconditions.checkNotNull;
  * @param  the type of data emitted
  */
 public class FlinkKinesisConsumer extends RichParallelSourceFunction
-   implements CheckpointedAsynchronously>, ResultTypeQueryable {
+   implements CheckpointedAsynchronously>, ResultTypeQueryable {
 
private static final long serialVersionUID = 4724006128720664870L;
 
@@ -92,10 +93,10 @@ public class FlinkKinesisConsumer extends 
RichParallelSourceFunction
private transient KinesisDataFetcher fetcher;
 
/** The sequence numbers of the last fetched data records from Kinesis 
by this task */
-   private transient HashMap lastSequenceNums;
+   private transient HashMap 
lastSequenceNums;
 
/** The sequence numbers to restore to upon restore from failure */
-   private transient HashMap 
sequenceNumsToRestore;
+   private transient HashMap 
sequenceNumsToRestore;
 
private volatile boolean hasAssignedShards;
 
@@ -227,14 +228,14 @@ public class FlinkKinesisConsumer extends 
RichParallelSourceFunction
LOG.info("Consumer task {} is restoring 
sequence numbers from previous checkpointed state", thisConsumerTaskIndex);
}
 
-   for (Map.Entry 
restoreSequenceNum : sequenceNumsToRestore.entrySet()) {
+   for (Map.Entry 
restoreSequenceNum : sequenceNumsToRestore.entrySet()) {
// advance the corresponding shard to the last 
known sequence number

fetcher.advanceSequenceNumberTo(restoreSequenceNum.getKey(), 
restoreSequenceNum.getValue());

flink git commit: [FLINK-4033] Polish up Kinesis connector documentation

2016-06-29 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master fa42cdabf -> 256c9c4da


[FLINK-4033] Polish up Kinesis connector documentation

Includes:
1. Scala examples for consumer and producer
2. Add information about AWS Kinesis service usage
3. Add Kinesis connecter to the fault tolerance guarantees table
4. Minor typo fix in Kafka documentation

This closes #2181


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/256c9c4d
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/256c9c4d
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/256c9c4d

Branch: refs/heads/master
Commit: 256c9c4daf11349128eaa2e71a434f609e57053c
Parents: fa42cda
Author: Gordon Tai 
Authored: Wed Jun 29 18:29:08 2016 +0800
Committer: Robert Metzger 
Committed: Wed Jun 29 13:41:35 2016 +0200

--
 docs/apis/streaming/connectors/kafka.md   |  2 +-
 docs/apis/streaming/connectors/kinesis.md | 63 ++
 docs/apis/streaming/fault_tolerance.md| 16 +--
 3 files changed, 68 insertions(+), 13 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/256c9c4d/docs/apis/streaming/connectors/kafka.md
--
diff --git a/docs/apis/streaming/connectors/kafka.md 
b/docs/apis/streaming/connectors/kafka.md
index 9bd70be..e7cd05b 100644
--- a/docs/apis/streaming/connectors/kafka.md
+++ b/docs/apis/streaming/connectors/kafka.md
@@ -236,7 +236,7 @@ properties.setProperty("bootstrap.servers", 
"localhost:9092");
 properties.setProperty("zookeeper.connect", "localhost:2181");
 properties.setProperty("group.id", "test");
 
-val myConsumer = new FlinkKafkaConsumer08[Stirng]("topic", new 
SimpleStringSchema(), properties);
+val myConsumer = new FlinkKafkaConsumer08[String]("topic", new 
SimpleStringSchema(), properties);
 myConsumer.assignTimestampsAndWatermarks(new CustomWatermarkEmitter());
 stream = env
 .addSource(myConsumer)

http://git-wip-us.apache.org/repos/asf/flink/blob/256c9c4d/docs/apis/streaming/connectors/kinesis.md
--
diff --git a/docs/apis/streaming/connectors/kinesis.md 
b/docs/apis/streaming/connectors/kinesis.md
index 66c078a..db3a9c4 100644
--- a/docs/apis/streaming/connectors/kinesis.md
+++ b/docs/apis/streaming/connectors/kinesis.md
@@ -52,12 +52,16 @@ mvn clean install -Pinclude-kinesis -DskipTests
 
 
 Note that the streaming connectors are not part of the binary distribution. 
-See linking with them for cluster execution 
[here]({{site.baseurl}}/apis/cluster_execution.html#linking-with-modules-not-contained-in-the-binary-distribution).
+See how to link with them for cluster execution 
[here]({{site.baseurl}}/apis/cluster_execution.html#linking-with-modules-not-contained-in-the-binary-distribution).
 
- Usage of Consumer
+### Using the Amazon Kinesis Streams Service
+Follow the instructions from the [Amazon Kinesis Streams Developer 
Guide](https://docs.aws.amazon.com/streams/latest/dev/learning-kinesis-module-one-create-stream.html)
+to setup Kinesis streams. Make sure to create the appropriate IAM policy and 
user to read / write to the Kinesis streams.
+
+### Kinesis Consumer
 
 The `FlinkKinesisConsumer` can be used to pull data from multiple Kinesis 
streams within the same AWS region in parallel.
-It participates in Flink's distributed snapshot checkpointing and provides 
exactly-once processing guarantees. Note
+It participates in Flink's distributed snapshot checkpointing and provides 
exactly-once user-defined state update guarantees. Note
 that the current version can not handle resharding of Kinesis streams. When 
Kinesis streams are resharded, the consumer
 will fail and the Flink streaming job must be resubmitted.
 
@@ -78,10 +82,28 @@ 
kinesisConsumerConfig.put(KinesisConfigConstants.CONFIG_STREAM_INIT_POSITION_TYP
 
 StreamExecutionEnvironment env = StreamExecutionEnvironment.getEnvironment();
 
-DataStream kinesisRecords = env.addSource(new FlinkKinesisConsumer<>(
+DataStream kinesis = env.addSource(new FlinkKinesisConsumer<>(
 "kinesis_stream_name", new SimpleStringSchema(), kinesisConsumerConfig))
 {% endhighlight %}
 
+
+{% highlight scala %}
+val kinesisConsumerConfig = new Properties();
+kinesisConsumerConfig.put(KinesisConfigConstants.CONFIG_AWS_REGION, 
"us-east-1");
+kinesisConsumerConfig.put(
+KinesisConfigConstants.CONFIG_AWS_CREDENTIALS_PROVIDER_BASIC_ACCESSKEYID,
+"aws_access_key_id_here");
+kinesisConsumerConfig.put(
+KinesisConfigConstants.CONFIG_AWS_CREDENTIALS_PROVIDER_BASIC_SECRETKEY,
+"aws_secret_key_here");
+kinesisConsumerConfig.put(KinesisConfigConstants.CONFIG_STREAM_INIT_POSITION_TYPE,
 "LATEST");
+
+val env = StreamExecutionEnvironment.getEnvironment
+

flink git commit: [FLINK-4085][Kinesis] Set Flink-specific user agent

2016-06-29 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master 256c9c4da -> 123be2276


[FLINK-4085][Kinesis] Set Flink-specific user agent

This closes #2175


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/123be227
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/123be227
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/123be227

Branch: refs/heads/master
Commit: 123be22765ebf32a26e3b74dbbd00a36999a5862
Parents: 256c9c4
Author: Robert Metzger 
Authored: Tue Jun 28 13:54:12 2016 +0200
Committer: Robert Metzger 
Committed: Wed Jun 29 13:43:04 2016 +0200

--
 .../connectors/kinesis/proxy/KinesisProxy.java | 17 +++--
 1 file changed, 11 insertions(+), 6 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/123be227/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/proxy/KinesisProxy.java
--
diff --git 
a/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/proxy/KinesisProxy.java
 
b/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/proxy/KinesisProxy.java
index b4ae3f5..d035c03 100644
--- 
a/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/proxy/KinesisProxy.java
+++ 
b/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/proxy/KinesisProxy.java
@@ -17,6 +17,8 @@
 
 package org.apache.flink.streaming.connectors.kinesis.proxy;
 
+import com.amazonaws.ClientConfiguration;
+import com.amazonaws.ClientConfigurationFactory;
 import com.amazonaws.regions.Region;
 import com.amazonaws.regions.Regions;
 import com.amazonaws.services.kinesis.AmazonKinesisClient;
@@ -29,6 +31,7 @@ import 
com.amazonaws.services.kinesis.model.LimitExceededException;
 import com.amazonaws.services.kinesis.model.ResourceNotFoundException;
 import com.amazonaws.services.kinesis.model.StreamStatus;
 import com.amazonaws.services.kinesis.model.Shard;
+import org.apache.flink.runtime.util.EnvironmentInformation;
 import 
org.apache.flink.streaming.connectors.kinesis.config.KinesisConfigConstants;
 import org.apache.flink.streaming.connectors.kinesis.model.KinesisStreamShard;
 import org.apache.flink.streaming.connectors.kinesis.util.AWSUtil;
@@ -58,9 +61,6 @@ public class KinesisProxy {
/** The actual Kinesis client from the AWS SDK that we will be using to 
make calls */
private final AmazonKinesisClient kinesisClient;
 
-   /** The AWS region that this proxy will be making calls to */
-   private final String regionId;
-
/** Configuration properties of this Flink Kinesis Connector */
private final Properties configProps;
 
@@ -72,9 +72,14 @@ public class KinesisProxy {
public KinesisProxy(Properties configProps) {
this.configProps = checkNotNull(configProps);
 
-   this.regionId = 
configProps.getProperty(KinesisConfigConstants.CONFIG_AWS_REGION);
-   AmazonKinesisClient client = new 
AmazonKinesisClient(AWSUtil.getCredentialsProvider(configProps).getCredentials());
-   
client.setRegion(Region.getRegion(Regions.fromName(this.regionId)));
+   /* The AWS region that this proxy will be making calls to */
+   String regionId = 
configProps.getProperty(KinesisConfigConstants.CONFIG_AWS_REGION);
+   // set Flink as a user agent
+   ClientConfiguration config = new 
ClientConfigurationFactory().getConfig();
+   config.setUserAgent("Apache Flink " + 
EnvironmentInformation.getVersion() + " (" + 
EnvironmentInformation.getRevisionInformation().commitId + ") Kinesis 
Connector");
+   AmazonKinesisClient client = new 
AmazonKinesisClient(AWSUtil.getCredentialsProvider(configProps).getCredentials(),
 config);
+
+   client.setRegion(Region.getRegion(Regions.fromName(regionId)));
 
this.kinesisClient = client;
}



flink git commit: [FLINK-4027] Flush FlinkKafkaProducer on checkpoints

2016-07-04 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master ffaf10d22 -> 7206b0ed2


[FLINK-4027] Flush FlinkKafkaProducer on checkpoints

This closes #2108

This closes #2058 because its an invalid pull request.


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/7206b0ed
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/7206b0ed
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/7206b0ed

Branch: refs/heads/master
Commit: 7206b0ed2adb10c94e1ffd3dbe851250b44edcf4
Parents: ffaf10d
Author: Robert Metzger 
Authored: Wed Jun 15 17:50:38 2016 +0200
Committer: Robert Metzger 
Committed: Mon Jul 4 11:55:09 2016 +0200

--
 .../connectors/kafka/FlinkKafkaProducer08.java  |  18 ++
 .../kafka/KafkaTestEnvironmentImpl.java |   4 +-
 .../connectors/kafka/FlinkKafkaProducer09.java  |   7 +
 .../kafka/KafkaTestEnvironmentImpl.java |   4 +-
 .../kafka/FlinkKafkaProducerBase.java   | 120 --
 .../kafka/AtLeastOnceProducerTest.java  | 218 +++
 .../connectors/kafka/KafkaConsumerTestBase.java |   3 +-
 7 files changed, 358 insertions(+), 16 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/7206b0ed/flink-streaming-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer08.java
--
diff --git 
a/flink-streaming-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer08.java
 
b/flink-streaming-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer08.java
index 4975f9a..e509d2f 100644
--- 
a/flink-streaming-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer08.java
+++ 
b/flink-streaming-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer08.java
@@ -125,4 +125,22 @@ public class FlinkKafkaProducer08 extends 
FlinkKafkaProducerBase  {
super(topicId, serializationSchema, producerConfig, 
customPartitioner);
}
 
+   @Override
+   protected void flush() {
+   // The Kafka 0.8 producer doesn't support flushing, we wait here
+   // until all pending records are confirmed
+   //noinspection SynchronizeOnNonFinalField
+   synchronized (pendingRecordsLock) {
+   while (pendingRecords > 0) {
+   try {
+   pendingRecordsLock.wait();
+   } catch (InterruptedException e) {
+   // this can be interrupted when the 
Task has been cancelled.
+   // by throwing an exception, we ensure 
that this checkpoint doesn't get confirmed
+   throw new RuntimeException("Flushing 
got interrupted while checkpointing", e);
+   }
+   }
+   }
+   }
+
 }

http://git-wip-us.apache.org/repos/asf/flink/blob/7206b0ed/flink-streaming-connectors/flink-connector-kafka-0.8/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTestEnvironmentImpl.java
--
diff --git 
a/flink-streaming-connectors/flink-connector-kafka-0.8/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTestEnvironmentImpl.java
 
b/flink-streaming-connectors/flink-connector-kafka-0.8/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTestEnvironmentImpl.java
index 77d41ac..75ca9ed 100644
--- 
a/flink-streaming-connectors/flink-connector-kafka-0.8/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTestEnvironmentImpl.java
+++ 
b/flink-streaming-connectors/flink-connector-kafka-0.8/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTestEnvironmentImpl.java
@@ -97,7 +97,9 @@ public class KafkaTestEnvironmentImpl extends 
KafkaTestEnvironment {
 
@Override
public  FlinkKafkaProducerBase getProducer(String topic, 
KeyedSerializationSchema serSchema, Properties props, KafkaPartitioner 
partitioner) {
-   return new FlinkKafkaProducer08(topic, serSchema, props, 
partitioner);
+   FlinkKafkaProducer08 prod = new 
FlinkKafkaProducer08(topic, serSchema, props, partitioner);
+   prod.setFlushOnCheckpoint(true);
+   return prod;
}
 
@Override


flink git commit: [hotfix] Fix NPE in NetworkStackThroughputITCase

2016-07-04 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master 7206b0ed2 -> 7ee0425a3


[hotfix] Fix NPE in NetworkStackThroughputITCase


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/7ee0425a
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/7ee0425a
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/7ee0425a

Branch: refs/heads/master
Commit: 7ee0425a3467f9cf8d1dc967c99b5bb6c29c1745
Parents: 7206b0e
Author: Robert Metzger 
Authored: Mon Jul 4 12:32:10 2016 +0200
Committer: Robert Metzger 
Committed: Mon Jul 4 12:32:10 2016 +0200

--
 .../apache/flink/test/runtime/NetworkStackThroughputITCase.java   | 3 +++
 1 file changed, 3 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/7ee0425a/flink-tests/src/test/java/org/apache/flink/test/runtime/NetworkStackThroughputITCase.java
--
diff --git 
a/flink-tests/src/test/java/org/apache/flink/test/runtime/NetworkStackThroughputITCase.java
 
b/flink-tests/src/test/java/org/apache/flink/test/runtime/NetworkStackThroughputITCase.java
index 5506f55..3d0e5ab 100644
--- 
a/flink-tests/src/test/java/org/apache/flink/test/runtime/NetworkStackThroughputITCase.java
+++ 
b/flink-tests/src/test/java/org/apache/flink/test/runtime/NetworkStackThroughputITCase.java
@@ -287,9 +287,12 @@ public class NetworkStackThroughputITCase {
config.setInteger(NUM_SLOTS_PER_TM_CONFIG_KEY, 
(Integer) p[5]);
 
TestBaseWrapper test = new TestBaseWrapper(config);
+   test.startCluster();
 
System.out.println(Arrays.toString(p));
test.testProgram();
+
+   test.stopCluster();
}
}
 



flink git commit: [FLINK-4394] RMQSource: QueueName accessible for subclasses

2016-08-15 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master 90fdae452 -> 5ccd90715


[FLINK-4394] RMQSource: QueueName accessible for subclasses

The queueName is needed if the subclasses override `setupQueue`.

This closes #2373


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/5ccd9071
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/5ccd9071
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/5ccd9071

Branch: refs/heads/master
Commit: 5ccd9071580e196d150905b2d05eef71e399a24c
Parents: 90fdae4
Author: Dominik 
Authored: Mon Aug 15 09:41:57 2016 +0200
Committer: Robert Metzger 
Committed: Mon Aug 15 16:19:49 2016 +0200

--
 .../org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java  | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/5ccd9071/flink-streaming-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java
--
diff --git 
a/flink-streaming-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java
 
b/flink-streaming-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java
index 33cf52c..ee9c3b9 100644
--- 
a/flink-streaming-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java
+++ 
b/flink-streaming-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java
@@ -74,7 +74,7 @@ public class RMQSource extends 
MultipleIdsMessageAcknowledgingSourceBase schema;
 



flink git commit: [hotfix] Fix Redis Sink to fail at opening if Redis is not initialized.

2016-08-18 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master 47acdeadf -> e9a067229


[hotfix] Fix Redis Sink to fail at opening if Redis is not initialized.

This closes #2245


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/e9a06722
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/e9a06722
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/e9a06722

Branch: refs/heads/master
Commit: e9a067229d787c3e874dc40303ac8661045cc54f
Parents: 47acdea
Author: kl0u 
Authored: Wed Jul 13 18:40:17 2016 +0200
Committer: Robert Metzger 
Committed: Thu Aug 18 16:41:53 2016 +0200

--
 .../streaming/connectors/redis/RedisSink.java   |  8 +-
 .../common/container/RedisClusterContainer.java | 12 ++-
 .../container/RedisCommandsContainer.java   |  7 ++
 .../redis/common/container/RedisContainer.java  | 17 -
 .../connectors/redis/RedisSinkITCase.java   |  2 -
 .../connectors/redis/RedisSinkTest.java | 79 
 6 files changed, 117 insertions(+), 8 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/e9a06722/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/RedisSink.java
--
diff --git 
a/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/RedisSink.java
 
b/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/RedisSink.java
index 43518e8..f6b0fd7 100644
--- 
a/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/RedisSink.java
+++ 
b/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/RedisSink.java
@@ -166,7 +166,13 @@ public class RedisSink extends RichSinkFunction {
  */
@Override
public void open(Configuration parameters) throws Exception {
-   this.redisCommandsContainer = 
RedisCommandsContainerBuilder.build(this.flinkJedisConfigBase);
+   try {
+   this.redisCommandsContainer = 
RedisCommandsContainerBuilder.build(this.flinkJedisConfigBase);
+   this.redisCommandsContainer.open();
+   } catch (Exception e) {
+   LOG.error("Redis has not been properly initialized: ", 
e);
+   throw e;
+   }
}
 
/**

http://git-wip-us.apache.org/repos/asf/flink/blob/e9a06722/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/common/container/RedisClusterContainer.java
--
diff --git 
a/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/common/container/RedisClusterContainer.java
 
b/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/common/container/RedisClusterContainer.java
index 7551c9e..d6621d6 100644
--- 
a/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/common/container/RedisClusterContainer.java
+++ 
b/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/common/container/RedisClusterContainer.java
@@ -33,7 +33,7 @@ public class RedisClusterContainer implements 
RedisCommandsContainer, Closeable
 
private static final Logger LOG = 
LoggerFactory.getLogger(RedisClusterContainer.class);
 
-   private JedisCluster jedisCluster;
+   private transient JedisCluster jedisCluster;
 
/**
 * Initialize Redis command container for Redis cluster.
@@ -47,6 +47,16 @@ public class RedisClusterContainer implements 
RedisCommandsContainer, Closeable
}
 
@Override
+   public void open() throws Exception {
+
+   // echo() tries to open a connection and echos back the
+   // message passed as argument. Here we use it to monitor
+   // if we can communicate with the cluster.
+
+   jedisCluster.echo("Test");
+   }
+
+   @Override
public void hset(final String key, final String hashField, final String 
value) {
try {
jedisCluster.hset(key, hashField, value);

http://git-wip-us.apache.org/repos/asf/flink/blob/e9a06722/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/common/container/RedisCommandsContainer.java
--
diff --git 

svn commit: r18075 - /release/flink/flink-1.2.0/

2017-02-02 Thread rmetzger
Author: rmetzger
Date: Thu Feb  2 11:02:16 2017
New Revision: 18075

Log:
Add files for Flink 1.2.0 release

Added:
release/flink/flink-1.2.0/
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz   (with 
props)
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.asc
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.md5
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.sha
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.11.tgz   (with 
props)
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.11.tgz.asc
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.11.tgz.md5
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.11.tgz.sha
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.10.tgz   (with 
props)
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.10.tgz.asc
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.10.tgz.md5
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.10.tgz.sha
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.11.tgz   (with 
props)
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.11.tgz.asc
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.11.tgz.md5
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.11.tgz.sha
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.10.tgz   (with 
props)
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.10.tgz.asc
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.10.tgz.md5
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.10.tgz.sha
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.11.tgz   (with 
props)
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.11.tgz.asc
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.11.tgz.md5
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.11.tgz.sha
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.10.tgz   (with 
props)
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.10.tgz.asc
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.10.tgz.md5
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.10.tgz.sha
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.11.tgz   (with 
props)
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.11.tgz.asc
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.11.tgz.md5
release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.11.tgz.sha
release/flink/flink-1.2.0/flink-1.2.0-src.tgz   (with props)
release/flink/flink-1.2.0/flink-1.2.0-src.tgz.asc
release/flink/flink-1.2.0/flink-1.2.0-src.tgz.md5
release/flink/flink-1.2.0/flink-1.2.0-src.tgz.sha

Added: release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz
==
Binary file - no diff available.

Propchange: release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz
--
svn:mime-type = application/octet-stream

Added: release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.asc
==
--- release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.asc (added)
+++ release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.asc Thu 
Feb  2 11:02:16 2017
@@ -0,0 +1,17 @@
+-BEGIN PGP SIGNATURE-
+Version: GnuPG v1
+
+iQIcBAABAgAGBQJYjl9uAAoJEBg/aUTZg5FZ/zcP/2oADsw2nizGwCpgjBnozG6D
+84oGwELWAEFVP74VvN8a94UmEt/G0SkQdIPKDFc9SlxjTwwd36q8OkSPEQpkjc2z
+EHyVjIcglygng5bQA2li7WQXhowlW+ZLG78Knktay/E+Pql4056JNZ2lBgAZqTvr
+H1ajHjBZe1jNR7NxioOqa4yibR56otBsIgK1X34goT1MuW8xe2E9UWnxNbDRSgle
+nMCp+HB92DRq8vXF6JEseVs9LwzIXR6VRd/0H8WXc/QIL5/4bpZ3flKBKB8xV4Fo
+MSelQYryfIQ0Vx1UMUqyjrmDXyOElD6DsYRXsKoOBet05FUuPGCqRn1TRSj4dOcu
+hIVohbsQW203rVylGNYI6mWrjhY/F3ayTDkVcJ0PI8yUcX5SneM0Ug3y5Sp5ckRv
+Z9hMrS52/26YqpPyvWXa8sTX0L+9csEy1+HWZVT/8+Ng7hmlxL2y2Vkc3vMOvcGP
+SIoLL1+KrVo7B4lZ1NaS2cDu2jNN2kpe5O7DfWlo+atDxhENa0AcToZNXrmt2lw3
+znXAdzrLJdljln7qjSeLQpzmTF4j9YfGDSKn3T6vd9A1CUP+c3di8rnhKZ1WKlRq
+AqW50MRjJy/YAasgkSBR8EOtJjghmqZzC5PMfVPjQA79u7BckVb4OzT3XiIOF9jf
+NNLcdmxmPYUsqsww5ybO
+=NyCB
+-END PGP SIGNATURE-

Added: release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.md5
==
--- release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.md5 (added)
+++ release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.md5 Thu 
Feb  2 11:02:16 2017
@@ -0,0 +1 @@
+792225bf4256b8ae2b2d9582a505c41c  flink-1.2.0-bin-hadoop2-scala_2.10.tgz

Added: release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.sha
==
--- release

svn commit: r18076 - in /release/flink: flink-0.10.2/ flink-1.0.3/ flink-1.1.1/ flink-1.1.2/ flink-1.1.3/

2017-02-02 Thread rmetzger
Author: rmetzger
Date: Thu Feb  2 11:08:41 2017
New Revision: 18076

Log:
Delete old release files (they are still in the archive)

Removed:
release/flink/flink-0.10.2/
release/flink/flink-1.0.3/
release/flink/flink-1.1.1/
release/flink/flink-1.1.2/
release/flink/flink-1.1.3/



[flink] Git Push Summary

2017-02-07 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/release-1.2.0-rc0 [deleted] f3c59ceda


[flink] Git Push Summary

2017-02-07 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/release-1.2.0-rc2 [deleted] 8b5b6a8b2


[flink] Git Push Summary

2017-02-07 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/release-1.2.0-rc1 [deleted] 1bca12088


[flink] Git Push Summary

2017-02-07 Thread rmetzger
Repository: flink
Updated Tags:  refs/tags/release-1.2.0 [created] 1c659cf4a


[2/2] flink git commit: [FLINK-5153] Support YARN application tags

2017-02-08 Thread rmetzger
[FLINK-5153] Support YARN application tags

Adds a new config option `yarn.tags`, a comma-separated list of strings
passed to YARN as application tags.


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/d9c116e5
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/d9c116e5
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/d9c116e5

Branch: refs/heads/master
Commit: d9c116e542889b6ad00485529b97652dc2d59cad
Parents: e24a866
Author: Patrick Lucas 
Authored: Fri Feb 3 19:17:55 2017 -0500
Committer: Robert Metzger 
Committed: Wed Feb 8 18:39:47 2017 +0100

--
 docs/setup/config.md|   2 +
 .../flink/configuration/ConfigConstants.java|   5 +
 .../yarn/AbstractYarnClusterDescriptor.java | 114 ++-
 3 files changed, 93 insertions(+), 28 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/d9c116e5/docs/setup/config.md
--
diff --git a/docs/setup/config.md b/docs/setup/config.md
index 2872cfa..2accdc2 100644
--- a/docs/setup/config.md
+++ b/docs/setup/config.md
@@ -431,6 +431,8 @@ use the `env.java.opts` setting, which is the `%jvmopts%` 
variable in the String
 
   For example when running Flink on YARN on an environment with a restrictive 
firewall, this option allows specifying a range of allowed ports.
 
+- `yarn.tags` A comma-separated list of tags to apply to the Flink YARN 
application.
+
 ### Mesos
 
 

http://git-wip-us.apache.org/repos/asf/flink/blob/d9c116e5/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java
--
diff --git 
a/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java 
b/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java
index 14ba9dd..c608fde 100644
--- 
a/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java
+++ 
b/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java
@@ -442,6 +442,11 @@ public final class ConfigConstants {
 */
public static final String YARN_APPLICATION_MASTER_PORT = 
"yarn.application-master.port";
 
+   /**
+* A comma-separated list of strings to use as YARN application tags.
+*/
+   public static final String YARN_APPLICATION_TAGS = "yarn.tags";
+
 
//  Mesos Configuration 
 

http://git-wip-us.apache.org/repos/asf/flink/blob/d9c116e5/flink-yarn/src/main/java/org/apache/flink/yarn/AbstractYarnClusterDescriptor.java
--
diff --git 
a/flink-yarn/src/main/java/org/apache/flink/yarn/AbstractYarnClusterDescriptor.java
 
b/flink-yarn/src/main/java/org/apache/flink/yarn/AbstractYarnClusterDescriptor.java
index b537e09..21599c1 100644
--- 
a/flink-yarn/src/main/java/org/apache/flink/yarn/AbstractYarnClusterDescriptor.java
+++ 
b/flink-yarn/src/main/java/org/apache/flink/yarn/AbstractYarnClusterDescriptor.java
@@ -855,7 +855,7 @@ public abstract class AbstractYarnClusterDescriptor 
implements ClusterDescriptor
name = customName;
}
 
-   appContext.setApplicationName(name); // application name
+   appContext.setApplicationName(name);
appContext.setApplicationType("Apache Flink");
appContext.setAMContainerSpec(amContainer);
appContext.setResource(capability);
@@ -863,6 +863,8 @@ public abstract class AbstractYarnClusterDescriptor 
implements ClusterDescriptor
appContext.setQueue(yarnQueue);
}
 
+   setApplicationTags(appContext);
+
// add a hook to clean up in case deployment fails
Thread deploymentFailureHook = new 
DeploymentFailureHook(yarnClient, yarnApplication);
Runtime.getRuntime().addShutdownHook(deploymentFailureHook);
@@ -1024,75 +1026,117 @@ public abstract class AbstractYarnClusterDescriptor 
implements ClusterDescriptor
customName = name;
}
 
-   private void 
activateHighAvailabilitySupport(ApplicationSubmissionContext appContext) throws 
InvocationTargetException, IllegalAccessException {
+   private void 
activateHighAvailabilitySupport(ApplicationSubmissionContext appContext) throws
+   InvocationTargetException, IllegalAccessException {
+
ApplicationSubmissionContextReflector reflector = 
ApplicationSubmissionContextReflector.getInstance();
 

reflector.setKeepContainersAcrossApplicationAttempts(appContext, true);

[1/2] flink git commit: [FLINK-5153] Add test for YARN application tags

2017-02-08 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master e24a866bf -> 95765b6d8


[FLINK-5153] Add test for YARN application tags


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/95765b6d
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/95765b6d
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/95765b6d

Branch: refs/heads/master
Commit: 95765b6d8fdda9c601ad61dc39ce02043ecefa05
Parents: d9c116e
Author: Patrick Lucas 
Authored: Tue Feb 7 12:47:21 2017 -0500
Committer: Robert Metzger 
Committed: Wed Feb 8 18:39:47 2017 +0100

--
 .../YARNSessionCapacitySchedulerITCase.java | 43 
 1 file changed, 35 insertions(+), 8 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/95765b6d/flink-yarn-tests/src/test/java/org/apache/flink/yarn/YARNSessionCapacitySchedulerITCase.java
--
diff --git 
a/flink-yarn-tests/src/test/java/org/apache/flink/yarn/YARNSessionCapacitySchedulerITCase.java
 
b/flink-yarn-tests/src/test/java/org/apache/flink/yarn/YARNSessionCapacitySchedulerITCase.java
index ec66eb2..2a3b6c6 100644
--- 
a/flink-yarn-tests/src/test/java/org/apache/flink/yarn/YARNSessionCapacitySchedulerITCase.java
+++ 
b/flink-yarn-tests/src/test/java/org/apache/flink/yarn/YARNSessionCapacitySchedulerITCase.java
@@ -21,6 +21,7 @@ import com.fasterxml.jackson.databind.JsonNode;
 import com.fasterxml.jackson.databind.ObjectMapper;
 import com.fasterxml.jackson.databind.node.ArrayNode;
 import com.google.common.base.Joiner;
+import com.google.common.collect.Sets;
 import org.apache.commons.io.FileUtils;
 import org.apache.flink.configuration.ConfigConstants;
 import org.apache.flink.configuration.GlobalConfiguration;
@@ -53,13 +54,9 @@ import org.slf4j.LoggerFactory;
 import java.io.File;
 import java.io.FilenameFilter;
 import java.io.IOException;
-import java.util.EnumSet;
-import java.util.LinkedList;
-import java.util.List;
-import java.util.Map;
-import java.util.Collections;
-import java.util.Comparator;
-import java.util.Arrays;
+import java.lang.reflect.InvocationTargetException;
+import java.lang.reflect.Method;
+import java.util.*;
 import java.util.concurrent.ConcurrentMap;
 import java.util.regex.Matcher;
 import java.util.regex.Pattern;
@@ -413,11 +410,14 @@ public class YARNSessionCapacitySchedulerITCase extends 
YarnTestBase {
throw new RuntimeException(e);
}
 
-   Runner runner = startWithArgs(new String[]{"run", "-m", 
"yarn-cluster", "-yj", flinkUberjar.getAbsolutePath(),
+   Runner runner = startWithArgs(new String[]{
+   "run", "-m", "yarn-cluster",
+   "-yj", flinkUberjar.getAbsolutePath(),
"-yt", flinkLibFolder.getAbsolutePath(),
"-yn", "1",
"-yjm", "768",
"-yD", "yarn.heap-cutoff-ratio=0.5", // test if 
the cutoff is passed correctly
+   "-yD", "yarn.tags=test-tag",
"-ytm", "1024",
"-ys", "2", // test requesting slots from YARN.
"--yarndetached", job,
@@ -516,6 +516,7 @@ public class YARNSessionCapacitySchedulerITCase extends 
YarnTestBase {
LOG.info("Got report {}", rep);
} while(rep.getYarnApplicationState() == 
YarnApplicationState.RUNNING);
 
+   verifyApplicationTags(rep);
} catch(Throwable t) {
LOG.warn("Error while detached yarn session was 
running", t);
Assert.fail(t.getMessage());
@@ -543,6 +544,32 @@ public class YARNSessionCapacitySchedulerITCase extends 
YarnTestBase {
}
}
 
+   /**
+* Ensures that the YARN application tags were set properly.
+*
+* Since YARN application tags were only added in Hadoop 2.4, but Flink 
still supports Hadoop 2.3, reflection is
+* required to invoke the methods. If the method does not exist, this 
test passes.
+*/
+   private void verifyApplicationTags(final ApplicationReport report) 
throws InvocationTargetException,
+   IllegalAccessException {
+
+   final Method applicationTagsMethod;
+
+   Class clazz = ApplicationReport.class;
+   try {
+   // this method is only supported by Hadoop 2.4.0 onwards
+   applicationTagsMethod = 
clazz.getMethod("getApplicationTags");
+   } catch (NoSuchMethodException e) {
+

[2/4] flink-web git commit: Add changelog for 1.2.0 release

2017-02-02 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/5b25f84e/content/blog/release_1.2.0-changelog.html
--
diff --git a/content/blog/release_1.2.0-changelog.html 
b/content/blog/release_1.2.0-changelog.html
new file mode 100644
index 000..01ab822
--- /dev/null
+++ b/content/blog/release_1.2.0-changelog.html
@@ -0,0 +1,1545 @@
+
+
+  
+
+
+
+
+Apache Flink: Release 1.2.0 – Changelog
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+Release 1.2.0 – Changelog
+
+   
+  Changelog
+
+
+Changelog
+
+The 1.2.0 release https://issues.apache.org/jira/issues/?jql=project+%3D+FLINK+AND+fixVersion+%3D+1.2.0;>resolved
 650 JIRA issues in total.
+
+Sub-task
+
+
+[https://issues.apache.org/jira/browse/FLINK-3202;>FLINK-3202] -   
  Make Timer/Trigger Service Scoped to Key/Namespace
+
+[https://issues.apache.org/jira/browse/FLINK-3580;>FLINK-3580] -   
  Reintroduce Date/Time and implement scalar functions for it
+
+[https://issues.apache.org/jira/browse/FLINK-3660;>FLINK-3660] -   
  Measure latency of elements and expose it as a metric
+
+[https://issues.apache.org/jira/browse/FLINK-3761;>FLINK-3761] -   
  Refactor State Backends/Make Keyed State Key-Group Aware
+
+[https://issues.apache.org/jira/browse/FLINK-3950;>FLINK-3950] -   
  Add Meter Metric Type
+
+[https://issues.apache.org/jira/browse/FLINK-4174;>FLINK-4174] -   
  Enhance Window Evictor
+
+[https://issues.apache.org/jira/browse/FLINK-4221;>FLINK-4221] -   
  Show metrics in WebFrontend
+
+[https://issues.apache.org/jira/browse/FLINK-4273;>FLINK-4273] -   
  Refactor JobClientActor to watch already submitted jobs 
+
+[https://issues.apache.org/jira/browse/FLINK-4312;>FLINK-4312] -   
  Remove Serializabiliy of ExecutionGraph
+
+[https://issues.apache.org/jira/browse/FLINK-4317;>FLINK-4317] -   
  Restructure documentation layout
+
+[https://issues.apache.org/jira/browse/FLINK-4320;>FLINK-4320] -   
  Fix misleading ScheduleMode names
+
+[https://issues.apache.org/jira/browse/FLINK-4324;>FLINK-4324] -   
  Enable Akka SSL
+
+[https://issues.apache.org/jira/browse/FLINK-4325;>FLINK-4325] -   
  Implement Web UI HTTPS
+
+[https://issues.apache.org/jira/browse/FLINK-4380;>FLINK-4380] -   
  Introduce KeyGroupAssigner and Max-Parallelism Parameter
+
+[https://issues.apache.org/jira/browse/FLINK-4381;>FLINK-4381] -   
  Refactor State to Prepare For Key-Group State Backends
+
+[https://issues.apache.org/jira/browse/FLINK-4384;>FLINK-4384] -   
  Add a scheduleRunAsync() feature to the RpcEndpoint
+
+[https://issues.apache.org/jira/browse/FLINK-4386;>FLINK-4386] -   
  Add as way to assert that code runs in the RpcEndpoints Main Thread
+
+[https://issues.apache.org/jira/browse/FLINK-4389;>FLINK-4389] -   
  Expose metrics to Webfrontend
+
+[https://issues.apache.org/jira/browse/FLINK-4392;>FLINK-4392] -   
  Make RPC Service Thread Safe
+
+[https://issues.apache.org/jira/browse/FLINK-4403;>FLINK-4403] -   
  RPC proxy classloading should use Flink 

[1/4] flink-web git commit: Add changelog for 1.2.0 release

2017-02-02 Thread rmetzger
Repository: flink-web
Updated Branches:
  refs/heads/asf-site 49439c679 -> 5b25f84e4


http://git-wip-us.apache.org/repos/asf/flink-web/blob/5b25f84e/content/q/quickstart-scala-SNAPSHOT.sh
--
diff --git a/content/q/quickstart-scala-SNAPSHOT.sh 
b/content/q/quickstart-scala-SNAPSHOT.sh
index 9cfeb2e..95a6ac2 100755
--- a/content/q/quickstart-scala-SNAPSHOT.sh
+++ b/content/q/quickstart-scala-SNAPSHOT.sh
@@ -24,7 +24,7 @@ PACKAGE=quickstart
 mvn archetype:generate 
\
   -DarchetypeGroupId=org.apache.flink  \
   -DarchetypeArtifactId=flink-quickstart-scala \
-  -DarchetypeVersion=1.2-SNAPSHOT  \
+  -DarchetypeVersion=1.3-SNAPSHOT  \
   -DgroupId=org.myorg.quickstart   \
   -DartifactId=$PACKAGE
\
   -Dversion=0.1
\

http://git-wip-us.apache.org/repos/asf/flink-web/blob/5b25f84e/content/q/quickstart-scala.sh
--
diff --git a/content/q/quickstart-scala.sh b/content/q/quickstart-scala.sh
index 7dcbe34..a561efe 100755
--- a/content/q/quickstart-scala.sh
+++ b/content/q/quickstart-scala.sh
@@ -24,7 +24,7 @@ PACKAGE=quickstart
 mvn archetype:generate 
\
   -DarchetypeGroupId=org.apache.flink  \
   -DarchetypeArtifactId=flink-quickstart-scala \
-  -DarchetypeVersion=1.1.4 
\
+  -DarchetypeVersion=1.2.0 
\
   -DgroupId=org.myorg.quickstart   \
   -DartifactId=$PACKAGE
\
   -Dversion=0.1
\



[3/4] flink-web git commit: Add changelog for 1.2.0 release

2017-02-02 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/5b25f84e/blog/release_1.2.0-changelog.md
--
diff --git a/blog/release_1.2.0-changelog.md b/blog/release_1.2.0-changelog.md
new file mode 100755
index 000..b42e0d5
--- /dev/null
+++ b/blog/release_1.2.0-changelog.md
@@ -0,0 +1,1370 @@
+---
+title: "Release 1.2.0 – Changelog"
+---
+
+* toc
+{:toc}
+
+## Changelog
+
+The 1.2.0 release [resolved 650 JIRA 
issues](https://issues.apache.org/jira/issues/?jql=project+%3D+FLINK+AND+fixVersion+%3D+1.2.0)
 in total.
+
+
+
+
+Sub-task
+
+
+[FLINK-3202] -   
  Make Timer/Trigger Service Scoped to Key/Namespace
+
+[FLINK-3580] -   
  Reintroduce Date/Time and implement scalar functions for it
+
+[FLINK-3660] -   
  Measure latency of elements and expose it as a metric
+
+[FLINK-3761] -   
  Refactor State Backends/Make Keyed State Key-Group Aware
+
+[FLINK-3950] -   
  Add Meter Metric Type
+
+[FLINK-4174] -   
  Enhance Window Evictor
+
+[FLINK-4221] -   
  Show metrics in WebFrontend
+
+[FLINK-4273] -   
  Refactor JobClientActor to watch already submitted jobs 
+
+[FLINK-4312] -   
  Remove Serializabiliy of ExecutionGraph
+
+[FLINK-4317] -   
  Restructure documentation layout
+
+[FLINK-4320] -   
  Fix misleading ScheduleMode names
+
+[FLINK-4324] -   
  Enable Akka SSL
+
+[FLINK-4325] -   
  Implement Web UI HTTPS
+
+[FLINK-4380] -   
  Introduce KeyGroupAssigner and Max-Parallelism Parameter
+
+[FLINK-4381] -   
  Refactor State to Prepare For Key-Group State Backends
+
+[FLINK-4384] -   
  Add a scheduleRunAsync() feature to the RpcEndpoint
+
+[FLINK-4386] -   
  Add as way to assert that code runs in the RpcEndpoints Main Thread
+
+[FLINK-4389] -   
  Expose metrics to Webfrontend
+
+[FLINK-4392] -   
  Make RPC Service Thread Safe
+
+[FLINK-4403] -   
  RPC proxy classloading should use Flink class classloader
+
+[FLINK-4404] -   
  Implement Data Transfer SSL
+
+[FLINK-4405] -   
  Implement Blob Server SSL
+
+[FLINK-4434] -   
  Add a testing RPC service
+
+[FLINK-4490] -   
  Decouple Slot and Instance
+
+[FLINK-4507] -   
  Deprecate savepoint backend config
+
+[FLINK-4509] -   
  Specify savepoint directory per savepoint
+
+[FLINK-4510] -   
  Always create CheckpointCoordinator
+
+[FLINK-4512] -   
  Add option for persistent checkpoints
+
+[FLINK-4670] -   
  Add watch mechanism on current RPC framework
+
+[FLINK-4685] -   
  Gather operator checkpoint durations data sizes from the runtime
+
+[FLINK-4691] -   
  Add group-windows for streaming tables   
+
+[FLINK-4697] -   
  Gather more detailed checkpoint stats in CheckpointStatsTracker
+
+[FLINK-4698] -   
  Visualize additional checkpoint information
+
+[FLINK-4748] -   
  Fix shutdown of automatic watermark context
+
+[FLINK-4749] -   
  Remove redundant processing time timers and futures in the window operator
+
+[FLINK-4750] -   
  Ensure that active processing time triggers complete before closing operators
+
+[FLINK-4768] -   
  Migrate High Availability configuration options
+
+[FLINK-4826] -   
  Add keytab based kerberos support for Mesos environment
+
+[FLINK-4877] -   
  Refactorings around FLINK-3674 (User Function Timers)
+
+[FLINK-4884] -   
  Eagerly Store MergingWindowSet in State in WindowOperator
+
+[FLINK-4899] -   
  Implement DCOS package
+
+[FLINK-4900] -   
  Implement Docker image support
+
+[FLINK-4901] -   
  Build DCOS Docker image
+
+[FLINK-4918] -   
  Add SSL support to Mesos artifact server
+
+[FLINK-4921] -   
  Upgrade to Mesos 1.0.1
+
+[FLINK-4922] -   
  Write documentation
+
+[FLINK-4937] -   
  Add incremental group window aggregation for streaming Table API
+
+[FLINK-4959] -   
  Write Documentation for ProcessFunction
+
+[FLINK-4984] -   
  Add Cancellation Barriers to BarrierTracker and BarrierBuffer
+
+[FLINK-4985] -   
  Report Declined/Canceled Checkpoints to Checkpoint Coordinator
+
+[FLINK-4987] -   
  Harden slot pool logic
+
+[FLINK-4993] -   
  Dont Allow Trigger.onMerge() to return TriggerResult
+
+[FLINK-4994] -   
  Dont Clear Trigger State and Merging Window Set When Purging
+
+[FLINK-5042] -   
  Convert old savepoints to new savepoints
+
+[FLINK-5043] -   
  Converting keyed state from Flink 1.1 backend implementations to their new 
counterparts in 1.2
+
+[FLINK-5044] -   
  Converting operator and function state from Flink 1.1 for all changed 
operators in 1.2
+
+[FLINK-5091] -   
  Formalize the AppMaster environment for docker compability
+
+[FLINK-5290] -   
  Ensure backwards compatibility of the hashes used to generate JobVertexIds
+
+[FLINK-5293] -   
  Make the Kafka consumer backwards compatible.
+
+[FLINK-5294] -   
  Make the WindowOperator backwards 

[4/4] flink-web git commit: Add changelog for 1.2.0 release

2017-02-02 Thread rmetzger
Add changelog for 1.2.0 release


Project: http://git-wip-us.apache.org/repos/asf/flink-web/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink-web/commit/5b25f84e
Tree: http://git-wip-us.apache.org/repos/asf/flink-web/tree/5b25f84e
Diff: http://git-wip-us.apache.org/repos/asf/flink-web/diff/5b25f84e

Branch: refs/heads/asf-site
Commit: 5b25f84e46253d3f2976c8db415fa48155ac57e8
Parents: 49439c6
Author: Robert Metzger 
Authored: Thu Feb 2 13:33:30 2017 +0100
Committer: Robert Metzger 
Committed: Thu Feb 2 13:33:30 2017 +0100

--
 blog/release_1.2.0-changelog.md   | 1370 +
 content/blog/release_1.2.0-changelog.html | 1545 
 content/q/quickstart-scala-SNAPSHOT.sh|2 +-
 content/q/quickstart-scala.sh |2 +-
 4 files changed, 2917 insertions(+), 2 deletions(-)
--




[1/4] flink-web git commit: Put Flink 1.2.0 on website

2017-02-06 Thread rmetzger
Repository: flink-web
Updated Branches:
  refs/heads/asf-site 5b25f84e4 -> 5c98e415d


http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/content/usecases.html
--
diff --git a/content/usecases.html b/content/usecases.html
index 0b4407b..132b2f3 100644
--- a/content/usecases.html
+++ b/content/usecases.html
@@ -90,20 +90,20 @@
 
 
 
 
   Documentation
 
 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.3; 
target="_blank">1.3 (Snapshot) 
 
   
 
 
 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
 
 
 

http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/downloads.md
--
diff --git a/downloads.md b/downloads.md
index e1f8efb..e47107e 100644
--- a/downloads.md
+++ b/downloads.md
@@ -36,12 +36,6 @@ pick the Hadoop 1 version.
 
 
 
-Hadoop® 1.2.1
-Download
-
-
-
-
 Hadoop® 2.3.0
 Download
 Download
@@ -83,7 +77,6 @@ pick the Hadoop 1 version.
 
 You can add the following dependencies to your `pom.xml` to include Apache 
Flink in your project. These dependencies include a local execution environment 
and thus support local testing.
 
-- **Hadoop 1**: If you want to interact with Hadoop 1, use `{{ 
site.FLINK_VERSION_HADOOP_1_STABLE }}` as the version.
 - **Scala API**: To use the Scala API, replace the `flink-java` artifact id 
with `flink-scala_2.10` and `flink-streaming-java_2.10` with 
`flink-streaming-scala_2.10`. For Scala 2.11 dependencies, use the suffix 
`_2.11` instead of `_2.10`.
 
 ```xml

http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/q/quickstart-SNAPSHOT.sh
--
diff --git a/q/quickstart-SNAPSHOT.sh b/q/quickstart-SNAPSHOT.sh
index 1b1f2d2..7670328 100755
--- a/q/quickstart-SNAPSHOT.sh
+++ b/q/quickstart-SNAPSHOT.sh
@@ -24,7 +24,7 @@ PACKAGE=quickstart
 mvn archetype:generate 
\
   -DarchetypeGroupId=org.apache.flink  \
   -DarchetypeArtifactId=flink-quickstart-java  \
-  -DarchetypeVersion=1.2-SNAPSHOT  \
+  -DarchetypeVersion=1.3-SNAPSHOT  \
   -DgroupId=org.myorg.quickstart   \
   -DartifactId=$PACKAGE
\
   -Dversion=0.1
\

http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/q/quickstart.sh
--
diff --git a/q/quickstart.sh b/q/quickstart.sh
index 972400a..b1322e3 100755
--- a/q/quickstart.sh
+++ b/q/quickstart.sh
@@ -24,7 +24,7 @@ PACKAGE=quickstart
 mvn archetype:generate 
\
   -DarchetypeGroupId=org.apache.flink  \
   -DarchetypeArtifactId=flink-quickstart-java  \
-  -DarchetypeVersion=1.1.3 
\
+  -DarchetypeVersion=1.2.0 
\
   -DgroupId=org.myorg.quickstart   \
   -DartifactId=$PACKAGE
\
   -Dversion=0.1
\

http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/q/sbt-quickstart.sh
--
diff --git a/q/sbt-quickstart.sh b/q/sbt-quickstart.sh
index 78abcb8..48df995 100755
--- a/q/sbt-quickstart.sh
+++ b/q/sbt-quickstart.sh
@@ -42,7 +42,7 @@ defaultProjectName="Flink Project"
 defaultOrganization="org.example"
 defaultVersion="0.1-SNAPSHOT"
 defaultScalaVersion="2.11.7"
-defaultFlinkVersion="1.1.3"
+defaultFlinkVersion="1.2.0"
 
 echo "This script creates a Flink project using Scala and SBT."
 

http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/q/stable-docs.html

[4/4] flink-web git commit: Put Flink 1.2.0 on website

2017-02-06 Thread rmetzger
Put Flink 1.2.0 on website


Project: http://git-wip-us.apache.org/repos/asf/flink-web/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink-web/commit/5c98e415
Tree: http://git-wip-us.apache.org/repos/asf/flink-web/tree/5c98e415
Diff: http://git-wip-us.apache.org/repos/asf/flink-web/diff/5c98e415

Branch: refs/heads/asf-site
Commit: 5c98e415dd22e1bb90c00a3341e67e255b51b13f
Parents: 5b25f84
Author: Robert Metzger 
Authored: Mon Feb 6 16:50:45 2017 +0100
Committer: Robert Metzger 
Committed: Mon Feb 6 16:50:45 2017 +0100

--
 _config.yml |  46 +-
 _includes/navbar.html   |   4 +-
 _posts/2017-02-06-release-1.2.0.md  | 244 +++
 content/blog/feed.xml   | 276 +++-
 content/blog/index.html |  51 +-
 content/blog/page2/index.html   |  53 +-
 content/blog/page3/index.html   |  51 +-
 content/blog/page4/index.html   |  56 +-
 content/blog/page5/index.html   | 651 +++
 .../release_1.0.0-changelog_known_issues.html   |   8 +-
 content/blog/release_1.1.0-changelog.html   |   8 +-
 content/blog/release_1.2.0-changelog.html   |   8 +-
 content/community.html  |  12 +-
 content/contribute-code.html|  16 +-
 content/contribute-documentation.html   |   8 +-
 content/documentation.html  |   8 +-
 content/downloads.html  |  47 +-
 content/ecosystem.html  |   8 +-
 content/faq.html|  28 +-
 content/features.html   |   8 +-
 content/how-to-contribute.html  |   8 +-
 content/improve-website.html|   8 +-
 content/index.html  |  16 +-
 content/introduction.html   |   8 +-
 content/material.html   |   8 +-
 content/news/2014/08/26/release-0.6.html|   8 +-
 content/news/2014/09/26/release-0.6.1.html  |   8 +-
 content/news/2014/10/03/upcoming_events.html|   8 +-
 content/news/2014/11/04/release-0.7.0.html  |   8 +-
 .../news/2014/11/18/hadoop-compatibility.html   |   8 +-
 content/news/2015/01/06/december-in-flink.html  |   8 +-
 content/news/2015/01/21/release-0.8.html|   8 +-
 content/news/2015/02/04/january-in-flink.html   |   8 +-
 content/news/2015/02/09/streaming-example.html  |   8 +-
 .../news/2015/03/02/february-2015-in-flink.html |   8 +-
 .../peeking-into-Apache-Flinks-Engine-Room.html |   8 +-
 content/news/2015/04/07/march-in-flink.html |   8 +-
 .../2015/04/13/release-0.9.0-milestone1.html|   8 +-
 .../05/11/Juggling-with-Bits-and-Bytes.html |   8 +-
 .../news/2015/05/14/Community-update-April.html |   8 +-
 .../announcing-apache-flink-0.9.0-release.html  |   8 +-
 .../2015/08/24/introducing-flink-gelly.html |   8 +-
 content/news/2015/09/01/release-0.9.1.html  |  10 +-
 content/news/2015/09/03/flink-forward.html  |   8 +-
 content/news/2015/09/16/off-heap-memory.html|   8 +-
 content/news/2015/11/16/release-0.10.0.html |   8 +-
 content/news/2015/11/27/release-0.10.1.html |   8 +-
 .../news/2015/12/04/Introducing-windows.html|   8 +-
 .../news/2015/12/11/storm-compatibility.html|   8 +-
 content/news/2015/12/18/a-year-in-review.html   |   8 +-
 content/news/2016/02/11/release-0.10.2.html |   8 +-
 content/news/2016/03/08/release-1.0.0.html  |   8 +-
 content/news/2016/04/06/cep-monitoring.html |   8 +-
 content/news/2016/04/06/release-1.0.1.html  |   8 +-
 .../news/2016/04/14/flink-forward-announce.html |   8 +-
 content/news/2016/04/22/release-1.0.2.html  |   8 +-
 content/news/2016/05/11/release-1.0.3.html  |   8 +-
 content/news/2016/05/24/stream-sql.html |   8 +-
 content/news/2016/08/08/release-1.1.0.html  |   8 +-
 content/news/2016/08/11/release-1.1.1.html  |   8 +-
 .../news/2016/08/24/ff16-keynotes-panels.html   |   8 +-
 content/news/2016/09/05/release-1.1.2.html  |   8 +-
 content/news/2016/10/12/release-1.1.3.html  |   8 +-
 .../news/2016/12/19/2016-year-in-review.html|   8 +-
 content/news/2016/12/21/release-1.1.4.html  |   8 +-
 content/news/2017/02/06/release-1.2.0.html  | 466 +
 content/poweredby.html  |   8 +-
 content/privacy-policy.html |   8 +-
 content/project.html|   8 +-
 content/q/quickstart-SNAPSHOT.sh|   2 +-
 content/q/quickstart.sh |   2 +-
 content/q/sbt-quickstart.sh |   2 +-
 content/q/stable-docs.html  |   2 +-
 content/slides.html |  10 +-
 content/usecases.html  

[2/4] flink-web git commit: Put Flink 1.2.0 on website

2017-02-06 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/content/news/2015/09/03/flink-forward.html
--
diff --git a/content/news/2015/09/03/flink-forward.html 
b/content/news/2015/09/03/flink-forward.html
index 37e7c8c..83c8cf2 100644
--- a/content/news/2015/09/03/flink-forward.html
+++ b/content/news/2015/09/03/flink-forward.html
@@ -90,20 +90,20 @@
 
 
 
 
   Documentation
 
 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.3; 
target="_blank">1.3 (Snapshot) 
 
   
 
 
 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
 
 
 

http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/content/news/2015/09/16/off-heap-memory.html
--
diff --git a/content/news/2015/09/16/off-heap-memory.html 
b/content/news/2015/09/16/off-heap-memory.html
index 60d7bda..7d72a7d 100644
--- a/content/news/2015/09/16/off-heap-memory.html
+++ b/content/news/2015/09/16/off-heap-memory.html
@@ -90,20 +90,20 @@
 
 
 
 
   Documentation
 
 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.3; 
target="_blank">1.3 (Snapshot) 
 
   
 
 
 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
 
 
 

http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/content/news/2015/11/16/release-0.10.0.html
--
diff --git a/content/news/2015/11/16/release-0.10.0.html 
b/content/news/2015/11/16/release-0.10.0.html
index 93bad1f..cbd11a1 100644
--- a/content/news/2015/11/16/release-0.10.0.html
+++ b/content/news/2015/11/16/release-0.10.0.html
@@ -90,20 +90,20 @@
 
 
 
 
   Documentation
 
 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.3; 
target="_blank">1.3 (Snapshot) 
 
   
 
 
 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
 
 
 

http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/content/news/2015/11/27/release-0.10.1.html
--
diff --git a/content/news/2015/11/27/release-0.10.1.html 
b/content/news/2015/11/27/release-0.10.1.html
index 01e6b7f..d4ef924 100644
--- a/content/news/2015/11/27/release-0.10.1.html
+++ b/content/news/2015/11/27/release-0.10.1.html
@@ -90,20 +90,20 @@
 
 
 
 
   Documentation
 
 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+  

flink git commit: [FLINK-5450] Fix restore from legacy log message

2017-01-23 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/release-1.2 085288e7e -> f4cd7d0ec


[FLINK-5450] Fix restore from legacy log message

This closes #3134


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/f4cd7d0e
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/f4cd7d0e
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/f4cd7d0e

Branch: refs/heads/release-1.2
Commit: f4cd7d0ec3dd70e51e1db27c8336481838d50e63
Parents: 085288e
Author: kl0u 
Authored: Mon Jan 16 15:27:38 2017 +0100
Committer: Robert Metzger 
Committed: Mon Jan 23 14:32:30 2017 +0100

--
 .../operators/windowing/WindowOperator.java | 27 +---
 1 file changed, 18 insertions(+), 9 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/f4cd7d0e/flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/operators/windowing/WindowOperator.java
--
diff --git 
a/flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/operators/windowing/WindowOperator.java
 
b/flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/operators/windowing/WindowOperator.java
index 628d663..5ed5a4e 100644
--- 
a/flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/operators/windowing/WindowOperator.java
+++ 
b/flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/operators/windowing/WindowOperator.java
@@ -804,9 +804,6 @@ public class WindowOperator
 
public void registerRestoredLegacyStateState() throws Exception {
 
-   LOG.info("{} (taskIdx={}) re-registering state from an older 
Flink version.",
-   getClass().getSimpleName(), legacyWindowOperatorType, 
getRuntimeContext().getIndexOfThisSubtask());
-
switch (legacyWindowOperatorType) {
case NONE:
reregisterStateFromLegacyWindowOperator();
@@ -983,14 +980,22 @@ public class WindowOperator
// if we restore from an older version,
// we have to re-register the recovered state.
 
-   if (restoredFromLegacyEventTimeTimers != null) {
+   if (restoredFromLegacyEventTimeTimers != null && 
!restoredFromLegacyEventTimeTimers.isEmpty()) {
+
+   LOG.info("{} (taskIdx={}) re-registering event-time 
timers from an older Flink version.",
+   getClass().getSimpleName(), 
getRuntimeContext().getIndexOfThisSubtask());
+
for (Timer timer : 
restoredFromLegacyEventTimeTimers) {
setCurrentKey(timer.key);

internalTimerService.registerEventTimeTimer(timer.window, timer.timestamp);
}
}
 
-   if (restoredFromLegacyProcessingTimeTimers != null) {
+   if (restoredFromLegacyProcessingTimeTimers != null && 
!restoredFromLegacyProcessingTimeTimers.isEmpty()) {
+
+   LOG.info("{} (taskIdx={}) re-registering 
processing-time timers from an older Flink version.",
+   getClass().getSimpleName(), 
getRuntimeContext().getIndexOfThisSubtask());
+
for (Timer timer : 
restoredFromLegacyProcessingTimeTimers) {
setCurrentKey(timer.key);

internalTimerService.registerProcessingTimeTimer(timer.window, timer.timestamp);
@@ -1003,16 +1008,20 @@ public class WindowOperator
}
 
public void reregisterStateFromLegacyAlignedWindowOperator() throws 
Exception {
-   if (restoredFromLegacyAlignedOpRecords != null) {
+   if (restoredFromLegacyAlignedOpRecords != null && 
!restoredFromLegacyAlignedOpRecords.isEmpty()) {
+
+   LOG.info("{} (taskIdx={}) re-registering timers from 
legacy {} from an older Flink version.",
+   getClass().getSimpleName(), 
getRuntimeContext().getIndexOfThisSubtask(), legacyWindowOperatorType);
+
while (!restoredFromLegacyAlignedOpRecords.isEmpty()) {
StreamRecord record = 
restoredFromLegacyAlignedOpRecords.poll();

setCurrentKey(keySelector.getKey(record.getValue()));
processElement(record);
}
-
-   // gc friendliness
-   restoredFromLegacyAlignedOpRecords = null;
}
+
+   // gc friendliness
+   

[1/2] flink git commit: [FLINK-5613][query] querying a non-existing key is inconsistent among state backends

2017-01-23 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/release-1.2 b323f66a6 -> 44f7098d7


[FLINK-5613][query] querying a non-existing key is inconsistent among state 
backends

Querying for a non-existing key for a state that has a default value set
currently results in an UnknownKeyOrNamespace exception when the
MemoryStateBackend or FsStateBackend is used but results in the default value
if RocksDBStateBackend is set.

This removes the special handling from the RocksDBStateBackend and makes it
consistent with the other two back-ends, i.e. returning null which results
in the mentioned UnknownKeyOrNamespace exception.


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/44f7098d
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/44f7098d
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/44f7098d

Branch: refs/heads/release-1.2
Commit: 44f7098d74476ab561f38c8ced5f3a0b5f9bd0c2
Parents: 2e175fb
Author: Nico Kruber 
Authored: Tue Jan 17 14:26:16 2017 +0100
Committer: Robert Metzger 
Committed: Tue Jan 24 08:44:08 2017 +0100

--
 .../streaming/state/RocksDBValueState.java  |  12 --
 .../flink/test/query/QueryableStateITCase.java  | 160 ++-
 2 files changed, 153 insertions(+), 19 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/44f7098d/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java
--
diff --git 
a/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java
 
b/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java
index 9563ed8..d8d89ba 100644
--- 
a/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java
+++ 
b/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java
@@ -23,7 +23,6 @@ import org.apache.flink.api.common.state.ValueStateDescriptor;
 import org.apache.flink.api.common.typeutils.TypeSerializer;
 import org.apache.flink.core.memory.DataInputViewStreamWrapper;
 import org.apache.flink.core.memory.DataOutputViewStreamWrapper;
-import org.apache.flink.runtime.query.netty.message.KvStateRequestSerializer;
 import org.rocksdb.ColumnFamilyHandle;
 import org.rocksdb.RocksDBException;
 import org.rocksdb.WriteOptions;
@@ -102,15 +101,4 @@ public class RocksDBValueState
throw new RuntimeException("Error while adding data to 
RocksDB", e);
}
}
-
-   @Override
-   public byte[] getSerializedValue(byte[] serializedKeyAndNamespace) 
throws Exception {
-   byte[] value = 
super.getSerializedValue(serializedKeyAndNamespace);
-
-   if (value != null) {
-   return value;
-   } else {
-   return 
KvStateRequestSerializer.serializeValue(stateDesc.getDefaultValue(), 
stateDesc.getSerializer());
-   }
-   }
 }

http://git-wip-us.apache.org/repos/asf/flink/blob/44f7098d/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java
--
diff --git 
a/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java
 
b/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java
index 327a715..c2df6ae 100644
--- 
a/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java
+++ 
b/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java
@@ -39,6 +39,7 @@ import org.apache.flink.api.java.tuple.Tuple2;
 import org.apache.flink.configuration.ConfigConstants;
 import org.apache.flink.configuration.Configuration;
 import org.apache.flink.configuration.QueryableStateOptions;
+import org.apache.flink.contrib.streaming.state.RocksDBStateBackend;
 import org.apache.flink.runtime.akka.AkkaUtils;
 import org.apache.flink.runtime.executiongraph.ExecutionAttemptID;
 import org.apache.flink.runtime.executiongraph.ExecutionGraph;
@@ -50,10 +51,14 @@ import 
org.apache.flink.runtime.messages.JobManagerMessages.JobFound;
 import org.apache.flink.runtime.query.KvStateLocation;
 import org.apache.flink.runtime.query.KvStateMessage;
 import org.apache.flink.runtime.query.QueryableStateClient;
+import org.apache.flink.runtime.query.netty.UnknownKeyOrNamespace;
 import org.apache.flink.runtime.query.netty.message.KvStateRequestSerializer;
+import org.apache.flink.runtime.state.AbstractStateBackend;
 import org.apache.flink.runtime.state.CheckpointListener;
 

[2/2] flink git commit: [FLINK-5395] [Build System] support locally build distribution by script create_release_files.sh

2017-01-23 Thread rmetzger
[FLINK-5395] [Build System] support locally build distribution by script 
create_release_files.sh

This closes #3049


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/2e175fb3
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/2e175fb3
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/2e175fb3

Branch: refs/heads/release-1.2
Commit: 2e175fb301cc84ca0c3195d1fe325319b34fd57a
Parents: b323f66
Author: shijinkui 
Authored: Wed Jan 11 17:42:16 2017 +0800
Committer: Robert Metzger 
Committed: Tue Jan 24 08:44:08 2017 +0100

--
 .gitignore|   2 +
 tools/create_release_files.sh | 156 +
 2 files changed, 124 insertions(+), 34 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/2e175fb3/.gitignore
--
diff --git a/.gitignore b/.gitignore
index 9012d0a..3c9e4e8 100644
--- a/.gitignore
+++ b/.gitignore
@@ -32,3 +32,5 @@ out/
 /docs/.jekyll-metadata
 *.ipr
 *.iws
+tools/flink
+tools/flink-*

http://git-wip-us.apache.org/repos/asf/flink/blob/2e175fb3/tools/create_release_files.sh
--
diff --git a/tools/create_release_files.sh b/tools/create_release_files.sh
index fdf50a5..ed01d97 100755
--- a/tools/create_release_files.sh
+++ b/tools/create_release_files.sh
@@ -66,16 +66,20 @@ fi
 GPG_PASSPHRASE=${GPG_PASSPHRASE:-XXX}
 GPG_KEY=${GPG_KEY:-XXX}
 GIT_AUTHOR=${GIT_AUTHOR:-"Your name "}
-OLD_VERSION=${OLD_VERSION:-1.1-SNAPSHOT}
-RELEASE_VERSION=${NEW_VERSION}
-RELEASE_CANDIDATE=${RELEASE_CANDIDATE:-rc1}
+OLD_VERSION=${OLD_VERSION:-1.2-SNAPSHOT}
+RELEASE_VERSION=${NEW_VERSION:-1.3-SNAPSHOT}
+RELEASE_CANDIDATE=${RELEASE_CANDIDATE:-none}
 RELEASE_BRANCH=${RELEASE_BRANCH:-master}
 USER_NAME=${USER_NAME:-yourapacheidhere}
 MVN=${MVN:-mvn}
 GPG=${GPG:-gpg}
 sonatype_user=${sonatype_user:-yourapacheidhere}
 sonatype_pw=${sonatype_pw:-XXX}
-
+# whether only build the dist local and don't release to apache
+IS_LOCAL_DIST=${IS_LOCAL_DIST:-false}
+GIT_REPO=${GIT_REPO:-git-wip-us.apache.org/repos/asf/flink.git}
+SCALA_VERSION=none
+HADOOP_VERSION=none
 
 if [ "$(uname)" == "Darwin" ]; then
 SHASUM="shasum -a 512"
@@ -85,18 +89,81 @@ else
 MD5SUM="md5sum"
 fi
 
+usage() {
+  set +x
+  echo "./create_release_files.sh --scala-version 2.11 --hadoop-version 2.7.2"
+  echo ""
+  echo "usage:"
+  echo "[--scala-version ] [--hadoop-version ]"
+  echo ""
+  echo "example 1: build apache release"
+  echo "  sonatype_user=APACHEID sonatype_pw=APACHEIDPASSWORD \ "
+  echo "  NEW_VERSION=1.2.0 RELEASE_CANDIDATE="rc1" 
RELEASE_BRANCH=release-1.2.0 OLD_VERSION=1.1-SNAPSHOT \ "
+  echo "  USER_NAME=APACHEID GPG_PASSPHRASE=XXX GPG_KEY=KEYID \ "
+  echo "  GIT_AUTHOR=\"`git config --get user.name` <`git config --get 
user.email`>\" \ "
+  echo "  GIT_REPO=github.com/apache/flink.git \ "
+  echo "  ./create_release_files.sh --scala-version 2.11 --hadoop-version 
2.7.2"
+  echo ""
+  echo "example 2: build local release"
+  echo "  NEW_VERSION=1.2.0 RELEASE_BRANCH=master OLD_VERSION=1.2-SNAPSHOT \ "
+  echo "  GPG_PASSPHRASE=XXX GPG_KEY=XXX IS_LOCAL_DIST=true \ "
+  echo "  ./create_release_files.sh --scala-version 2.11 --hadoop-version 
2.7.2"
+
+  exit 1
+}
+
+# Parse arguments
+while (( "$#" )); do
+  case $1 in
+--scala-version)
+  SCALA_VERSION="$2"
+  shift
+  ;;
+--hadoop-version)
+  HADOOP_VERSION="$2"
+  shift
+  ;;
+--help)
+  usage
+  ;;
+*)
+  break
+  ;;
+  esac
+  shift
+done
+
+###
 
 prepare() {
   # prepare
-  git clone http://git-wip-us.apache.org/repos/asf/flink.git flink
+  target_branch=release-$RELEASE_VERSION
+  if [ "$RELEASE_CANDIDATE" != "none" ]; then
+target_branch=$target_branch-$RELEASE_CANDIDATE
+  fi
+
+  if [ ! -d ./flink ]; then
+git clone http://$GIT_REPO flink
+  else
+# if flink git repo exist, delete target branch, delete builded 
distribution
+rm -rf flink-*.tgz
+cd flink
+# try-catch
+{
+  git pull --all
+  git checkout master
+  git branch -D $target_branch -f
+} || {
+  echo "branch $target_branch not found"
+}
+cd ..
+  fi
+
   cd flink
-  git checkout -b "release-$RELEASE_VERSION-$RELEASE_CANDIDATE" 
origin/$RELEASE_BRANCH
-  rm -f .gitignore
-  rm -f .gitattributes
-  rm -f .travis.yml
-  rm -f deploysettings.xml
-  rm -f CHANGELOG
-  rm -rf .github
+
+  git checkout -b $target_branch origin/$RELEASE_BRANCH
+  rm -rf .gitignore .gitattributes .travis.yml deploysettings.xml CHANGELOG 
.github
+
   cd ..
 }
 
@@ -117,10 +184,13 @@ make_source_release() {
   perl -pi -e "s#^version_short: 

flink git commit: [FLINK-5395] [Build System] support locally build distribution by script create_release_files.sh

2017-01-24 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master 00d1ad86a -> 6b55e2ca3


[FLINK-5395] [Build System] support locally build distribution by script 
create_release_files.sh

This closes #3049


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/6b55e2ca
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/6b55e2ca
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/6b55e2ca

Branch: refs/heads/master
Commit: 6b55e2ca30c5ecbf9998e44b2638a7f2d94a5963
Parents: 00d1ad8
Author: shijinkui 
Authored: Wed Jan 11 17:42:16 2017 +0800
Committer: Robert Metzger 
Committed: Tue Jan 24 09:11:56 2017 +0100

--
 .gitignore|   2 +
 tools/create_release_files.sh | 156 +
 2 files changed, 124 insertions(+), 34 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/6b55e2ca/.gitignore
--
diff --git a/.gitignore b/.gitignore
index 9012d0a..3c9e4e8 100644
--- a/.gitignore
+++ b/.gitignore
@@ -32,3 +32,5 @@ out/
 /docs/.jekyll-metadata
 *.ipr
 *.iws
+tools/flink
+tools/flink-*

http://git-wip-us.apache.org/repos/asf/flink/blob/6b55e2ca/tools/create_release_files.sh
--
diff --git a/tools/create_release_files.sh b/tools/create_release_files.sh
index fdf50a5..ed01d97 100755
--- a/tools/create_release_files.sh
+++ b/tools/create_release_files.sh
@@ -66,16 +66,20 @@ fi
 GPG_PASSPHRASE=${GPG_PASSPHRASE:-XXX}
 GPG_KEY=${GPG_KEY:-XXX}
 GIT_AUTHOR=${GIT_AUTHOR:-"Your name "}
-OLD_VERSION=${OLD_VERSION:-1.1-SNAPSHOT}
-RELEASE_VERSION=${NEW_VERSION}
-RELEASE_CANDIDATE=${RELEASE_CANDIDATE:-rc1}
+OLD_VERSION=${OLD_VERSION:-1.2-SNAPSHOT}
+RELEASE_VERSION=${NEW_VERSION:-1.3-SNAPSHOT}
+RELEASE_CANDIDATE=${RELEASE_CANDIDATE:-none}
 RELEASE_BRANCH=${RELEASE_BRANCH:-master}
 USER_NAME=${USER_NAME:-yourapacheidhere}
 MVN=${MVN:-mvn}
 GPG=${GPG:-gpg}
 sonatype_user=${sonatype_user:-yourapacheidhere}
 sonatype_pw=${sonatype_pw:-XXX}
-
+# whether only build the dist local and don't release to apache
+IS_LOCAL_DIST=${IS_LOCAL_DIST:-false}
+GIT_REPO=${GIT_REPO:-git-wip-us.apache.org/repos/asf/flink.git}
+SCALA_VERSION=none
+HADOOP_VERSION=none
 
 if [ "$(uname)" == "Darwin" ]; then
 SHASUM="shasum -a 512"
@@ -85,18 +89,81 @@ else
 MD5SUM="md5sum"
 fi
 
+usage() {
+  set +x
+  echo "./create_release_files.sh --scala-version 2.11 --hadoop-version 2.7.2"
+  echo ""
+  echo "usage:"
+  echo "[--scala-version ] [--hadoop-version ]"
+  echo ""
+  echo "example 1: build apache release"
+  echo "  sonatype_user=APACHEID sonatype_pw=APACHEIDPASSWORD \ "
+  echo "  NEW_VERSION=1.2.0 RELEASE_CANDIDATE="rc1" 
RELEASE_BRANCH=release-1.2.0 OLD_VERSION=1.1-SNAPSHOT \ "
+  echo "  USER_NAME=APACHEID GPG_PASSPHRASE=XXX GPG_KEY=KEYID \ "
+  echo "  GIT_AUTHOR=\"`git config --get user.name` <`git config --get 
user.email`>\" \ "
+  echo "  GIT_REPO=github.com/apache/flink.git \ "
+  echo "  ./create_release_files.sh --scala-version 2.11 --hadoop-version 
2.7.2"
+  echo ""
+  echo "example 2: build local release"
+  echo "  NEW_VERSION=1.2.0 RELEASE_BRANCH=master OLD_VERSION=1.2-SNAPSHOT \ "
+  echo "  GPG_PASSPHRASE=XXX GPG_KEY=XXX IS_LOCAL_DIST=true \ "
+  echo "  ./create_release_files.sh --scala-version 2.11 --hadoop-version 
2.7.2"
+
+  exit 1
+}
+
+# Parse arguments
+while (( "$#" )); do
+  case $1 in
+--scala-version)
+  SCALA_VERSION="$2"
+  shift
+  ;;
+--hadoop-version)
+  HADOOP_VERSION="$2"
+  shift
+  ;;
+--help)
+  usage
+  ;;
+*)
+  break
+  ;;
+  esac
+  shift
+done
+
+###
 
 prepare() {
   # prepare
-  git clone http://git-wip-us.apache.org/repos/asf/flink.git flink
+  target_branch=release-$RELEASE_VERSION
+  if [ "$RELEASE_CANDIDATE" != "none" ]; then
+target_branch=$target_branch-$RELEASE_CANDIDATE
+  fi
+
+  if [ ! -d ./flink ]; then
+git clone http://$GIT_REPO flink
+  else
+# if flink git repo exist, delete target branch, delete builded 
distribution
+rm -rf flink-*.tgz
+cd flink
+# try-catch
+{
+  git pull --all
+  git checkout master
+  git branch -D $target_branch -f
+} || {
+  echo "branch $target_branch not found"
+}
+cd ..
+  fi
+
   cd flink
-  git checkout -b "release-$RELEASE_VERSION-$RELEASE_CANDIDATE" 
origin/$RELEASE_BRANCH
-  rm -f .gitignore
-  rm -f .gitattributes
-  rm -f .travis.yml
-  rm -f deploysettings.xml
-  rm -f CHANGELOG
-  rm -rf .github
+
+  git checkout -b $target_branch origin/$RELEASE_BRANCH
+  rm -rf .gitignore .gitattributes .travis.yml deploysettings.xml CHANGELOG 
.github
+
   cd ..
 }
 
@@ -117,10 

[1/2] flink git commit: [FLINK-5473] Limit max parallelism to 1 for non-parallel operators

2017-01-24 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/release-1.2 908376ba9 -> 993a2e2fa


http://git-wip-us.apache.org/repos/asf/flink/blob/993a2e2f/flink-streaming-java/src/test/java/org/apache/flink/streaming/api/StreamExecutionEnvironmentTest.java
--
diff --git 
a/flink-streaming-java/src/test/java/org/apache/flink/streaming/api/StreamExecutionEnvironmentTest.java
 
b/flink-streaming-java/src/test/java/org/apache/flink/streaming/api/StreamExecutionEnvironmentTest.java
index 3fb4513..3fc1344 100644
--- 
a/flink-streaming-java/src/test/java/org/apache/flink/streaming/api/StreamExecutionEnvironmentTest.java
+++ 
b/flink-streaming-java/src/test/java/org/apache/flink/streaming/api/StreamExecutionEnvironmentTest.java
@@ -17,19 +17,12 @@
 
 package org.apache.flink.streaming.api;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
-
-import java.util.Arrays;
-import java.util.Iterator;
-import java.util.List;
-import java.util.NoSuchElementException;
-
+import org.apache.flink.api.common.functions.FlatMapFunction;
 import org.apache.flink.api.common.typeinfo.BasicTypeInfo;
 import org.apache.flink.api.common.typeinfo.TypeInformation;
 import org.apache.flink.streaming.api.datastream.DataStream;
 import org.apache.flink.streaming.api.datastream.DataStreamSource;
+import org.apache.flink.streaming.api.datastream.SingleOutputStreamOperator;
 import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
 import org.apache.flink.streaming.api.functions.sink.DiscardingSink;
 import org.apache.flink.streaming.api.functions.source.FromElementsFunction;
@@ -38,10 +31,20 @@ import 
org.apache.flink.streaming.api.functions.source.StatefulSequenceSource;
 import org.apache.flink.streaming.api.graph.StreamGraph;
 import org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator;
 import org.apache.flink.streaming.api.operators.StreamOperator;
+import org.apache.flink.util.Collector;
 import org.apache.flink.util.SplittableIterator;
-
+import org.junit.Assert;
 import org.junit.Test;
 
+import java.util.Arrays;
+import java.util.Iterator;
+import java.util.List;
+import java.util.NoSuchElementException;
+
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertTrue;
+import static org.junit.Assert.fail;
+
 public class StreamExecutionEnvironmentTest {
 
@Test
@@ -124,6 +127,102 @@ public class StreamExecutionEnvironmentTest {
assertTrue(getFunctionFromDataSource(src4) instanceof 
FromElementsFunction);
}
 
+   @Test
+   public void testParallelismBounds() {
+   StreamExecutionEnvironment env = 
StreamExecutionEnvironment.getExecutionEnvironment();
+
+   SourceFunction srcFun = new SourceFunction() {
+   private static final long serialVersionUID = 1L;
+
+   @Override
+   public void run(SourceContext ctx) throws 
Exception {
+   }
+
+   @Override
+   public void cancel() {
+   }
+   };
+
+
+   SingleOutputStreamOperator operator =
+   env.addSource(srcFun).flatMap(new 
FlatMapFunction() {
+
+   private static final long serialVersionUID = 1L;
+
+   @Override
+   public void flatMap(Integer value, Collector 
out) throws Exception {
+
+   }
+   });
+
+   // default value for max parallelism
+   Assert.assertEquals(-1, 
operator.getTransformation().getMaxParallelism());
+
+   // bounds for parallelism 1
+   try {
+   operator.setParallelism(0);
+   Assert.fail();
+   } catch (IllegalArgumentException expected) {
+   }
+
+   // bounds for parallelism 2
+   operator.setParallelism(1);
+   Assert.assertEquals(1, operator.getParallelism());
+
+   // bounds for parallelism 3
+   operator.setParallelism(1 << 15);
+   Assert.assertEquals(1 << 15, operator.getParallelism());
+
+   // default value after generating
+   env.getStreamGraph().getJobGraph();
+   Assert.assertEquals(-1, 
operator.getTransformation().getMaxParallelism());
+
+   // configured value after generating
+   env.setMaxParallelism(42);
+   env.getStreamGraph().getJobGraph();
+   Assert.assertEquals(42, 
operator.getTransformation().getMaxParallelism());
+
+   // bounds configured parallelism 1
+   try {
+   env.setMaxParallelism(0);
+   Assert.fail();
+   } catch 

[2/2] flink git commit: [FLINK-5473] Limit max parallelism to 1 for non-parallel operators

2017-01-24 Thread rmetzger
[FLINK-5473] Limit max parallelism to 1 for non-parallel operators

[FLINK-5473] Better default behaviours for unspecified maximum parallelism

This closes #3182.


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/993a2e2f
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/993a2e2f
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/993a2e2f

Branch: refs/heads/release-1.2
Commit: 993a2e2fa0ceecff0979a267ace7cd7b8e05d359
Parents: 908376b
Author: Stefan Richter 
Authored: Mon Jan 16 14:31:22 2017 +0100
Committer: Till Rohrmann 
Committed: Tue Jan 24 09:49:52 2017 +0100

--
 .../checkpoint/CheckpointCoordinator.java   |   4 +-
 .../checkpoint/StateAssignmentOperation.java| 285 ++-
 .../checkpoint/savepoint/SavepointLoader.java   |   8 +-
 .../ResultPartitionDeploymentDescriptor.java|  18 +-
 .../runtime/executiongraph/ExecutionGraph.java  |   8 +-
 .../executiongraph/ExecutionJobVertex.java  | 115 ++--
 .../runtime/executiongraph/ExecutionVertex.java |  31 +-
 .../api/writer/ResultPartitionWriter.java   |   4 +
 .../io/network/partition/ResultPartition.java   |   8 +
 .../flink/runtime/jobgraph/JobVertex.java   |   7 +-
 .../runtime/state/KeyGroupRangeAssignment.java  |  42 ++-
 .../apache/flink/runtime/taskmanager/Task.java  |   1 +
 .../checkpoint/CheckpointCoordinatorTest.java   |   1 +
 .../savepoint/SavepointLoaderTest.java  |   1 +
 ...ResultPartitionDeploymentDescriptorTest.java |   1 +
 .../executiongraph/ExecutionJobVertexTest.java  | 140 +
 .../ExecutionVertexDeploymentTest.java  |  27 +-
 .../network/partition/ResultPartitionTest.java  |   1 +
 .../consumer/LocalInputChannelTest.java |   1 +
 .../runtime/jobmanager/JobManagerTest.java  |   2 +
 .../runtime/taskmanager/TaskManagerTest.java|   5 +-
 .../streaming/api/datastream/KeyedStream.java   |  13 +-
 .../datastream/SingleOutputStreamOperator.java  |  29 +-
 .../environment/StreamExecutionEnvironment.java |  12 +-
 .../flink/streaming/api/graph/StreamGraph.java  |  37 +--
 .../api/graph/StreamGraphGenerator.java |  26 +-
 .../flink/streaming/api/graph/StreamNode.java   |   2 -
 .../api/graph/StreamingJobGraphGenerator.java   |  13 +-
 .../transformations/StreamTransformation.java   |   5 +
 .../partitioner/KeyGroupStreamPartitioner.java  |   1 +
 .../streaming/runtime/tasks/OperatorChain.java  |  10 +
 .../api/StreamExecutionEnvironmentTest.java | 119 +++-
 .../api/graph/StreamGraphGeneratorTest.java |   9 +-
 .../test/checkpointing/RescalingITCase.java |  25 +-
 34 files changed, 723 insertions(+), 288 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/993a2e2f/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CheckpointCoordinator.java
--
diff --git 
a/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CheckpointCoordinator.java
 
b/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CheckpointCoordinator.java
index 9132897..78cad91 100644
--- 
a/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CheckpointCoordinator.java
+++ 
b/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CheckpointCoordinator.java
@@ -885,8 +885,10 @@ public class CheckpointCoordinator {
 
LOG.info("Restoring from latest valid checkpoint: {}.", 
latest);
 
+   final Map taskStates = 
latest.getTaskStates();
+
StateAssignmentOperation stateAssignmentOperation =
-   new StateAssignmentOperation(LOG, 
tasks, latest, allowNonRestoredState);
+   new StateAssignmentOperation(LOG, 
tasks, taskStates, allowNonRestoredState);
 
stateAssignmentOperation.assignStates();
 

http://git-wip-us.apache.org/repos/asf/flink/blob/993a2e2f/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperation.java
--
diff --git 
a/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperation.java
 
b/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperation.java
index 6c23f02..6d075db 100644
--- 
a/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperation.java
+++ 
b/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperation.java
@@ -45,29 +45,34 @@ public class StateAssignmentOperation {
 
private final Logger logger;
private final Map 

[2/2] flink git commit: [FLINK-5617] Change reference version for API stability checks to 1.1.4

2017-01-24 Thread rmetzger
[FLINK-5617] Change reference version for API stability checks to 1.1.4


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/20c76a98
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/20c76a98
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/20c76a98

Branch: refs/heads/master
Commit: 20c76a98b0b52d3fc164fbc0b8808a93cb50804f
Parents: f0d96e3
Author: Robert Metzger 
Authored: Mon Jan 23 14:31:27 2017 +0100
Committer: Robert Metzger 
Committed: Tue Jan 24 12:01:35 2017 +0100

--
 flink-core/pom.xml   |  9 +
 flink-streaming-scala/pom.xml|  2 +-
 pom.xml  |  2 +-
 tools/collect_japicmp_reports.sh | 31 +++
 4 files changed, 42 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/20c76a98/flink-core/pom.xml
--
diff --git a/flink-core/pom.xml b/flink-core/pom.xml
index 3debc72..e9738a2 100644
--- a/flink-core/pom.xml
+++ b/flink-core/pom.xml
@@ -156,6 +156,15 @@ under the License.

org.apache.flink.api.common.ExecutionConfig#CONFIG_KEY

org.apache.flink.core.fs.FileSystem$FSKey

org.apache.flink.api.java.typeutils.WritableTypeInfo
+   
+   
org.apache.flink.configuration.ConfigConstants#DEFAULT_JOB_MANAGER_MAX_ATTEMPTS_HISTORY_SIZE
+   
org.apache.flink.configuration.ConfigConstants#DEFAULT_NETWORK_REQUEST_BACKOFF_INITIAL
+   
org.apache.flink.configuration.ConfigConstants#DEFAULT_NETWORK_REQUEST_BACKOFF_MAX
+   
org.apache.flink.configuration.ConfigConstants#DEFAULT_TASK_CANCELLATION_TIMEOUT_MILLIS
+   
org.apache.flink.configuration.ConfigConstants#ENABLE_QUARANTINE_MONITOR
+   
org.apache.flink.configuration.ConfigConstants#NETWORK_REQUEST_BACKOFF_INITIAL_KEY
+   
org.apache.flink.configuration.ConfigConstants#NETWORK_REQUEST_BACKOFF_MAX_KEY




http://git-wip-us.apache.org/repos/asf/flink/blob/20c76a98/flink-streaming-scala/pom.xml
--
diff --git a/flink-streaming-scala/pom.xml b/flink-streaming-scala/pom.xml
index d73ed38..f95ad79 100644
--- a/flink-streaming-scala/pom.xml
+++ b/flink-streaming-scala/pom.xml
@@ -237,7 +237,7 @@ under the License.
japicmp-maven-plugin


-   
+   


*\$\$anon\$*


http://git-wip-us.apache.org/repos/asf/flink/blob/20c76a98/pom.xml
--
diff --git a/pom.xml b/pom.xml
index 389af84..2279b0b 100644
--- a/pom.xml
+++ b/pom.xml
@@ -1242,7 +1242,7 @@ under the License.


org.apache.flink

${project.artifactId}
-   
1.0.0
+   
1.1.4

${project.packaging}



http://git-wip-us.apache.org/repos/asf/flink/blob/20c76a98/tools/collect_japicmp_reports.sh
--
diff --git a/tools/collect_japicmp_reports.sh b/tools/collect_japicmp_reports.sh
new file mode 100755
index 000..6ff1238
--- /dev/null
+++ b/tools/collect_japicmp_reports.sh
@@ -0,0 +1,31 @@
+#!/bin/sh
+
+#
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license agreements.  See the NOTICE file distributed with
+# this work for additional information regarding copyright 

[1/2] flink git commit: [FLINK-5617] Mark some methods as PublicEvolving or Internal

2017-01-24 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master f0d96e30d -> 19ef82cf3


[FLINK-5617] Mark some methods as PublicEvolving or Internal

This closes #3195


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/19ef82cf
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/19ef82cf
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/19ef82cf

Branch: refs/heads/master
Commit: 19ef82cf358da54bf204e678bf203f25147b11f9
Parents: 20c76a9
Author: Robert Metzger 
Authored: Mon Jan 23 16:28:48 2017 +0100
Committer: Robert Metzger 
Committed: Tue Jan 24 12:01:35 2017 +0100

--
 .../main/java/org/apache/flink/api/common/ExecutionConfig.java  | 2 ++
 .../src/main/java/org/apache/flink/core/fs/FileSystem.java  | 4 
 .../flink/streaming/api/datastream/AllWindowedStream.java   | 5 -
 .../apache/flink/streaming/api/datastream/WindowedStream.java   | 4 
 4 files changed, 14 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/19ef82cf/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java
--
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java 
b/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java
index 3cde5e7..32ea0a3 100644
--- a/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java
+++ b/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java
@@ -19,6 +19,7 @@
 package org.apache.flink.api.common;
 
 import com.esotericsoftware.kryo.Serializer;
+import org.apache.flink.annotation.Internal;
 import org.apache.flink.annotation.Public;
 import org.apache.flink.annotation.PublicEvolving;
 import org.apache.flink.api.common.restartstrategy.RestartStrategies;
@@ -848,6 +849,7 @@ public class ExecutionConfig implements Serializable, 
Archiveablehttp://git-wip-us.apache.org/repos/asf/flink/blob/19ef82cf/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java
--
diff --git a/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java 
b/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java
index e6313aa..33addbb 100644
--- a/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java
+++ b/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java
@@ -25,6 +25,7 @@
 
 package org.apache.flink.core.fs;
 
+import org.apache.flink.annotation.Internal;
 import org.apache.flink.annotation.Public;
 import org.apache.flink.configuration.ConfigConstants;
 import org.apache.flink.configuration.Configuration;
@@ -97,6 +98,7 @@ public abstract class FileSystem {
 * Create a SafetyNetCloseableRegistry for a Task. This method should 
be called at the beginning of the task's
 * main thread.
 */
+   @Internal
public static void createFileSystemCloseableRegistryForTask() {
SafetyNetCloseableRegistry oldRegistry = REGISTRIES.get();
if (null != oldRegistry) {
@@ -111,6 +113,7 @@ public abstract class FileSystem {
 * Create a SafetyNetCloseableRegistry for a Task. This method should 
be called at the end of the task's
 * main thread or when the task should be canceled.
 */
+   @Internal
public static void disposeFileSystemCloseableRegistryForTask() {
SafetyNetCloseableRegistry registry = REGISTRIES.get();
if (null != registry) {
@@ -200,6 +203,7 @@ public abstract class FileSystem {
}
}
 
+   @Internal
public static FileSystem getUnguardedFileSystem(URI uri) throws 
IOException {
FileSystem fs;
 

http://git-wip-us.apache.org/repos/asf/flink/blob/19ef82cf/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AllWindowedStream.java
--
diff --git 
a/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AllWindowedStream.java
 
b/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AllWindowedStream.java
index ae71ce5..31ea001 100644
--- 
a/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AllWindowedStream.java
+++ 
b/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AllWindowedStream.java
@@ -199,7 +199,7 @@ public class AllWindowedStream {
 * @param function The window function.
 * @return The data stream that is the result of applying the window 
function to the window.
 */
-
+   @PublicEvolving
public  SingleOutputStreamOperator 

flink git commit: Commit for release 1.2.0

2017-01-24 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/release-1.2.0-rc1 [created] 1bca12088


Commit for release 1.2.0


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/1bca1208
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/1bca1208
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/1bca1208

Branch: refs/heads/release-1.2.0-rc1
Commit: 1bca120885f1950dc5d4436ae9f19025d26aad7c
Parents: 993a2e2
Author: Robert Metzger 
Authored: Tue Jan 24 13:00:32 2017 +
Committer: Robert Metzger 
Committed: Tue Jan 24 13:00:32 2017 +

--
 .gitattributes  |  3 --
 .github/CONTRIBUTING.md | 11 
 .github/PULL_REQUEST_TEMPLATE.md| 17 --
 .gitignore  | 36 -
 .travis.yml | 57 
 deploysettings.xml  | 38 -
 docs/_config.yml|  4 +-
 flink-annotations/pom.xml   |  2 +-
 flink-clients/pom.xml   |  2 +-
 flink-connectors/flink-avro/pom.xml |  2 +-
 .../flink-connector-cassandra/pom.xml   |  2 +-
 .../flink-connector-elasticsearch/pom.xml   |  2 +-
 .../flink-connector-elasticsearch2/pom.xml  |  2 +-
 .../flink-connector-filesystem/pom.xml  |  2 +-
 .../flink-connector-kafka-0.10/pom.xml  |  2 +-
 .../flink-connector-kafka-0.8/pom.xml   |  2 +-
 .../flink-connector-kafka-0.9/pom.xml   |  2 +-
 .../flink-connector-kafka-base/pom.xml  |  2 +-
 .../flink-connector-kinesis/pom.xml |  2 +-
 flink-connectors/flink-connector-nifi/pom.xml   |  2 +-
 .../flink-connector-rabbitmq/pom.xml|  2 +-
 .../flink-connector-twitter/pom.xml |  2 +-
 .../flink-hadoop-compatibility/pom.xml  |  2 +-
 flink-connectors/flink-hbase/pom.xml|  2 +-
 flink-connectors/flink-hcatalog/pom.xml |  2 +-
 flink-connectors/flink-jdbc/pom.xml |  2 +-
 flink-connectors/pom.xml|  2 +-
 flink-contrib/flink-connector-wikiedits/pom.xml |  2 +-
 .../flink-statebackend-rocksdb/pom.xml  |  2 +-
 flink-contrib/flink-storm-examples/pom.xml  |  2 +-
 flink-contrib/flink-storm/pom.xml   |  2 +-
 flink-contrib/flink-streaming-contrib/pom.xml   |  2 +-
 flink-contrib/flink-tweet-inputformat/pom.xml   |  2 +-
 flink-contrib/pom.xml   |  2 +-
 flink-core/pom.xml  |  2 +-
 flink-dist/pom.xml  |  2 +-
 flink-examples/flink-examples-batch/pom.xml |  2 +-
 flink-examples/flink-examples-streaming/pom.xml |  2 +-
 flink-examples/pom.xml  |  2 +-
 flink-fs-tests/pom.xml  |  2 +-
 flink-java/pom.xml  |  2 +-
 flink-java8/pom.xml |  2 +-
 flink-libraries/flink-cep-scala/pom.xml |  2 +-
 flink-libraries/flink-cep/pom.xml   |  2 +-
 flink-libraries/flink-gelly-examples/pom.xml|  2 +-
 flink-libraries/flink-gelly-scala/pom.xml   |  2 +-
 flink-libraries/flink-gelly/pom.xml |  2 +-
 flink-libraries/flink-ml/pom.xml|  2 +-
 flink-libraries/flink-python/pom.xml|  2 +-
 flink-libraries/flink-table/pom.xml |  2 +-
 flink-libraries/pom.xml |  2 +-
 flink-mesos/pom.xml |  2 +-
 flink-metrics/flink-metrics-core/pom.xml|  2 +-
 flink-metrics/flink-metrics-dropwizard/pom.xml  |  2 +-
 flink-metrics/flink-metrics-ganglia/pom.xml |  2 +-
 flink-metrics/flink-metrics-graphite/pom.xml|  2 +-
 flink-metrics/flink-metrics-jmx/pom.xml |  2 +-
 flink-metrics/flink-metrics-statsd/pom.xml  |  2 +-
 flink-metrics/pom.xml   |  2 +-
 flink-optimizer/pom.xml |  2 +-
 flink-quickstart/flink-quickstart-java/pom.xml  |  2 +-
 .../main/resources/archetype-resources/pom.xml  |  2 +-
 flink-quickstart/flink-quickstart-scala/pom.xml |  2 +-
 .../main/resources/archetype-resources/pom.xml  |  2 +-
 flink-quickstart/pom.xml|  2 +-
 flink-runtime-web/pom.xml   |  2 +-
 flink-runtime/pom.xml   |  2 +-
 flink-scala-shell/pom.xml   |  2 +-
 flink-scala/pom.xml |  2 +-
 .../flink-shaded-curator-recipes/pom.xml|  2 +-
 .../flink-shaded-curator-test/pom.xml   |  2 +-
 flink-shaded-curator/pom.xml|  2 +-
 .../flink-shaded-hadoop2/pom.xml|  2 +-
 .../flink-shaded-include-yarn-tests/pom.xml |  2 +-
 

flink git commit: Commit for release 1.2.0

2017-01-29 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/release-1.2.0-rc3 [created] 1c659cf4a


Commit for release 1.2.0


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/1c659cf4
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/1c659cf4
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/1c659cf4

Branch: refs/heads/release-1.2.0-rc3
Commit: 1c659cf4acca2cbc8c36494156d9ec28423e43c6
Parents: 617ff50
Author: Robert Metzger 
Authored: Sun Jan 29 21:19:15 2017 +
Committer: Robert Metzger 
Committed: Sun Jan 29 21:19:15 2017 +

--
 .gitattributes  |  3 --
 .github/CONTRIBUTING.md | 11 
 .github/PULL_REQUEST_TEMPLATE.md| 17 --
 .gitignore  | 36 -
 .travis.yml | 57 
 deploysettings.xml  | 38 -
 docs/_config.yml|  4 +-
 flink-annotations/pom.xml   |  2 +-
 flink-clients/pom.xml   |  2 +-
 flink-connectors/flink-avro/pom.xml |  2 +-
 .../flink-connector-cassandra/pom.xml   |  2 +-
 .../flink-connector-elasticsearch/pom.xml   |  2 +-
 .../flink-connector-elasticsearch2/pom.xml  |  2 +-
 .../flink-connector-filesystem/pom.xml  |  2 +-
 .../flink-connector-kafka-0.10/pom.xml  |  2 +-
 .../flink-connector-kafka-0.8/pom.xml   |  2 +-
 .../flink-connector-kafka-0.9/pom.xml   |  2 +-
 .../flink-connector-kafka-base/pom.xml  |  2 +-
 .../flink-connector-kinesis/pom.xml |  2 +-
 flink-connectors/flink-connector-nifi/pom.xml   |  2 +-
 .../flink-connector-rabbitmq/pom.xml|  2 +-
 .../flink-connector-twitter/pom.xml |  2 +-
 .../flink-hadoop-compatibility/pom.xml  |  2 +-
 flink-connectors/flink-hbase/pom.xml|  2 +-
 flink-connectors/flink-hcatalog/pom.xml |  2 +-
 flink-connectors/flink-jdbc/pom.xml |  2 +-
 flink-connectors/pom.xml|  2 +-
 flink-contrib/flink-connector-wikiedits/pom.xml |  2 +-
 .../flink-statebackend-rocksdb/pom.xml  |  2 +-
 flink-contrib/flink-storm-examples/pom.xml  |  2 +-
 flink-contrib/flink-storm/pom.xml   |  2 +-
 flink-contrib/flink-streaming-contrib/pom.xml   |  2 +-
 flink-contrib/flink-tweet-inputformat/pom.xml   |  2 +-
 flink-contrib/pom.xml   |  2 +-
 flink-core/pom.xml  |  2 +-
 flink-dist/pom.xml  |  2 +-
 flink-examples/flink-examples-batch/pom.xml |  2 +-
 flink-examples/flink-examples-streaming/pom.xml |  2 +-
 flink-examples/pom.xml  |  2 +-
 flink-fs-tests/pom.xml  |  2 +-
 flink-java/pom.xml  |  2 +-
 flink-java8/pom.xml |  2 +-
 flink-libraries/flink-cep-scala/pom.xml |  2 +-
 flink-libraries/flink-cep/pom.xml   |  2 +-
 flink-libraries/flink-gelly-examples/pom.xml|  2 +-
 flink-libraries/flink-gelly-scala/pom.xml   |  2 +-
 flink-libraries/flink-gelly/pom.xml |  2 +-
 flink-libraries/flink-ml/pom.xml|  2 +-
 flink-libraries/flink-python/pom.xml|  2 +-
 flink-libraries/flink-table/pom.xml |  2 +-
 flink-libraries/pom.xml |  2 +-
 flink-mesos/pom.xml |  2 +-
 flink-metrics/flink-metrics-core/pom.xml|  2 +-
 flink-metrics/flink-metrics-dropwizard/pom.xml  |  2 +-
 flink-metrics/flink-metrics-ganglia/pom.xml |  2 +-
 flink-metrics/flink-metrics-graphite/pom.xml|  2 +-
 flink-metrics/flink-metrics-jmx/pom.xml |  2 +-
 flink-metrics/flink-metrics-statsd/pom.xml  |  2 +-
 flink-metrics/pom.xml   |  2 +-
 flink-optimizer/pom.xml |  2 +-
 flink-quickstart/flink-quickstart-java/pom.xml  |  2 +-
 .../main/resources/archetype-resources/pom.xml  |  2 +-
 flink-quickstart/flink-quickstart-scala/pom.xml |  2 +-
 .../main/resources/archetype-resources/pom.xml  |  2 +-
 flink-quickstart/pom.xml|  2 +-
 flink-runtime-web/pom.xml   |  2 +-
 flink-runtime/pom.xml   |  2 +-
 flink-scala-shell/pom.xml   |  2 +-
 flink-scala/pom.xml |  2 +-
 .../flink-shaded-curator-recipes/pom.xml|  2 +-
 .../flink-shaded-curator-test/pom.xml   |  2 +-
 flink-shaded-curator/pom.xml|  2 +-
 .../flink-shaded-hadoop2/pom.xml|  2 +-
 .../flink-shaded-include-yarn-tests/pom.xml |  2 +-
 

flink git commit: [FLINK-5492] Fix TaskManager log entry

2017-01-25 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/release-1.2 d4b92bcf7 -> 887c074a7


[FLINK-5492] Fix TaskManager log entry


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/887c074a
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/887c074a
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/887c074a

Branch: refs/heads/release-1.2
Commit: 887c074a769d821cc210468dbe82b7b702ff6db5
Parents: d4b92bc
Author: Robert Metzger 
Authored: Wed Jan 25 14:27:29 2017 +0100
Committer: Robert Metzger 
Committed: Wed Jan 25 14:44:02 2017 +0100

--
 .../org/apache/flink/runtime/taskmanager/TaskManager.scala | 6 +++---
 1 file changed, 3 insertions(+), 3 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/887c074a/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala
--
diff --git 
a/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala
 
b/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala
index d2fb717..f1e74cd 100644
--- 
a/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala
+++ 
b/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala
@@ -1736,11 +1736,11 @@ object TaskManager {
   taskManagerClass: Class[_ <: TaskManager])
 : Unit = {
 
-LOG.info(s"Starting TaskManager")
+LOG.info("Starting TaskManager")
 
 // Bring up the TaskManager actor system first, bind it to the given 
address.
-
-LOG.info("Starting TaskManager actor system at {}:{}.", 
taskManagerHostname, actorSystemPort)
+
+LOG.info(s"Starting TaskManager actor system at 
$taskManagerHostname:$actorSystemPort.")
 
 val taskManagerSystem = try {
   val akkaConfig = AkkaUtils.getAkkaConfig(



flink git commit: [FLINK-5492] Fix TaskManager log entry

2017-01-25 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master 34e106f63 -> cf6b3fb22


[FLINK-5492] Fix TaskManager log entry


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/cf6b3fb2
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/cf6b3fb2
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/cf6b3fb2

Branch: refs/heads/master
Commit: cf6b3fb22c67a540dbfa96497d66041ba95ad358
Parents: 34e106f
Author: Robert Metzger 
Authored: Wed Jan 25 14:27:29 2017 +0100
Committer: Robert Metzger 
Committed: Wed Jan 25 14:45:04 2017 +0100

--
 .../org/apache/flink/runtime/taskmanager/TaskManager.scala | 6 +++---
 1 file changed, 3 insertions(+), 3 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/cf6b3fb2/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala
--
diff --git 
a/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala
 
b/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala
index 3c142cd..7cb1902 100644
--- 
a/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala
+++ 
b/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala
@@ -1745,11 +1745,11 @@ object TaskManager {
   taskManagerClass: Class[_ <: TaskManager])
 : Unit = {
 
-LOG.info(s"Starting TaskManager")
+LOG.info("Starting TaskManager")
 
 // Bring up the TaskManager actor system first, bind it to the given 
address.
-
-LOG.info("Starting TaskManager actor system at {}:{}.", 
taskManagerHostname, actorSystemPort)
+
+LOG.info(s"Starting TaskManager actor system at 
$taskManagerHostname:$actorSystemPort.")
 
 val taskManagerSystem = try {
   val akkaConfig = AkkaUtils.getAkkaConfig(



flink git commit: [FLINK-5637] Avoid warning while parsing YAML configuration files

2017-01-25 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/release-1.2 80c9f8018 -> f2240eb93


[FLINK-5637] Avoid warning while parsing YAML configuration files

This closes #3216


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/f2240eb9
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/f2240eb9
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/f2240eb9

Branch: refs/heads/release-1.2
Commit: f2240eb9394b368b392ac83d37106482a005e0a2
Parents: 80c9f80
Author: Robert Metzger 
Authored: Wed Jan 25 16:31:11 2017 +0100
Committer: Robert Metzger 
Committed: Wed Jan 25 18:01:25 2017 +0100

--
 .../org/apache/flink/configuration/GlobalConfiguration.java | 9 +
 .../apache/flink/configuration/GlobalConfigurationTest.java | 3 ++-
 flink-dist/src/main/resources/flink-conf.yaml   | 2 +-
 3 files changed, 8 insertions(+), 6 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/f2240eb9/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java
--
diff --git 
a/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java
 
b/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java
index dca6307..ea9f8bf 100644
--- 
a/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java
+++ 
b/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java
@@ -144,11 +144,12 @@ public final class GlobalConfiguration {
try (BufferedReader reader = new BufferedReader(new 
InputStreamReader(new FileInputStream(file{
 
String line;
+   int lineNo = 0;
while ((line = reader.readLine()) != null) {
-
+   lineNo++;
// 1. check for comments
String[] comments = line.split("#", 2);
-   String conf = comments[0];
+   String conf = comments[0].trim();
 
// 2. get key and value
if (conf.length() > 0) {
@@ -156,7 +157,7 @@ public final class GlobalConfiguration {
 
// skip line with no valid key-value 
pair
if (kv.length == 1) {
-   LOG.warn("Error while trying to 
split key and value in configuration file " + file + ": " + line);
+   LOG.warn("Error while trying to 
split key and value in configuration file " + file + ":" + lineNo + ": \"" + 
line + "\"");
continue;
}
 
@@ -165,7 +166,7 @@ public final class GlobalConfiguration {
 
// sanity check
if (key.length() == 0 || value.length() 
== 0) {
-   LOG.warn("Error after splitting 
key and value in configuration file " + file + ": " + line);
+   LOG.warn("Error after splitting 
key and value in configuration file " + file + ":" + lineNo + ": \"" + line + 
"\"");
continue;
}
 

http://git-wip-us.apache.org/repos/asf/flink/blob/f2240eb9/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java
--
diff --git 
a/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java
 
b/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java
index 6336a73..322e64d 100644
--- 
a/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java
+++ 
b/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java
@@ -56,7 +56,8 @@ public class GlobalConfigurationTest extends TestLogger {
pw.println("mykey3:myvalue3"); // SKIP, missing 
white space after colon
pw.println(" some nonsense without colon and 
whitespace separator"); // SKIP
pw.println(" :  "); // SKIP
-   pw.println("   "); // SKIP
+   pw.println("   "); // SKIP (silently)
+   pw.println(" "); // SKIP (silently)
pw.println("mykey4: myvalue4# some comments"); 
// OK, skip comments only
   

flink git commit: [FLINK-5637] Avoid warning while parsing YAML configuration files

2017-01-25 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master b9ea4cfdf -> 54d3e2668


[FLINK-5637] Avoid warning while parsing YAML configuration files


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/54d3e266
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/54d3e266
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/54d3e266

Branch: refs/heads/master
Commit: 54d3e26686998a0e5e472651cd9d709d491e7093
Parents: b9ea4cf
Author: Robert Metzger 
Authored: Wed Jan 25 16:31:11 2017 +0100
Committer: Robert Metzger 
Committed: Wed Jan 25 18:02:00 2017 +0100

--
 .../org/apache/flink/configuration/GlobalConfiguration.java | 9 +
 .../apache/flink/configuration/GlobalConfigurationTest.java | 3 ++-
 flink-dist/src/main/resources/flink-conf.yaml   | 2 +-
 3 files changed, 8 insertions(+), 6 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/54d3e266/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java
--
diff --git 
a/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java
 
b/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java
index dca6307..ea9f8bf 100644
--- 
a/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java
+++ 
b/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java
@@ -144,11 +144,12 @@ public final class GlobalConfiguration {
try (BufferedReader reader = new BufferedReader(new 
InputStreamReader(new FileInputStream(file{
 
String line;
+   int lineNo = 0;
while ((line = reader.readLine()) != null) {
-
+   lineNo++;
// 1. check for comments
String[] comments = line.split("#", 2);
-   String conf = comments[0];
+   String conf = comments[0].trim();
 
// 2. get key and value
if (conf.length() > 0) {
@@ -156,7 +157,7 @@ public final class GlobalConfiguration {
 
// skip line with no valid key-value 
pair
if (kv.length == 1) {
-   LOG.warn("Error while trying to 
split key and value in configuration file " + file + ": " + line);
+   LOG.warn("Error while trying to 
split key and value in configuration file " + file + ":" + lineNo + ": \"" + 
line + "\"");
continue;
}
 
@@ -165,7 +166,7 @@ public final class GlobalConfiguration {
 
// sanity check
if (key.length() == 0 || value.length() 
== 0) {
-   LOG.warn("Error after splitting 
key and value in configuration file " + file + ": " + line);
+   LOG.warn("Error after splitting 
key and value in configuration file " + file + ":" + lineNo + ": \"" + line + 
"\"");
continue;
}
 

http://git-wip-us.apache.org/repos/asf/flink/blob/54d3e266/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java
--
diff --git 
a/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java
 
b/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java
index 6336a73..322e64d 100644
--- 
a/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java
+++ 
b/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java
@@ -56,7 +56,8 @@ public class GlobalConfigurationTest extends TestLogger {
pw.println("mykey3:myvalue3"); // SKIP, missing 
white space after colon
pw.println(" some nonsense without colon and 
whitespace separator"); // SKIP
pw.println(" :  "); // SKIP
-   pw.println("   "); // SKIP
+   pw.println("   "); // SKIP (silently)
+   pw.println(" "); // SKIP (silently)
pw.println("mykey4: myvalue4# some comments"); 
// OK, skip comments only
pw.println("   mykey5   

flink git commit: Commit for release 1.2.0

2017-01-25 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/release-1.2.0-rc2 [created] 8b5b6a8b2


Commit for release 1.2.0


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/8b5b6a8b
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/8b5b6a8b
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/8b5b6a8b

Branch: refs/heads/release-1.2.0-rc2
Commit: 8b5b6a8b27516578ddf05c5e77c840f5666bb28f
Parents: f2240eb
Author: Robert Metzger 
Authored: Wed Jan 25 17:17:29 2017 +
Committer: Robert Metzger 
Committed: Wed Jan 25 17:17:29 2017 +

--
 .gitattributes  |  3 --
 .github/CONTRIBUTING.md | 11 
 .github/PULL_REQUEST_TEMPLATE.md| 17 --
 .gitignore  | 36 -
 .travis.yml | 57 
 deploysettings.xml  | 38 -
 docs/_config.yml|  4 +-
 flink-annotations/pom.xml   |  2 +-
 flink-clients/pom.xml   |  2 +-
 flink-connectors/flink-avro/pom.xml |  2 +-
 .../flink-connector-cassandra/pom.xml   |  2 +-
 .../flink-connector-elasticsearch/pom.xml   |  2 +-
 .../flink-connector-elasticsearch2/pom.xml  |  2 +-
 .../flink-connector-filesystem/pom.xml  |  2 +-
 .../flink-connector-kafka-0.10/pom.xml  |  2 +-
 .../flink-connector-kafka-0.8/pom.xml   |  2 +-
 .../flink-connector-kafka-0.9/pom.xml   |  2 +-
 .../flink-connector-kafka-base/pom.xml  |  2 +-
 .../flink-connector-kinesis/pom.xml |  2 +-
 flink-connectors/flink-connector-nifi/pom.xml   |  2 +-
 .../flink-connector-rabbitmq/pom.xml|  2 +-
 .../flink-connector-twitter/pom.xml |  2 +-
 .../flink-hadoop-compatibility/pom.xml  |  2 +-
 flink-connectors/flink-hbase/pom.xml|  2 +-
 flink-connectors/flink-hcatalog/pom.xml |  2 +-
 flink-connectors/flink-jdbc/pom.xml |  2 +-
 flink-connectors/pom.xml|  2 +-
 flink-contrib/flink-connector-wikiedits/pom.xml |  2 +-
 .../flink-statebackend-rocksdb/pom.xml  |  2 +-
 flink-contrib/flink-storm-examples/pom.xml  |  2 +-
 flink-contrib/flink-storm/pom.xml   |  2 +-
 flink-contrib/flink-streaming-contrib/pom.xml   |  2 +-
 flink-contrib/flink-tweet-inputformat/pom.xml   |  2 +-
 flink-contrib/pom.xml   |  2 +-
 flink-core/pom.xml  |  2 +-
 flink-dist/pom.xml  |  2 +-
 flink-examples/flink-examples-batch/pom.xml |  2 +-
 flink-examples/flink-examples-streaming/pom.xml |  2 +-
 flink-examples/pom.xml  |  2 +-
 flink-fs-tests/pom.xml  |  2 +-
 flink-java/pom.xml  |  2 +-
 flink-java8/pom.xml |  2 +-
 flink-libraries/flink-cep-scala/pom.xml |  2 +-
 flink-libraries/flink-cep/pom.xml   |  2 +-
 flink-libraries/flink-gelly-examples/pom.xml|  2 +-
 flink-libraries/flink-gelly-scala/pom.xml   |  2 +-
 flink-libraries/flink-gelly/pom.xml |  2 +-
 flink-libraries/flink-ml/pom.xml|  2 +-
 flink-libraries/flink-python/pom.xml|  2 +-
 flink-libraries/flink-table/pom.xml |  2 +-
 flink-libraries/pom.xml |  2 +-
 flink-mesos/pom.xml |  2 +-
 flink-metrics/flink-metrics-core/pom.xml|  2 +-
 flink-metrics/flink-metrics-dropwizard/pom.xml  |  2 +-
 flink-metrics/flink-metrics-ganglia/pom.xml |  2 +-
 flink-metrics/flink-metrics-graphite/pom.xml|  2 +-
 flink-metrics/flink-metrics-jmx/pom.xml |  2 +-
 flink-metrics/flink-metrics-statsd/pom.xml  |  2 +-
 flink-metrics/pom.xml   |  2 +-
 flink-optimizer/pom.xml |  2 +-
 flink-quickstart/flink-quickstart-java/pom.xml  |  2 +-
 .../main/resources/archetype-resources/pom.xml  |  2 +-
 flink-quickstart/flink-quickstart-scala/pom.xml |  2 +-
 .../main/resources/archetype-resources/pom.xml  |  2 +-
 flink-quickstart/pom.xml|  2 +-
 flink-runtime-web/pom.xml   |  2 +-
 flink-runtime/pom.xml   |  2 +-
 flink-scala-shell/pom.xml   |  2 +-
 flink-scala/pom.xml |  2 +-
 .../flink-shaded-curator-recipes/pom.xml|  2 +-
 .../flink-shaded-curator-test/pom.xml   |  2 +-
 flink-shaded-curator/pom.xml|  2 +-
 .../flink-shaded-hadoop2/pom.xml|  2 +-
 .../flink-shaded-include-yarn-tests/pom.xml |  2 +-
 

flink git commit: FLINK-5731 Spilt up tests into three disjoint groups

2017-02-22 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master b0f0f3722 -> e7a914d4e


FLINK-5731 Spilt up tests into three disjoint groups

This closes #3344


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/e7a914d4
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/e7a914d4
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/e7a914d4

Branch: refs/heads/master
Commit: e7a914d4e926216e1382e622556be8c8358b27c9
Parents: b0f0f37
Author: Robert Metzger 
Authored: Mon Feb 13 21:43:25 2017 +0100
Committer: Robert Metzger 
Committed: Wed Feb 22 20:29:49 2017 +0100

--
 .travis.yml |  9 +
 pom.xml | 13 +++--
 2 files changed, 20 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/e7a914d4/.travis.yml
--
diff --git a/.travis.yml b/.travis.yml
index 1445f8d..74bec6b 100644
--- a/.travis.yml
+++ b/.travis.yml
@@ -19,21 +19,30 @@ matrix:
   env: PROFILE="-Dhadoop.version=2.7.2 -Dscala-2.11 
-Pinclude-yarn-tests,flink-fast-tests-a,include-kinesis 
-Dmaven.javadoc.skip=true"
 - jdk: "oraclejdk8"
   env: PROFILE="-Dhadoop.version=2.7.2 -Dscala-2.11 
-Pinclude-yarn-tests,flink-fast-tests-b,include-kinesis 
-Dmaven.javadoc.skip=true"
+- jdk: "oraclejdk8"
+  env: PROFILE="-Dhadoop.version=2.7.2 -Dscala-2.11 
-Pinclude-yarn-tests,flink-fast-tests-c,include-kinesis 
-Dmaven.javadoc.skip=true"
 
 - jdk: "oraclejdk8"
   env: PROFILE="-Dhadoop.version=2.6.3 
-Pinclude-yarn-tests,flink-fast-tests-a,include-kinesis 
-Dmaven.javadoc.skip=true"
 - jdk: "oraclejdk8"
   env: PROFILE="-Dhadoop.version=2.6.3 
-Pinclude-yarn-tests,flink-fast-tests-b,include-kinesis 
-Dmaven.javadoc.skip=true"
+- jdk: "oraclejdk8"
+  env: PROFILE="-Dhadoop.version=2.6.3 
-Pinclude-yarn-tests,flink-fast-tests-c,include-kinesis 
-Dmaven.javadoc.skip=true"
 
 - jdk: "openjdk7"
   env: PROFILE="-Dhadoop.version=2.4.1 -Dscala-2.11 
-Pinclude-yarn-tests,flink-fast-tests-a,include-kinesis"
 - jdk: "openjdk7"
   env: PROFILE="-Dhadoop.version=2.4.1 -Dscala-2.11 
-Pinclude-yarn-tests,flink-fast-tests-b,include-kinesis"
+- jdk: "openjdk7"
+  env: PROFILE="-Dhadoop.version=2.4.1 -Dscala-2.11 
-Pinclude-yarn-tests,flink-fast-tests-c,include-kinesis"
 
 - jdk: "oraclejdk7"
   env: PROFILE="-Dhadoop.version=2.3.0 
-Pflink-fast-tests-a,include-kinesis"
 - jdk: "oraclejdk7"
   env: PROFILE="-Dhadoop.version=2.3.0 
-Pflink-fast-tests-b,include-kinesis"
+- jdk: "oraclejdk7"
+  env: PROFILE="-Dhadoop.version=2.3.0 
-Pflink-fast-tests-c,include-kinesis"
+
 
 
 git:

http://git-wip-us.apache.org/repos/asf/flink/blob/e7a914d4/pom.xml
--
diff --git a/pom.xml b/pom.xml
index d81296e..0070835 100644
--- a/pom.xml
+++ b/pom.xml
@@ -767,13 +767,22 @@ under the License.

flink-fast-tests-a

-   
%regex[.*/[A-M].*]
+   
+   
%regex[.*/[J-Z].*]



flink-fast-tests-b

-   
%regex[.*/[N-Z].*]
+   
+   
%regex[.*/[A-IS-Z].*]
+   
+   
+   
+   flink-fast-tests-c
+   
+   
+   
%regex[.*/[A-R].*]






flink git commit: [FLINK-5690][docs] Add note on shading to best practices guide

2017-02-14 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master 8efacf588 -> d32281444


[FLINK-5690][docs] Add note on shading to best practices guide

This closes #3300


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/d3228144
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/d3228144
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/d3228144

Branch: refs/heads/master
Commit: d322814447e9ce437eb1adffef24699f26803038
Parents: 8efacf5
Author: Robert Metzger 
Authored: Mon Feb 13 20:50:23 2017 +0100
Committer: Robert Metzger 
Committed: Tue Feb 14 15:05:16 2017 +0100

--
 docs/dev/best_practices.md| 314 +
 docs/monitoring/best_practices.md | 312 
 docs/monitoring/debugging_classloading.md |  14 ++
 3 files changed, 328 insertions(+), 312 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/d3228144/docs/dev/best_practices.md
--
diff --git a/docs/dev/best_practices.md b/docs/dev/best_practices.md
new file mode 100644
index 000..b2111c4
--- /dev/null
+++ b/docs/dev/best_practices.md
@@ -0,0 +1,314 @@
+---
+title: "Best Practices"
+nav-parent_id: dev
+nav-pos: 90
+---
+
+
+This page contains a collection of best practices for Flink programmers on how 
to solve frequently encountered problems.
+
+
+* This will be replaced by the TOC
+{:toc}
+
+## Parsing command line arguments and passing them around in your Flink 
application
+
+
+Almost all Flink applications, both batch and streaming rely on external 
configuration parameters.
+For example for specifying input and output sources (like paths or addresses), 
also system parameters (parallelism, runtime configuration) and application 
specific parameters (often used within the user functions).
+
+Since version 0.9 we are providing a simple utility called `ParameterTool` to 
provide at least some basic tooling for solving these problems.
+
+Please note that you don't have to use the `ParameterTool` explained here. 
Other frameworks such as [Commons 
CLI](https://commons.apache.org/proper/commons-cli/),
+[argparse4j](http://argparse4j.sourceforge.net/) and others work well with 
Flink as well.
+
+
+### Getting your configuration values into the `ParameterTool`
+
+The `ParameterTool` provides a set of predefined static methods for reading 
the configuration. The tool is internally expecting a `Map`, so 
its very easy to integrate it with your own configuration style.
+
+
+ From `.properties` files
+
+The following method will read a 
[Properties](https://docs.oracle.com/javase/tutorial/essential/environment/properties.html)
 file and provide the key/value pairs:
+{% highlight java %}
+String propertiesFile = "/home/sam/flink/myjob.properties";
+ParameterTool parameter = ParameterTool.fromPropertiesFile(propertiesFile);
+{% endhighlight %}
+
+
+ From the command line arguments
+
+This allows getting arguments like `--input hdfs:///mydata --elements 42` from 
the command line.
+{% highlight java %}
+public static void main(String[] args) {
+   ParameterTool parameter = ParameterTool.fromArgs(args);
+   // .. regular code ..
+{% endhighlight %}
+
+
+ From system properties
+
+When starting a JVM, you can pass system properties to it: 
`-Dinput=hdfs:///mydata`. You can also initialize the `ParameterTool` from 
these system properties:
+
+{% highlight java %}
+ParameterTool parameter = ParameterTool.fromSystemProperties();
+{% endhighlight %}
+
+
+### Using the parameters in your Flink program
+
+Now that we've got the parameters from somewhere (see above) we can use them 
in various ways.
+
+**Directly from the `ParameterTool`**
+
+The `ParameterTool` itself has methods for accessing the values.
+{% highlight java %}
+ParameterTool parameters = // ...
+parameter.getRequired("input");
+parameter.get("output", "myDefaultValue");
+parameter.getLong("expectedCount", -1L);
+parameter.getNumberOfParameters()
+// .. there are more methods available.
+{% endhighlight %}
+
+You can use the return values of these methods directly in the main() method 
(=the client submitting the application).
+For example you could set the parallelism of a operator like this:
+
+{% highlight java %}
+ParameterTool parameters = ParameterTool.fromArgs(args);
+int parallelism = parameters.get("mapParallelism", 2);
+DataSet> counts = text.flatMap(new 
Tokenizer()).setParallelism(parallelism);
+{% endhighlight %}
+
+Since the `ParameterTool` is serializable, you can pass it to the functions 
itself:
+
+{% highlight java %}
+ParameterTool parameters = ParameterTool.fromArgs(args);
+DataSet> counts = 

flink git commit: [FLINK-5613][query] querying a non-existing key is inconsistent among state backends

2017-01-23 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master 24fe08148 -> da26bdc2e


[FLINK-5613][query] querying a non-existing key is inconsistent among state 
backends

Querying for a non-existing key for a state that has a default value set
currently results in an UnknownKeyOrNamespace exception when the
MemoryStateBackend or FsStateBackend is used but results in the default value
if RocksDBStateBackend is set.

This removes the special handling from the RocksDBStateBackend and makes it
consistent with the other two back-ends, i.e. returning null which results
in the mentioned UnknownKeyOrNamespace exception.

This closes #3193


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/da26bdc2
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/da26bdc2
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/da26bdc2

Branch: refs/heads/master
Commit: da26bdc2eda1c8d70ada8b1969db1d48d2095ed1
Parents: 24fe081
Author: Nico Kruber 
Authored: Tue Jan 17 14:26:16 2017 +0100
Committer: Robert Metzger 
Committed: Mon Jan 23 18:47:44 2017 +0100

--
 .../streaming/state/RocksDBValueState.java  |  12 --
 .../flink/test/query/QueryableStateITCase.java  | 160 ++-
 2 files changed, 153 insertions(+), 19 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/da26bdc2/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java
--
diff --git 
a/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java
 
b/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java
index 7724f02..b2a4fba 100644
--- 
a/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java
+++ 
b/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java
@@ -23,7 +23,6 @@ import org.apache.flink.api.common.state.ValueStateDescriptor;
 import org.apache.flink.api.common.typeutils.TypeSerializer;
 import org.apache.flink.core.memory.DataInputViewStreamWrapper;
 import org.apache.flink.core.memory.DataOutputViewStreamWrapper;
-import org.apache.flink.runtime.query.netty.message.KvStateRequestSerializer;
 import org.apache.flink.runtime.state.internal.InternalValueState;
 import org.rocksdb.ColumnFamilyHandle;
 import org.rocksdb.RocksDBException;
@@ -103,15 +102,4 @@ public class RocksDBValueState
throw new RuntimeException("Error while adding data to 
RocksDB", e);
}
}
-
-   @Override
-   public byte[] getSerializedValue(byte[] serializedKeyAndNamespace) 
throws Exception {
-   byte[] value = 
super.getSerializedValue(serializedKeyAndNamespace);
-
-   if (value != null) {
-   return value;
-   } else {
-   return 
KvStateRequestSerializer.serializeValue(stateDesc.getDefaultValue(), 
stateDesc.getSerializer());
-   }
-   }
 }

http://git-wip-us.apache.org/repos/asf/flink/blob/da26bdc2/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java
--
diff --git 
a/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java
 
b/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java
index 327a715..c2df6ae 100644
--- 
a/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java
+++ 
b/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java
@@ -39,6 +39,7 @@ import org.apache.flink.api.java.tuple.Tuple2;
 import org.apache.flink.configuration.ConfigConstants;
 import org.apache.flink.configuration.Configuration;
 import org.apache.flink.configuration.QueryableStateOptions;
+import org.apache.flink.contrib.streaming.state.RocksDBStateBackend;
 import org.apache.flink.runtime.akka.AkkaUtils;
 import org.apache.flink.runtime.executiongraph.ExecutionAttemptID;
 import org.apache.flink.runtime.executiongraph.ExecutionGraph;
@@ -50,10 +51,14 @@ import 
org.apache.flink.runtime.messages.JobManagerMessages.JobFound;
 import org.apache.flink.runtime.query.KvStateLocation;
 import org.apache.flink.runtime.query.KvStateMessage;
 import org.apache.flink.runtime.query.QueryableStateClient;
+import org.apache.flink.runtime.query.netty.UnknownKeyOrNamespace;
 import org.apache.flink.runtime.query.netty.message.KvStateRequestSerializer;
+import org.apache.flink.runtime.state.AbstractStateBackend;
 import 

flink git commit: [FLINK-5395][release script] Fix version commit and maven upload

2017-01-24 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master b351aa2dd -> 5f0d8c9dd


[FLINK-5395][release script] Fix version commit and maven upload


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/5f0d8c9d
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/5f0d8c9d
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/5f0d8c9d

Branch: refs/heads/master
Commit: 5f0d8c9dd56935f4a62ee3f57aa2d411f102bead
Parents: b351aa2
Author: Robert Metzger 
Authored: Tue Jan 24 15:14:55 2017 +0100
Committer: Robert Metzger 
Committed: Tue Jan 24 15:14:55 2017 +0100

--
 tools/create_release_files.sh | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/5f0d8c9d/tools/create_release_files.sh
--
diff --git a/tools/create_release_files.sh b/tools/create_release_files.sh
index ed01d97..f4c4673 100755
--- a/tools/create_release_files.sh
+++ b/tools/create_release_files.sh
@@ -185,7 +185,7 @@ make_source_release() {
   cd ..
 
   # local dist have no need to commit to remote
-  if [ ! IS_LOCAL_DIST ]; then
+  if [ "$IS_LOCAL_DIST" == "false" ]; then
 git commit --author="$GIT_AUTHOR" -am "Commit for release $RELEASE_VERSION"
 git remote add asf_push https://$USER_NAME@$GIT_REPO
 RELEASE_HASH=`git rev-parse HEAD`
@@ -299,7 +299,7 @@ else
   make_binary_release "hadoop2x" "-Dhadoop.version=$HADOOP_VERSION" 
"$SCALA_VERSION"
 fi
 
-if [ ! IS_LOCAL_DIST ] ; then
+if [ "$IS_LOCAL_DIST" == "false" ] ; then
 copy_data
 deploy_to_maven
 fi



flink git commit: [FLINK-3150] Make YARN container invocation configurable

2017-01-18 Thread rmetzger
Repository: flink
Updated Branches:
  refs/heads/master 6fb6967b9 -> 8f4139a42


[FLINK-3150] Make YARN container invocation configurable

This closes #3056


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/8f4139a4
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/8f4139a4
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/8f4139a4

Branch: refs/heads/master
Commit: 8f4139a42a619738a4efa45ea0b5d20009718aba
Parents: 6fb6967
Author: Nico Kruber 
Authored: Mon Jan 2 16:17:09 2017 +0100
Committer: Robert Metzger 
Committed: Wed Jan 18 11:17:02 2017 +0100

--
 docs/setup/config.md|   5 +
 .../flink/configuration/ConfigConstants.java|  12 ++
 .../clusterframework/BootstrapTools.java|  95 +---
 .../clusterframework/BootstrapToolsTest.java| 148 ++
 .../yarn/AbstractYarnClusterDescriptor.java |  46 --
 .../flink/yarn/YarnClusterDescriptorTest.java   | 149 +++
 6 files changed, 416 insertions(+), 39 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/flink/blob/8f4139a4/docs/setup/config.md
--
diff --git a/docs/setup/config.md b/docs/setup/config.md
index 1b2be8a..1a72e27 100644
--- a/docs/setup/config.md
+++ b/docs/setup/config.md
@@ -425,6 +425,11 @@ of the JobManager, because the same ActorSystem is used. 
Its not possible to use
 
 - `yarn.taskmanager.env.` Similar to the configuration prefix about, this 
prefix allows setting custom environment variables for the TaskManager 
processes.
 
+- `yarn.container-start-command-template`: Flink uses the following template 
when starting on YARN:
+`%java% %jvmmem% %jvmopts% %logging% %class% %args% %redirects%`. This 
configuration parameter allows users
+to pass custom settings (such as JVM paths, arguments etc.). Note that in most 
cases, it is sufficient to
+use the `env.java.opts` setting, which is the `%jvmopts%` variable in the 
String.
+
 - `yarn.application-master.port` (Default: 0, which lets the OS choose an 
ephemeral port) With this configuration option, users can specify a port, a 
range of ports or a list of ports for the  Application Master (and JobManager) 
RPC port. By default we recommend using the default value (0) to let the 
operating system choose an appropriate port. In particular when multiple AMs 
are running on the  same physical host, fixed port assignments prevent the AM 
from starting.
 
   For example when running Flink on YARN on an environment with a restrictive 
firewall, this option allows specifying a range of allowed ports.

http://git-wip-us.apache.org/repos/asf/flink/blob/8f4139a4/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java
--
diff --git 
a/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java 
b/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java
index fc389e0..14ba9dd 100644
--- 
a/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java
+++ 
b/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java
@@ -423,6 +423,12 @@ public final class ConfigConstants {
 */
@Deprecated
public static final String YARN_TASK_MANAGER_ENV_PREFIX = 
"yarn.taskmanager.env.";
+
+   /**
+* Template for the YARN container start incovation.
+*/
+   public static final String YARN_CONTAINER_START_COMMAND_TEMPLATE =
+   "yarn.container-start-command-template";

 /**
 * The config parameter defining the Akka actor system port for the 
ApplicationMaster and
@@ -1146,6 +1152,12 @@ public final class ConfigConstants {
public static final float DEFAULT_YARN_HEAP_CUTOFF_RATIO = 0.25f;
 
/**
+* Start command template for Flink on YARN containers
+*/
+   public static final String 
DEFAULT_YARN_CONTAINER_START_COMMAND_TEMPLATE =
+   "%java% %jvmmem% %jvmopts% %logging% %class% %args% 
%redirects%";
+
+   /**
 * Default port for the application master is 0, which means
 * the operating system assigns an ephemeral port
 */

http://git-wip-us.apache.org/repos/asf/flink/blob/8f4139a4/flink-runtime/src/main/java/org/apache/flink/runtime/clusterframework/BootstrapTools.java
--
diff --git 
a/flink-runtime/src/main/java/org/apache/flink/runtime/clusterframework/BootstrapTools.java
 
b/flink-runtime/src/main/java/org/apache/flink/runtime/clusterframework/BootstrapTools.java
index 2a33c44..1ef2cdc 100644
--- 

flink-web git commit: Added a page that redirects to the latest stable docs. We can link to this page from older versions of the docs.

2017-01-19 Thread rmetzger
Repository: flink-web
Updated Branches:
  refs/heads/asf-site 9ec0a879f -> ccdf85ae5


Added a page that redirects to the latest stable docs. We can link to this page 
from older versions of the docs.


Project: http://git-wip-us.apache.org/repos/asf/flink-web/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink-web/commit/ccdf85ae
Tree: http://git-wip-us.apache.org/repos/asf/flink-web/tree/ccdf85ae
Diff: http://git-wip-us.apache.org/repos/asf/flink-web/diff/ccdf85ae

Branch: refs/heads/asf-site
Commit: ccdf85ae593228ca0a728470b73b910122262217
Parents: 9ec0a87
Author: David Anderson 
Authored: Thu Jan 19 15:39:21 2017 +0100
Committer: David Anderson 
Committed: Thu Jan 19 15:39:21 2017 +0100

--
 content/q/stable-docs.html | 30 ++
 q/stable-docs.html | 30 ++
 2 files changed, 60 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/flink-web/blob/ccdf85ae/content/q/stable-docs.html
--
diff --git a/content/q/stable-docs.html b/content/q/stable-docs.html
new file mode 100644
index 000..70907a8
--- /dev/null
+++ b/content/q/stable-docs.html
@@ -0,0 +1,30 @@
+
+
+
+
+
+
+https://ci.apache.org/projects/flink/flink-docs-release-1.1/; />
+  
+

http://git-wip-us.apache.org/repos/asf/flink-web/blob/ccdf85ae/q/stable-docs.html
--
diff --git a/q/stable-docs.html b/q/stable-docs.html
new file mode 100644
index 000..70907a8
--- /dev/null
+++ b/q/stable-docs.html
@@ -0,0 +1,30 @@
+
+
+
+
+
+
+https://ci.apache.org/projects/flink/flink-docs-release-1.1/; />
+  
+



[03/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/js/jquery-2.1.0.js
--
diff --git a/content/visualizer/js/jquery-2.1.0.js 
b/content/visualizer/js/jquery-2.1.0.js
deleted file mode 100755
index f7f4227..000
--- a/content/visualizer/js/jquery-2.1.0.js
+++ /dev/null
@@ -1,9111 +0,0 @@
-/*!
- * jQuery JavaScript Library v2.1.0
- * http://jquery.com/
- *
- * Includes Sizzle.js
- * http://sizzlejs.com/
- *
- * Copyright 2005, 2014 jQuery Foundation, Inc. and other contributors
- * Released under the MIT license
- * http://jquery.org/license
- *
- * Date: 2014-01-23T21:10Z
- */
-
-(function( global, factory ) {
-
-   if ( typeof module === "object" && typeof module.exports === "object" ) 
{
-   // For CommonJS and CommonJS-like environments where a proper 
window is present,
-   // execute the factory and get jQuery
-   // For environments that do not inherently posses a window with 
a document
-   // (such as Node.js), expose a jQuery-making factory as 
module.exports
-   // This accentuates the need for the creation of a real window
-   // e.g. var jQuery = require("jquery")(window);
-   // See ticket #14549 for more info
-   module.exports = global.document ?
-   factory( global, true ) :
-   function( w ) {
-   if ( !w.document ) {
-   throw new Error( "jQuery requires a 
window with a document" );
-   }
-   return factory( w );
-   };
-   } else {
-   factory( global );
-   }
-
-// Pass this if window is not defined yet
-}(typeof window !== "undefined" ? window : this, function( window, noGlobal ) {
-
-// Can't do this because several apps including ASP.NET trace
-// the stack via arguments.caller.callee and Firefox dies if
-// you try to trace through "use strict" call chains. (#13335)
-// Support: Firefox 18+
-//
-
-var arr = [];
-
-var slice = arr.slice;
-
-var concat = arr.concat;
-
-var push = arr.push;
-
-var indexOf = arr.indexOf;
-
-var class2type = {};
-
-var toString = class2type.toString;
-
-var hasOwn = class2type.hasOwnProperty;
-
-var trim = "".trim;
-
-var support = {};
-
-
-
-var
-   // Use the correct document accordingly with window argument (sandbox)
-   document = window.document,
-
-   version = "2.1.0",
-
-   // Define a local copy of jQuery
-   jQuery = function( selector, context ) {
-   // The jQuery object is actually just the init constructor 
'enhanced'
-   // Need init if jQuery is called (just allow error to be thrown 
if not included)
-   return new jQuery.fn.init( selector, context );
-   },
-
-   // Matches dashed string for camelizing
-   rmsPrefix = /^-ms-/,
-   rdashAlpha = /-([\da-z])/gi,
-
-   // Used by jQuery.camelCase as callback to replace()
-   fcamelCase = function( all, letter ) {
-   return letter.toUpperCase();
-   };
-
-jQuery.fn = jQuery.prototype = {
-   // The current version of jQuery being used
-   jquery: version,
-
-   constructor: jQuery,
-
-   // Start with an empty selector
-   selector: "",
-
-   // The default length of a jQuery object is 0
-   length: 0,
-
-   toArray: function() {
-   return slice.call( this );
-   },
-
-   // Get the Nth element in the matched element set OR
-   // Get the whole matched element set as a clean array
-   get: function( num ) {
-   return num != null ?
-
-   // Return a 'clean' array
-   ( num < 0 ? this[ num + this.length ] : this[ num ] ) :
-
-   // Return just the object
-   slice.call( this );
-   },
-
-   // Take an array of elements and push it onto the stack
-   // (returning the new matched element set)
-   pushStack: function( elems ) {
-
-   // Build a new jQuery matched element set
-   var ret = jQuery.merge( this.constructor(), elems );
-
-   // Add the old object onto the stack (as a reference)
-   ret.prevObject = this;
-   ret.context = this.context;
-
-   // Return the newly-formed element set
-   return ret;
-   },
-
-   // Execute a callback for every element in the matched set.
-   // (You can seed the arguments with an array of args, but this is
-   // only used internally.)
-   each: function( callback, args ) {
-   return jQuery.each( this, callback, args );
-   },
-
-   map: function( callback ) {
-   return this.pushStack( jQuery.map(this, function( elem, i ) {
-   return 

[05/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/js/dagre-d3.js
--
diff --git a/content/visualizer/js/dagre-d3.js 
b/content/visualizer/js/dagre-d3.js
deleted file mode 100755
index 482ce82..000
--- a/content/visualizer/js/dagre-d3.js
+++ /dev/null
@@ -1,4560 +0,0 @@
-;(function e(t,n,r){function s(o,u){if(!n[o]){if(!t[o]){var a=typeof 
require=="function"&if(!u&)return a(o,!0);if(i)return i(o,!0);throw 
new Error("Cannot find module '"+o+"'")}var 
f=n[o]={exports:{}};t[o][0].call(f.exports,function(e){var n=t[o][1][e];return 
s(n?n:e)},f,f.exports,e,t,n,r)}return n[o].exports}var i=typeof 
require=="function"&for(var o=0;o

[07/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/css/graph.css
--
diff --git a/content/visualizer/css/graph.css b/content/visualizer/css/graph.css
deleted file mode 100755
index f4365cc..000
--- a/content/visualizer/css/graph.css
+++ /dev/null
@@ -1,48 +0,0 @@
-/*!
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-g.type-TK > rect {
-  fill: #00ffd0;
-}
-
-svg {
-  border: 0px solid #999;
-  overflow: hidden;
-}
-
-text {
-  font-weight: 300;
-  font-family: "Helvetica Neue", Helvetica, Arial, sans-serf;
-  font-size: 14px;
-}
-
-.node rect {
-  stroke: #999;
-  stroke-width: 0px;
-  fill: #fff;
-}
-
-.edgeLabel rect {
-  fill: #fff;
-}
-
-.edgePath path {
-  stroke: #333;
-  stroke-width: 3px;
-  fill: none;
-}

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/css/nephelefrontend.css
--
diff --git a/content/visualizer/css/nephelefrontend.css 
b/content/visualizer/css/nephelefrontend.css
deleted file mode 100755
index 3ca1837..000
--- a/content/visualizer/css/nephelefrontend.css
+++ /dev/null
@@ -1,198 +0,0 @@
-/*!
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-*
-{
-   padding: 0px;
-   margin: 0px;
-}
-
-html, body
-{
-   height: 100%;
-   position: relative;
-   overflow: auto;
-}
-
-h1
-{
-   font-family: verdana;
-   font-size: 36px;
-   font-weight: normal;
-   font-decoration: none;
-   font-style: normal;
-   font-variant: small-caps;
-}
-
-h1 img {
-   vertical-align: middle;
-   margin-right: 50px;
-}
-
-h3
-{
-   font-family: verdana;
-   font-size: 20px;
-   font-weight: normal;
-   font-decoration: none;
-   font-style: italic;
-   font-variant: normal;
-   text-align: center;
-   color: #66;
-   margin: 10px 10px 20px 10px;
-}
-
-h4
-{
-   font-family: Verdana;
-   font-size: 16px;
-   font-weight: normal;
-   font-decoration: none;
-   font-style: italic;
-   font-variant: normal;
-   text-align: left;
-   color: #00;
-   margin: 5px 5px 10px 5px;
-   padding: 2px 2px 5px 2px;
-   border-bottom: 1px dashed #44;
-}
-
-input
-{
-   font-family: verdana, Sans-Serif;
-   font-size: 13px;
-
-   border: solid 1.5px #33;
-   background-color: #ee;
-   color: #33;
-
-   padding: 5px;
-   margin: 5px;
-}
-
-.fadedPropertiesText
-{
-   font-family: verdana;
-   font-size: 18px;
-   color: #AA;
-   font-style: italic;
-}
-
-
-.mainHeading
-{
-   border: 1px solid #262A37;
-   margin: 5px;
-   text-align: left;
-   background-color: white;
-   background-repeat: no-repeat;
-   background-position: right;
-   height: 100px;
-   overflow: hidden;
-}
-
-
-.boxed
-{
-   border: 1px solid #262A37;
-   margin: 5px;
-}
-
-.spaced
-{
-   margin: 5px;
-}
-
-.footer
-{
-   position: absolute;
-   left: 0px;
-   right: 0px;
-   bottom: 0px;
-}
-
-.error_text
-{
-   font-family: verdana;
-   font-size: 14px;
-   font-weight: normal;
-   font-decoration: none;
-   font-style: italic;
-   font-variant: normal;
-   text-align: center;
-   color: #DF0101;
-}
-
-.translucent
-{
-   -moz-opacity:0 ;
-   

[01/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
Repository: flink-web
Updated Branches:
  refs/heads/asf-site 4bb81d3f1 -> 9ec0a879f


http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/js/program.js
--
diff --git a/content/visualizer/js/program.js b/content/visualizer/js/program.js
deleted file mode 100755
index 683185d..000
--- a/content/visualizer/js/program.js
+++ /dev/null
@@ -1,233 +0,0 @@
-/*!
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-var maxColumnWidth = 200;
-var minColumnWidth = 100;
-
-// global variable for the currently requested plan
-var pactPlanRequested = 0;
-
-/*
- * This function toggels the child checkbox on and of, depending on the 
parent's state
- */
-function toggleShowPlanBox(box)
-{
-  var child = $('#suspendJobDuringPlanCheck');
-  
-  if (box.is(':checked')) {
-child.attr('disabled', false);
-  }
-  else {
-child.attr('disabled', true);
-  }
-}
-
-/*
- * Shows an error message below the upload field.
- */
-function showUploadError(message)
-{
-  $('#upload_error_text').fadeOut("fast", function () { 
$('#upload_error_text')[0].innerHTML = "" + message;
-   
$('#upload_error_text').fadeIn("slow"); } );
-}
-
-/*
- * Checks the selected file and triggers an upload, if all is correct.
- */
-function processUpload()
-{
-
-  var filename = $('#upload_file_input').val();
-  var len = filename.length;
-  if (len == 0) {
-showUploadError("Please select a file.");
-  }
-  else if (len > 4 && filename.substr(len - 4, len) == ".jar") {
-$('#upload_form')[0].submit();
-  }
-  else {
-showUploadError("Please select a .jar file.");
-  }
-}
-
-/*
- * This function makes sure only one checkbox is selected.
- * Upon selection it initializes the drawing of the pact plan.
- * Upon deselection, it clears the pact plan.
- */
-function toggleCheckboxes(box)
-{
-
-  if (box.is(':checked')) {
-$('.jobItemCheckbox').attr('checked', false);
-box.attr('checked', true);
-var id = box.parentsUntil('.JobListItems').parent().attr('id').substr(4);
-var assemblerClass = box.attr('id');
-
-$('#mainCanvas').html('');
-pactPlanRequested = id;
-
-$.ajax({
-type: "GET",
-url: "pactPlan",
-data: { job: id, assemblerClass: assemblerClass},
-success: function(response) { showPreviewPlan(response); }
-});
-  }
-  else {
-$('#mainCanvas').html('');
-  }
-}
-
-/*
- * Function that takes the returned plan and draws it.
- */
-function showPreviewPlan(data)
-{
-   $("#mainCanvas").empty();
-var svgElement = "";
-$("#mainCanvas").append(svgElement);
-drawGraph(data.plan, "#svg-main");
-pactPlanRequested = 0;
-
-//activate zoom buttons
-activateZoomButtons();
-//  }
-}
-
-/*
- * Asynchronously loads the jobs and creates a list of them.
- */
-function loadJobList()
-{
-  $.get("jobs", { action: "list" }, createJobList);
-}
-
-/*
- * Triggers an AJAX request to delete a job.
- */
-function deleteJob(id)
-{
-  var name = id.substr(4);
-  $.get("jobs", { action: "delete", filename: name }, loadJobList);
-}
-
-/*
- * Creates and lists the returned jobs.
- */
-function createJobList(data)
-{
-  var markup = "";
-   
-  var lines = data.split("\n");
-  for (var i = 0; i < lines.length; i++)
-  {
-if (lines[i] == null || lines[i].length == 0) {
-  continue;
-}
-
-var date = "unknown date";
-var assemblerClass = "no entry class specified";
-
-var tokens = lines[i].split("\t");
-var name = tokens[0];
-if (tokens.length > 1) {
-  date = tokens[1];
-  if (tokens.length > 2) {
-assemblerClass = tokens[2];
-  }
-}
-
-var entries = assemblerClass.split("#");
-var classes = entries[0].split(",");
-
-markup += '';
-markup += '' + name + 
'';
-markup += '' + date + '';
-markup += '';
-
-var j = 0;
-for (var idx in classes) {
-  markup += '';
-  markup += '' + classes[idx] + '';
-}
-markup += '';
-  }
-  
-  // add the contents
-  $('#jobsContents').html(markup); 
-  
-  // register the event handler that triggers the delete when the 

[26/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/colors/flink_colors.pptx
--
diff --git a/content/img/logo/colors/flink_colors.pptx 
b/content/img/logo/colors/flink_colors.pptx
deleted file mode 100644
index 8ecd0be..000
Binary files a/content/img/logo/colors/flink_colors.pptx and /dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/100/flink_squirrel_100_black.png
--
diff --git a/content/img/logo/png/100/flink_squirrel_100_black.png 
b/content/img/logo/png/100/flink_squirrel_100_black.png
deleted file mode 100755
index 6869f58..000
Binary files a/content/img/logo/png/100/flink_squirrel_100_black.png and 
/dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/100/flink_squirrel_100_color.png
--
diff --git a/content/img/logo/png/100/flink_squirrel_100_color.png 
b/content/img/logo/png/100/flink_squirrel_100_color.png
deleted file mode 100755
index c508e1e..000
Binary files a/content/img/logo/png/100/flink_squirrel_100_color.png and 
/dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/100/flink_squirrel_100_white.png
--
diff --git a/content/img/logo/png/100/flink_squirrel_100_white.png 
b/content/img/logo/png/100/flink_squirrel_100_white.png
deleted file mode 100755
index 088fb27..000
Binary files a/content/img/logo/png/100/flink_squirrel_100_white.png and 
/dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink1000_black.png
--
diff --git a/content/img/logo/png/1000/flink1000_black.png 
b/content/img/logo/png/1000/flink1000_black.png
deleted file mode 100755
index 31af663..000
Binary files a/content/img/logo/png/1000/flink1000_black.png and /dev/null 
differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink1000_color_black.png
--
diff --git a/content/img/logo/png/1000/flink1000_color_black.png 
b/content/img/logo/png/1000/flink1000_color_black.png
deleted file mode 100755
index 2f4b991..000
Binary files a/content/img/logo/png/1000/flink1000_color_black.png and 
/dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink1000_color_white.png
--
diff --git a/content/img/logo/png/1000/flink1000_color_white.png 
b/content/img/logo/png/1000/flink1000_color_white.png
deleted file mode 100755
index cd0b56e..000
Binary files a/content/img/logo/png/1000/flink1000_color_white.png and 
/dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink1000_white.png
--
diff --git a/content/img/logo/png/1000/flink1000_white.png 
b/content/img/logo/png/1000/flink1000_white.png
deleted file mode 100755
index 3b7b253..000
Binary files a/content/img/logo/png/1000/flink1000_white.png and /dev/null 
differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink_squirrel_1000.png
--
diff --git a/content/img/logo/png/1000/flink_squirrel_1000.png 
b/content/img/logo/png/1000/flink_squirrel_1000.png
deleted file mode 100755
index 37c980e..000
Binary files a/content/img/logo/png/1000/flink_squirrel_1000.png and /dev/null 
differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink_squirrel_black_1000.png
--
diff --git a/content/img/logo/png/1000/flink_squirrel_black_1000.png 
b/content/img/logo/png/1000/flink_squirrel_black_1000.png
deleted file mode 100755
index b68621d..000
Binary files a/content/img/logo/png/1000/flink_squirrel_black_1000.png and 
/dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink_squirrel_white_1000.png
--
diff --git a/content/img/logo/png/1000/flink_squirrel_white_1000.png 
b/content/img/logo/png/1000/flink_squirrel_white_1000.png
deleted file mode 100755
index 14c7350..000
Binary files a/content/img/logo/png/1000/flink_squirrel_white_1000.png and 
/dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/200/flink2_200_black.png

[04/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/js/graphCreator.js
--
diff --git a/content/visualizer/js/graphCreator.js 
b/content/visualizer/js/graphCreator.js
deleted file mode 100755
index 86c8c81..000
--- a/content/visualizer/js/graphCreator.js
+++ /dev/null
@@ -1,445 +0,0 @@
-/*!
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-/*
- * Document initialization.
- */
-$(document).ready(function () 
-{
-   zoom = d3.behavior.zoom("#svg-main").on("zoom", function() {
-   var ev = d3.event;
-   d3.select("#svg-main g")
-   .attr("transform", "translate(" + ev.translate + ") 
scale(" + ev.scale + ")");
-   });
-   zoom.scaleExtent([0.3, 3])
-});
-
-//The current JSON file
-var JSONData; 
-//The informations for the iterations
-var iterationIds = new Array();
-var iterationGraphs = new Array();
-var iterationWidths = new Array();
-var iterationHeights = new Array();
-
-//The zoom element
-var zoom;
-//informations about the enviroment
-var svgWidth;
-var svgHeight;
-
-//Renders and draws the graph
-function drawGraph(data, svgID){
-   JSONData = data;
-   
-   //First step: precompute all iteration graphs
-   
-   //find all iterations
-   iterationNodes = searchForIterationNodes();
-   
-   //add the graphs of iterations and their sizes + Ids to arrays
-   if (iterationNodes != null) {
-   for (var i in iterationNodes) {
-   var itNode = iterationNodes[i];
-   iterationIds.push(itNode.id);
-   var g0 = loadJsonToDagre(itNode);
-   iterationGraphs.push(g0);
-   var r = new dagreD3.Renderer();
-   var l = dagreD3.layout()
-.nodeSep(20)
-.rankDir("LR");
-   l = r.layout(l).run(g0, d3.select("#svg-main"));
-   
-   iterationWidths.push(l._value.width);
-   iterationHeights.push(l._value.height);
-   
-   //Clean svg
-   $("#svg-main g").empty();
-   }
-   }
-   
-   //Continue normal
-   var g = loadJsonToDagre(data);
-   var selector = svgID + " g";
-   var renderer = new dagreD3.Renderer();
-   var layout = dagreD3.layout()
-   .nodeSep(20)
-   .rankDir("LR");
-   var svgElement = d3.select(selector);
-   layout = renderer.layout(layout).run(g, svgElement);
-   
-   svgHeight = layout._value.height;
-   svgWidth = layout._value.width;
-   
-var svg = d3.select("#svg-main")
-   .attr("width", $(document).width() - 15)
-   .attr("height", $(document).height() - 15 - 110)
-// .attr("viewBox", "0 0 "+ ($(document).width() - 150) +" 
"+($(document).height() - 15 - 110))
-   .call(zoom);
-   
-   // This should now draw the precomputed graphs in the svgs... . 
-   
-   for (var i in iterationIds) {
-   var workset = searchForNode(iterationIds[i]);
-   renderer = new dagreD3.Renderer();
-   layout = dagreD3.layout()
-.nodeSep(20)
-.rankDir("LR");
-   selector = "#svg-"+iterationIds[i]+" g";
-   svgElement = d3.select(selector);
-   layout = renderer.layout(layout).run(iterationGraphs[i], 
svgElement);
-   }
-   
-   //enable Overlays and register function for overlay-infos
-   $("a[rel]").overlay({ 
-   onBeforeLoad: function(){ 
-   var id = this.getTrigger().attr("nodeID")
-   showProperties(id);
-   }
-   });
-   
-   }
-
-//Creates the dagreD3 graph object
-//Responsible for adding nodes and edges
-function loadJsonToDagre(data){
-   
-   //stores all nodes that are in current graph -> no edges to nodes which 
are outside of current 

[77/78] [abbrv] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/blog/feed.xml
--
diff --git a/content/blog/feed.xml b/content/blog/feed.xml
new file mode 100644
index 000..066ae2f
--- /dev/null
+++ b/content/blog/feed.xml
@@ -0,0 +1,6041 @@
+
+http://www.w3.org/2005/Atom;>
+
+Flink Blog Feed
+Flink Blog
+http://flink.apache.org/blog
+http://flink.apache.org/blog/feed.xml; rel="self" 
type="application/rss+xml" />
+
+
+Apache Flink 1.1.4 Released
+pThe Apache Flink community released the next bugfix 
version of the Apache Flink 1.1 series./p
+
+pThis release includes major robustness improvements for checkpoint 
cleanup on failures and consumption of intermediate streams. We highly 
recommend all users to upgrade to Flink 1.1.4./p
+
+div class=highlightprecode 
class=language-xmlspan 
class=ntlt;dependencygt;/span
+  span 
class=ntlt;groupIdgt;/spanorg.apache.flinkspan
 class=ntlt;/groupIdgt;/span
+  span 
class=ntlt;artifactIdgt;/spanflink-javaspan
 class=ntlt;/artifactIdgt;/span
+  span 
class=ntlt;versiongt;/span1.1.4span 
class=ntlt;/versiongt;/span
+span class=ntlt;/dependencygt;/span
+span class=ntlt;dependencygt;/span
+  span 
class=ntlt;groupIdgt;/spanorg.apache.flinkspan
 class=ntlt;/groupIdgt;/span
+  span 
class=ntlt;artifactIdgt;/spanflink-streaming-java_2.10span
 class=ntlt;/artifactIdgt;/span
+  span 
class=ntlt;versiongt;/span1.1.4span 
class=ntlt;/versiongt;/span
+span class=ntlt;/dependencygt;/span
+span class=ntlt;dependencygt;/span
+  span 
class=ntlt;groupIdgt;/spanorg.apache.flinkspan
 class=ntlt;/groupIdgt;/span
+  span 
class=ntlt;artifactIdgt;/spanflink-clients_2.10span
 class=ntlt;/artifactIdgt;/span
+  span 
class=ntlt;versiongt;/span1.1.4span 
class=ntlt;/versiongt;/span
+span 
class=ntlt;/dependencygt;/span/code/pre/div
+
+pYou can find the binaries on the updated a 
href=http://flink.apache.org/downloads.htmlDownloads 
page/a./p
+
+h2 id=note-for-rocksdb-backend-usersNote for RocksDB 
Backend Users/h2
+
+pWe updated Flink’s RocksDB dependency version from 
code4.5.1/code to code4.11.2/code. Between 
these versions some of RocksDB’s internal configuration defaults changed that 
would affect the memory footprint of running Flink with RocksDB. Therefore, we 
manually reset them to the previous defaults. If you want to run with the new 
Rocks 4.11.2 defaults, you can do this via:/p
+
+div class=highlightprecode 
class=language-javaspan 
class=nRocksDBStateBackend/span span 
class=nbackend/span span 
class=o=/span span 
class=knew/span span 
class=nfRocksDBStateBackend/spanspan 
class=o(/spanspan 
class=squot;...quot;/spanspan 
class=o);/span
+span class=c1// Use the new default options. Otherwise, 
the default for RocksDB 4.5.1/span
+span class=c1// `PredefinedOptions.DEFAULT_ROCKS_4_5_1` 
will be used./span
+span class=nbackend/spanspan 
class=o./spanspan 
class=nasetPredefinedOptions/spanspan 
class=o(/spanspan 
class=nPredefinedOptions/spanspan 
class=o./spanspan 
class=naDEFAULT/spanspan 
class=o);/span/code/pre/div
+
+h2 id=release-notes---flink---version-114Release Notes - 
Flink - Version 1.1.4/h2
+
+h3 id=sub-taskSub-task/h3
+ul
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4510FLINK-4510/a;]
 - Always create CheckpointCoordinator
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4984FLINK-4984/a;]
 - Add Cancellation Barriers to BarrierTracker and BarrierBuffer
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4985FLINK-4985/a;]
 - Report Declined/Canceled Checkpoints to Checkpoint Coordinator
+/li
+/ul
+
+h3 id=bugBug/h3
+ul
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-2662FLINK-2662/a;]
 - CompilerException: quot;Bug: Plan generation for Unions picked 
a ship strategy between binary plan operators.quot;
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-3680FLINK-3680/a;]
 - Remove or improve (not set) text in the Job Plan UI
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-3813FLINK-3813/a;]
 - YARNSessionFIFOITCase.testDetachedMode failed on Travis
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4108FLINK-4108/a;]
 - NPE in Row.productArity
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4506FLINK-4506/a;]
 - CsvOutputFormat defaults allowNullValues to false, even though doc 
and declaration says true
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4581FLINK-4581/a;]
 - Table API throws quot;No suitable driver found for 
jdbc:calcitequot;
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4586FLINK-4586/a;]
 - NumberSequenceIterator and Accumulator threading issue
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4619FLINK-4619/a;]
 - JobManager does not answer to client when restore from savepoint 
fails
+/li
+li[a 

[44/78] [abbrv] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/js/dagre-d3.js
--
diff --git a/content/visualizer/js/dagre-d3.js 
b/content/visualizer/js/dagre-d3.js
new file mode 100755
index 000..482ce82
--- /dev/null
+++ b/content/visualizer/js/dagre-d3.js
@@ -0,0 +1,4560 @@
+;(function e(t,n,r){function s(o,u){if(!n[o]){if(!t[o]){var a=typeof 
require=="function"&if(!u&)return a(o,!0);if(i)return i(o,!0);throw 
new Error("Cannot find module '"+o+"'")}var 
f=n[o]={exports:{}};t[o][0].call(f.exports,function(e){var n=t[o][1][e];return 
s(n?n:e)},f,f.exports,e,t,n,r)}return n[o].exports}var i=typeof 
require=="function"&for(var o=0;o

[70/78] [abbrv] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/how-to-contribute.html
--
diff --git a/content/how-to-contribute.html b/content/how-to-contribute.html
new file mode 100644
index 000..87ad4a0
--- /dev/null
+++ b/content/how-to-contribute.html
@@ -0,0 +1,345 @@
+
+
+  
+
+
+
+
+Apache Flink: How To Contribute
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to 
Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+How To Contribute
+
+   Apache Flink is developed by an open and friendly community. 
Everybody is cordially welcome to join the community and contribute to Apache 
Flink. There are several ways to interact with the community and to contribute 
to Flink including asking questions, filing bug reports, proposing new 
features, joining discussions on the mailing lists, contributing code or 
documentation, improving the website, or testing release candidates.
+
+
+
+  Ask 
questions!
+  File a 
bug report
+  Propose an 
improvement or a new feature
+  Help others and join the 
discussions
+  Test a release candidate
+  Contribute 
code
+  Contribute documentation
+  Improve the website
+  More ways to contribute…
+  Submit a Contributor 
License Agreement
+  How to become a committer
+
+
+
+
+Ask questions!
+
+The Apache Flink community is eager to help and to answer your questions. 
We have a user mailing list, hang 
out in an IRC channel, and watch Stack 
Overflow on the http://stackoverflow.com/questions/tagged/apache-flink;>[apache-flink]
 tag.
+
+
+
+File a bug report
+
+Please let us know if you experienced a problem with Flink and file a bug 
report. Open http://issues.apache.org/jira/browse/FLINK;>Flink’s 
JIRA and click on the blue Create button at the top. 
Please give detailed information about the problem you encountered and, if 
possible, add a description that helps to reproduce the problem. Thank you very 
much.
+
+
+
+Propose an improvement or a 
new feature
+
+Our community is constantly looking for feedback to improve Apache Flink. 
If you have an idea how to improve Flink or have a new feature in mind that 
would be beneficial for Flink users, please open an issue in http://issues.apache.org/jira/browse/FLINK;>Flink’s JIRA. The 
improvement or new feature should be described in appropriate detail and 
include the scope and its requirements if possible. Detailed information is 
important for a few reasons:
+
+
+  It ensures your requirements are met when the improvement or feature is 
implemented.
+  It helps to estimate the effort and to design a solution that addresses 
your needs.
+  It allow for constructive discussions that might arise around this 
issue.
+
+
+Detailed information is also required, if you plan to contribute the 
improvement or feature you proposed yourself. Please read the Contribute code guide in this case as well.
+
+We recommend to first reach consensus with the community on whether a new 
feature is required and how to implement a new feature, before starting with 
the implementation. Some features might be out of scope of the project, and 
it’s best to 

[17/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/news/2015/03/02/february-2015-in-flink.html
--
diff --git a/content/news/2015/03/02/february-2015-in-flink.html 
b/content/news/2015/03/02/february-2015-in-flink.html
deleted file mode 100644
index a32ba35..000
--- a/content/news/2015/03/02/february-2015-in-flink.html
+++ /dev/null
@@ -1,308 +0,0 @@
-
-
-  
-
-
-
-
-Apache Flink: February 2015 in the Flink community
-
-
-
-
-https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
-
-
-
-
-
-
-
-
-https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
-
-
-
-
-  
-
-
-
-
-
-
-
-  
- 
-  
-
-
-
-  
-
-
-
-  
-  
-
-  
-
-  
-
-
-
-
-  
-
-
-Download Flink
-
-
-Home
-
-
-Introduction to Flink
-
-
-Flink Use Cases
-
-
-Powered by Flink
-
-
-Ecosystem
-
-
-Community  Project Info
-
-
-How to Contribute
-
-
-Flink 
Blog
-
-
-
-
-
-
-
-
-  Documentation
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
-
-  
-
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
-
-
-
-
-  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
-
-
-
-
-
-
-
-  
-
-
-
-  
-  
-
-
-Project FAQ
-
-
-https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
-
-
-Plan Visualizer 
-
-  
-
-
-
-  
-  
-  
-  
-
-  February 2015 in the Flink community
-
-  
-02 Mar 2015
-
-February might be the shortest month of the year, but this does not
-mean that the Flink community has not been busy adding features to the
-system and fixing bugs. Here’s a rundown of the activity in the Flink
-community last month.
-
-0.8.1 release
-
-Flink 0.8.1 was released. This bugfixing release resolves a total of 22 
issues.
-
-New committer
-
-https://github.com/mxm;>Max Michels has been voted a committer 
by the Flink PMC.
-
-Flink adapter for Apache SAMOA
-
-http://samoa.incubator.apache.org;>Apache SAMOA (incubating) 
is a
-distributed streaming machine learning (ML) framework with a
-programming abstraction for distributed streaming ML algorithms. SAMOA
-runs on a variety of backend engines, currently Apache Storm and
-Apache S4.  A https://github.com/apache/incubator-samoa/pull/11;>pull
-request is
-available at the SAMOA repository that adds a Flink adapter for SAMOA.
-
-Easy Flink deployment 
on Google Compute Cloud
-
-Flink is now integrated in bdutil, Google’s open source tool for
-creating and configuring (Hadoop) clusters in Google Compute
-Engine. Deployment of Flink clusters in now supported starting with
-https://groups.google.com/forum/#!topic/gcp-hadoop-announce/uVJ_6y9cGKM;>bdutil
-1.2.0.
-
-Flink on the Web
-
-A new blog post on http://flink.apache.org/news/2015/02/09/streaming-example.html;>Flink
-Streaming
-was published at the blog. Flink was mentioned in several articles on
-the web. Here are some examples:
-
-
-  
-http://dataconomy.com/how-flink-became-an-apache-top-level-project/;>How 
Flink became an Apache Top-Level Project
-  
-  
-https://www.linkedin.com/pulse/stale-synchronous-parallelism-new-frontier-apache-flink-nam-luc-tran?utm_content=buffer461afutm_medium=socialutm_source=linkedin.comutm_campaign=buffer;>Stale
 Synchronous Parallelism: The new frontier for Apache Flink?
-  
-  
-http://www.hadoopsphere.com/2015/02/distributed-data-processing-with-apache.html;>Distributed
 data processing with Apache Flink
-  
-  
-http://www.hadoopsphere.com/2015/02/ciao-latency-hallo-speed.html;>Ciao 
latency, hello speed
-  
-
-
-In the Flink master
-
-The following features have been now merged in Flink’s master 
repository.
-
-Gelly
-
-Gelly, Flink’s Graph API allows users to manipulate graph-shaped data
-directly. Here’s for example a calculation of shortest paths in a
-graph:
-
-GraphLong, Double, Double graph 
= Graph.fromDataSet(vertices, edges, env);
-

[06/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/js/d3.js
--
diff --git a/content/visualizer/js/d3.js b/content/visualizer/js/d3.js
deleted file mode 100755
index 2d09bc9..000
--- a/content/visualizer/js/d3.js
+++ /dev/null
@@ -1,9270 +0,0 @@
-/**
- * D3 renderer from http://d3js.org/
- * Copyright (c) 2010-2014, Michael Bostock
- * 
- * Under BSD License (as stated under http://d3js.org/)
- */
-
-!function() {
-  var d3 = {
-version: "3.4.6"
-  };
-  if (!Date.now) Date.now = function() {
-return +new Date();
-  };
-  var d3_arraySlice = [].slice, d3_array = function(list) {
-return d3_arraySlice.call(list);
-  };
-  var d3_document = document, d3_documentElement = 
d3_document.documentElement, d3_window = window;
-  try {
-d3_array(d3_documentElement.childNodes)[0].nodeType;
-  } catch (e) {
-d3_array = function(list) {
-  var i = list.length, array = new Array(i);
-  while (i--) array[i] = list[i];
-  return array;
-};
-  }
-  try {
-d3_document.createElement("div").style.setProperty("opacity", 0, "");
-  } catch (error) {
-var d3_element_prototype = d3_window.Element.prototype, 
d3_element_setAttribute = d3_element_prototype.setAttribute, 
d3_element_setAttributeNS = d3_element_prototype.setAttributeNS, 
d3_style_prototype = d3_window.CSSStyleDeclaration.prototype, 
d3_style_setProperty = d3_style_prototype.setProperty;
-d3_element_prototype.setAttribute = function(name, value) {
-  d3_element_setAttribute.call(this, name, value + "");
-};
-d3_element_prototype.setAttributeNS = function(space, local, value) {
-  d3_element_setAttributeNS.call(this, space, local, value + "");
-};
-d3_style_prototype.setProperty = function(name, value, priority) {
-  d3_style_setProperty.call(this, name, value + "", priority);
-};
-  }
-  d3.ascending = d3_ascending;
-  function d3_ascending(a, b) {
-return a < b ? -1 : a > b ? 1 : a >= b ? 0 : NaN;
-  }
-  d3.descending = function(a, b) {
-return b < a ? -1 : b > a ? 1 : b >= a ? 0 : NaN;
-  };
-  d3.min = function(array, f) {
-var i = -1, n = array.length, a, b;
-if (arguments.length === 1) {
-  while (++i < n && !((a = array[i]) != null && a <= a)) a = undefined;
-  while (++i < n) if ((b = array[i]) != null && a > b) a = b;
-} else {
-  while (++i < n && !((a = f.call(array, array[i], i)) != null && a <= a)) 
a = undefined;
-  while (++i < n) if ((b = f.call(array, array[i], i)) != null && a > b) a 
= b;
-}
-return a;
-  };
-  d3.max = function(array, f) {
-var i = -1, n = array.length, a, b;
-if (arguments.length === 1) {
-  while (++i < n && !((a = array[i]) != null && a <= a)) a = undefined;
-  while (++i < n) if ((b = array[i]) != null && b > a) a = b;
-} else {
-  while (++i < n && !((a = f.call(array, array[i], i)) != null && a <= a)) 
a = undefined;
-  while (++i < n) if ((b = f.call(array, array[i], i)) != null && b > a) a 
= b;
-}
-return a;
-  };
-  d3.extent = function(array, f) {
-var i = -1, n = array.length, a, b, c;
-if (arguments.length === 1) {
-  while (++i < n && !((a = c = array[i]) != null && a <= a)) a = c = 
undefined;
-  while (++i < n) if ((b = array[i]) != null) {
-if (a > b) a = b;
-if (c < b) c = b;
-  }
-} else {
-  while (++i < n && !((a = c = f.call(array, array[i], i)) != null && a <= 
a)) a = undefined;
-  while (++i < n) if ((b = f.call(array, array[i], i)) != null) {
-if (a > b) a = b;
-if (c < b) c = b;
-  }
-}
-return [ a, c ];
-  };
-  d3.sum = function(array, f) {
-var s = 0, n = array.length, a, i = -1;
-if (arguments.length === 1) {
-  while (++i < n) if (!isNaN(a = +array[i])) s += a;
-} else {
-  while (++i < n) if (!isNaN(a = +f.call(array, array[i], i))) s += a;
-}
-return s;
-  };
-  function d3_number(x) {
-return x != null && !isNaN(x);
-  }
-  d3.mean = function(array, f) {
-var s = 0, n = array.length, a, i = -1, j = n;
-if (arguments.length === 1) {
-  while (++i < n) if (d3_number(a = array[i])) s += a; else --j;
-} else {
-  while (++i < n) if (d3_number(a = f.call(array, array[i], i))) s += a; 
else --j;
-}
-return j ? s / j : undefined;
-  };
-  d3.quantile = function(values, p) {
-var H = (values.length - 1) * p + 1, h = Math.floor(H), v = +values[h - 
1], e = H - h;
-return e ? v + e * (values[h] - v) : v;
-  };
-  d3.median = function(array, f) {
-if (arguments.length > 1) array = array.map(f);
-array = array.filter(d3_number);
-return array.length ? d3.quantile(array.sort(d3_ascending), .5) : 
undefined;
-  };
-  function d3_bisector(compare) {
-return {
-  left: function(a, x, lo, hi) {
-if (arguments.length < 3) lo = 0;
-if (arguments.length < 4) hi = a.length;
-

[32/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/faq.html
--
diff --git a/content/faq.html b/content/faq.html
deleted file mode 100644
index 7653ff8..000
--- a/content/faq.html
+++ /dev/null
@@ -1,683 +0,0 @@
-
-
-  
-
-
-
-
-Apache Flink: Frequently Asked Questions (FAQ)
-
-
-
-
-https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
-
-
-
-
-
-
-
-
-https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
-
-
-
-
-  
-
-
-
-
-
-
-
-  
- 
-  
-
-
-
-  
-
-
-
-  
-  
-
-  
-
-  
-
-
-
-
-  
-
-
-Download Flink
-
-
-Home
-
-
-Introduction to Flink
-
-
-Flink Use Cases
-
-
-Powered by Flink
-
-
-Ecosystem
-
-
-Community  Project Info
-
-
-How to Contribute
-
-
-Flink 
Blog
-
-
-
-
-
-
-
-
-  Documentation
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
-
-  
-
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
-
-
-
-
-  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
-
-
-
-
-
-
-
-  
-
-
-
-  
-  
-
-
-Project 
FAQ
-
-
-https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
-
-
-Plan Visualizer 
-
-  
-
-
-
-  
-  
-  
-  
-Frequently Asked Questions (FAQ)
-
-   
-
-The following questions are frequently asked with regard to the Flink 
project in general. If you have further questions, make sure 
to consult the documentation or ask the 
community.
-
-
-
-  General
-  Is Flink a Hadoop Project?
-  Do I have to 
install Apache Hadoop to use Flink?
-
-  
-  Usage
-  How do I 
assess the progress of a Flink program?
-  How can I figure 
out why a program failed?
-  How do I debug Flink 
programs?
-  What is the 
parallelism? How do I set it?
-
-  
-  Errors
-  Why am I getting 
a “NonSerializableException” ?
-  In
 Scala API, I get an error about implicit values and evidence 
parameters
-  I
 get an error message saying that not enough buffers are available. How do I 
fix this?
-  My
 job fails early with a java.io.EOFException. What could be the cause?
-  My
 job fails with various exceptions from the HDFS/Hadoop code. What can I 
do?
-  In 
Eclipse, I get compilation errors in the Scala projects
-  My
 program does not compute the correct result. Why are my custom key 
types
-  I
 get a java.lang.InstantiationException for my data type, what is 
wrong?
-  I
 can’t stop Flink with the provided stop-scripts. What can I do?
-  I got an 
OutOfMemoryException. What can I do?
-  Why do the 
TaskManager log files become so huge?
-  The
 slot allocated for my task manager has been released. What should I 
do?
-
-  
-  YARN 
Deployment
-  The YARN session 
runs only for a few seconds
-  My
 YARN containers are killed because they use too much memory
-  The
 YARN session crashes with a HDFS permission exception during startup
-  My job is not 
reacting to a job cancellation?
-
-  
-  Features
-  What kind of 
fault-tolerance does Flink provide?
-  Are
 Hadoop-like utilities, such as Counters and the DistributedCache 
supported?
-
-  
-
-
-
-
-General
-
-Is Flink a Hadoop Project?
-
-Flink is a data processing system and an alternative to Hadoop’s
-MapReduce component. It comes with its own runtime rather 
than building on top
-of MapReduce. As such, it can work completely independently of the Hadoop
-ecosystem. However, Flink can also access Hadoop’s distributed file
-system (HDFS) to read and write data, and Hadoop’s next-generation resource
-manager (YARN) to provision cluster resources. Since most Flink users are
-using Hadoop HDFS to store their data, Flink already ships the required 
libraries to
-access HDFS.
-
-Do I have to install 
Apache Hadoop to use Flink?
-
-No. Flink can run without a Hadoop 
installation. However, a 

[14/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/news/2015/09/16/off-heap-memory.html
--
diff --git a/content/news/2015/09/16/off-heap-memory.html 
b/content/news/2015/09/16/off-heap-memory.html
deleted file mode 100644
index 60d7bda..000
--- a/content/news/2015/09/16/off-heap-memory.html
+++ /dev/null
@@ -1,1083 +0,0 @@
-
-
-  
-
-
-
-
-Apache Flink: Off-heap Memory in Apache Flink and the curious JIT 
compiler
-
-
-
-
-https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
-
-
-
-
-
-
-
-
-https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
-
-
-
-
-  
-
-
-
-
-
-
-
-  
- 
-  
-
-
-
-  
-
-
-
-  
-  
-
-  
-
-  
-
-
-
-
-  
-
-
-Download Flink
-
-
-Home
-
-
-Introduction to Flink
-
-
-Flink Use Cases
-
-
-Powered by Flink
-
-
-Ecosystem
-
-
-Community  Project Info
-
-
-How to Contribute
-
-
-Flink 
Blog
-
-
-
-
-
-
-
-
-  Documentation
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
-
-  
-
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
-
-
-
-
-  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
-
-
-
-
-
-
-
-  
-
-
-
-  
-  
-
-
-Project FAQ
-
-
-https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
-
-
-Plan Visualizer 
-
-  
-
-
-
-  
-  
-  
-  
-
-  Off-heap Memory in Apache Flink and the curious JIT compiler
-
-  
-16 Sep 2015 by Stephan Ewen (https://twitter.com/stephanewen;>@stephanewen)
-
-Running data-intensive code in the JVM and making it well-behaved is 
tricky. Systems that put billions of data objects naively onto the JVM heap 
face unpredictable OutOfMemoryErrors and Garbage Collection stalls. Of course, 
you still want to to keep your data in memory as much as possible, for speed 
and responsiveness of the processing applications. In that context, 
“off-heap” has become almost something like a magic word to solve these 
problems.
-
-In this blog post, we will look at how Flink exploits off-heap memory. The 
feature is part of the upcoming release, but you can try it out with the latest 
nightly builds. We will also give a few interesting insights into the behavior 
for Java’s JIT compiler for highly optimized methods and loops.
-
-Recap: Memory Management in 
Flink
-
-To understand Flink’s approach to off-heap memory, we need to recap 
Flink’s approach to custom managed memory. We have written an earlier blog post 
about how Flink manages JVM memory itself
-
-As a summary, the core part is that Flink implements its algorithms not 
against Java objects, arrays, or lists, but actually against a data structure 
similar to java.nio.ByteBuffer. Flink uses its own specialized 
version, called https://github.com/apache/flink/blob/release-0.9.1-rc1/flink-core/src/main/java/org/apache/flink/core/memory/MemorySegment.java;>MemorySegment
 on which algorithms put and get at specific positions ints, longs, byte 
arrays, etc, and compare and copy memory. The memory segments are held and 
distributed by a central component (called MemoryManager) from 
which algorithms request segments according to their calculated memory 
budgets.
-
-Don’t believe that this can be fast? Have a look at the benchmarks in the 
earlier blogpost, which show that it is actually often much faster than 
working on objects, due to better control over data layout (cache efficiency, 
data size), and reducing the pressure on Java’s Garbage Collector.
-
-This form of memory management has been in Flink for a long time. 
Anecdotally, the first public demo of Flink’s predecessor project 
Stratosphere, at the VLDB conference in 2010, was running its programs 
with custom managed memory (although I believe few attendees were aware of 
that).
-
-Why actually bother with 
off-heap memory?
-
-Given that Flink has a sophisticated level of managing on-heap 

[08/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/css/bootstrap.css
--
diff --git a/content/visualizer/css/bootstrap.css 
b/content/visualizer/css/bootstrap.css
deleted file mode 100755
index fc41022..000
--- a/content/visualizer/css/bootstrap.css
+++ /dev/null
@@ -1,5785 +0,0 @@
-/*!
- * Bootstrap v3.1.1 (http://getbootstrap.com)
- * Copyright 2011-2014 Twitter, Inc.
- * Licensed under MIT (https://github.com/twbs/bootstrap/blob/master/LICENSE)
- */
-
-/*! normalize.css v3.0.0 | MIT License | git.io/normalize */
-html {
-  font-family: sans-serif;
-  -webkit-text-size-adjust: 100%;
-  -ms-text-size-adjust: 100%;
-}
-body {
-  margin: 0;
-}
-article,
-aside,
-details,
-figcaption,
-figure,
-footer,
-header,
-hgroup,
-main,
-nav,
-section,
-summary {
-  display: block;
-}
-audio,
-canvas,
-progress,
-video {
-  display: inline-block;
-  vertical-align: baseline;
-}
-audio:not([controls]) {
-  display: none;
-  height: 0;
-}
-[hidden],
-template {
-  display: none;
-}
-a {
-  background: transparent;
-}
-a:active,
-a:hover {
-  outline: 0;
-}
-abbr[title] {
-  border-bottom: 1px dotted;
-}
-b,
-strong {
-  font-weight: bold;
-}
-dfn {
-  font-style: italic;
-}
-h1 {
-  margin: .67em 0;
-  font-size: 2em;
-}
-mark {
-  color: #000;
-  background: #ff0;
-}
-small {
-  font-size: 80%;
-}
-sub,
-sup {
-  position: relative;
-  font-size: 75%;
-  line-height: 0;
-  vertical-align: baseline;
-}
-sup {
-  top: -.5em;
-}
-sub {
-  bottom: -.25em;
-}
-img {
-  border: 0;
-}
-svg:not(:root) {
-  overflow: hidden;
-}
-figure {
-  margin: 1em 40px;
-}
-hr {
-  height: 0;
-  -moz-box-sizing: content-box;
-   box-sizing: content-box;
-}
-pre {
-  overflow: auto;
-}
-code,
-kbd,
-pre,
-samp {
-  font-family: monospace, monospace;
-  font-size: 1em;
-}
-button,
-input,
-optgroup,
-select,
-textarea {
-  margin: 0;
-  font: inherit;
-  color: inherit;
-}
-button {
-  overflow: visible;
-}
-button,
-select {
-  text-transform: none;
-}
-button,
-html input[type="button"],
-input[type="reset"],
-input[type="submit"] {
-  -webkit-appearance: button;
-  cursor: pointer;
-}
-button[disabled],
-html input[disabled] {
-  cursor: default;
-}
-button::-moz-focus-inner,
-input::-moz-focus-inner {
-  padding: 0;
-  border: 0;
-}
-input {
-  line-height: normal;
-}
-input[type="checkbox"],
-input[type="radio"] {
-  box-sizing: border-box;
-  padding: 0;
-}
-input[type="number"]::-webkit-inner-spin-button,
-input[type="number"]::-webkit-outer-spin-button {
-  height: auto;
-}
-input[type="search"] {
-  -webkit-box-sizing: content-box;
- -moz-box-sizing: content-box;
-  box-sizing: content-box;
-  -webkit-appearance: textfield;
-}
-input[type="search"]::-webkit-search-cancel-button,
-input[type="search"]::-webkit-search-decoration {
-  -webkit-appearance: none;
-}
-fieldset {
-  padding: .35em .625em .75em;
-  margin: 0 2px;
-  border: 1px solid #c0c0c0;
-}
-legend {
-  padding: 0;
-  border: 0;
-}
-textarea {
-  overflow: auto;
-}
-optgroup {
-  font-weight: bold;
-}
-table {
-  border-spacing: 0;
-  border-collapse: collapse;
-}
-td,
-th {
-  padding: 0;
-}
-@media print {
-  * {
-color: #000 !important;
-text-shadow: none !important;
-background: transparent !important;
-box-shadow: none !important;
-  }
-  a,
-  a:visited {
-text-decoration: underline;
-  }
-  a[href]:after {
-content: " (" attr(href) ")";
-  }
-  abbr[title]:after {
-content: " (" attr(title) ")";
-  }
-  a[href^="javascript:"]:after,
-  a[href^="#"]:after {
-content: "";
-  }
-  pre,
-  blockquote {
-border: 1px solid #999;
-
-page-break-inside: avoid;
-  }
-  thead {
-display: table-header-group;
-  }
-  tr,
-  img {
-page-break-inside: avoid;
-  }
-  img {
-max-width: 100% !important;
-  }
-  p,
-  h2,
-  h3 {
-orphans: 3;
-widows: 3;
-  }
-  h2,
-  h3 {
-page-break-after: avoid;
-  }
-  select {
-background: #fff !important;
-  }
-  .navbar {
-display: none;
-  }
-  .table td,
-  .table th {
-background-color: #fff !important;
-  }
-  .btn > .caret,
-  .dropup > .btn > .caret {
-border-top-color: #000 !important;
-  }
-  .label {
-border: 1px solid #000;
-  }
-  .table {
-border-collapse: collapse !important;
-  }
-  .table-bordered th,
-  .table-bordered td {
-border: 1px solid #ddd !important;
-  }
-}
-* {
-  -webkit-box-sizing: border-box;
- -moz-box-sizing: border-box;
-  box-sizing: border-box;
-}
-*:before,
-*:after {
-  -webkit-box-sizing: border-box;
- -moz-box-sizing: border-box;
-  box-sizing: border-box;
-}
-html {
-  font-size: 62.5%;
-
-  -webkit-tap-highlight-color: rgba(0, 0, 0, 0);
-}
-body {
-  font-family: "Helvetica Neue", Helvetica, Arial, sans-serif;
-  font-size: 14px;
-  line-height: 1.42857143;
-  color: #333;
-  background-color: #fff;
-}
-input,
-button,
-select,
-textarea {
-  font-family: inherit;
- 

[47/78] [abbrv] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/css/bootstrap.css
--
diff --git a/content/visualizer/css/bootstrap.css 
b/content/visualizer/css/bootstrap.css
new file mode 100755
index 000..fc41022
--- /dev/null
+++ b/content/visualizer/css/bootstrap.css
@@ -0,0 +1,5785 @@
+/*!
+ * Bootstrap v3.1.1 (http://getbootstrap.com)
+ * Copyright 2011-2014 Twitter, Inc.
+ * Licensed under MIT (https://github.com/twbs/bootstrap/blob/master/LICENSE)
+ */
+
+/*! normalize.css v3.0.0 | MIT License | git.io/normalize */
+html {
+  font-family: sans-serif;
+  -webkit-text-size-adjust: 100%;
+  -ms-text-size-adjust: 100%;
+}
+body {
+  margin: 0;
+}
+article,
+aside,
+details,
+figcaption,
+figure,
+footer,
+header,
+hgroup,
+main,
+nav,
+section,
+summary {
+  display: block;
+}
+audio,
+canvas,
+progress,
+video {
+  display: inline-block;
+  vertical-align: baseline;
+}
+audio:not([controls]) {
+  display: none;
+  height: 0;
+}
+[hidden],
+template {
+  display: none;
+}
+a {
+  background: transparent;
+}
+a:active,
+a:hover {
+  outline: 0;
+}
+abbr[title] {
+  border-bottom: 1px dotted;
+}
+b,
+strong {
+  font-weight: bold;
+}
+dfn {
+  font-style: italic;
+}
+h1 {
+  margin: .67em 0;
+  font-size: 2em;
+}
+mark {
+  color: #000;
+  background: #ff0;
+}
+small {
+  font-size: 80%;
+}
+sub,
+sup {
+  position: relative;
+  font-size: 75%;
+  line-height: 0;
+  vertical-align: baseline;
+}
+sup {
+  top: -.5em;
+}
+sub {
+  bottom: -.25em;
+}
+img {
+  border: 0;
+}
+svg:not(:root) {
+  overflow: hidden;
+}
+figure {
+  margin: 1em 40px;
+}
+hr {
+  height: 0;
+  -moz-box-sizing: content-box;
+   box-sizing: content-box;
+}
+pre {
+  overflow: auto;
+}
+code,
+kbd,
+pre,
+samp {
+  font-family: monospace, monospace;
+  font-size: 1em;
+}
+button,
+input,
+optgroup,
+select,
+textarea {
+  margin: 0;
+  font: inherit;
+  color: inherit;
+}
+button {
+  overflow: visible;
+}
+button,
+select {
+  text-transform: none;
+}
+button,
+html input[type="button"],
+input[type="reset"],
+input[type="submit"] {
+  -webkit-appearance: button;
+  cursor: pointer;
+}
+button[disabled],
+html input[disabled] {
+  cursor: default;
+}
+button::-moz-focus-inner,
+input::-moz-focus-inner {
+  padding: 0;
+  border: 0;
+}
+input {
+  line-height: normal;
+}
+input[type="checkbox"],
+input[type="radio"] {
+  box-sizing: border-box;
+  padding: 0;
+}
+input[type="number"]::-webkit-inner-spin-button,
+input[type="number"]::-webkit-outer-spin-button {
+  height: auto;
+}
+input[type="search"] {
+  -webkit-box-sizing: content-box;
+ -moz-box-sizing: content-box;
+  box-sizing: content-box;
+  -webkit-appearance: textfield;
+}
+input[type="search"]::-webkit-search-cancel-button,
+input[type="search"]::-webkit-search-decoration {
+  -webkit-appearance: none;
+}
+fieldset {
+  padding: .35em .625em .75em;
+  margin: 0 2px;
+  border: 1px solid #c0c0c0;
+}
+legend {
+  padding: 0;
+  border: 0;
+}
+textarea {
+  overflow: auto;
+}
+optgroup {
+  font-weight: bold;
+}
+table {
+  border-spacing: 0;
+  border-collapse: collapse;
+}
+td,
+th {
+  padding: 0;
+}
+@media print {
+  * {
+color: #000 !important;
+text-shadow: none !important;
+background: transparent !important;
+box-shadow: none !important;
+  }
+  a,
+  a:visited {
+text-decoration: underline;
+  }
+  a[href]:after {
+content: " (" attr(href) ")";
+  }
+  abbr[title]:after {
+content: " (" attr(title) ")";
+  }
+  a[href^="javascript:"]:after,
+  a[href^="#"]:after {
+content: "";
+  }
+  pre,
+  blockquote {
+border: 1px solid #999;
+
+page-break-inside: avoid;
+  }
+  thead {
+display: table-header-group;
+  }
+  tr,
+  img {
+page-break-inside: avoid;
+  }
+  img {
+max-width: 100% !important;
+  }
+  p,
+  h2,
+  h3 {
+orphans: 3;
+widows: 3;
+  }
+  h2,
+  h3 {
+page-break-after: avoid;
+  }
+  select {
+background: #fff !important;
+  }
+  .navbar {
+display: none;
+  }
+  .table td,
+  .table th {
+background-color: #fff !important;
+  }
+  .btn > .caret,
+  .dropup > .btn > .caret {
+border-top-color: #000 !important;
+  }
+  .label {
+border: 1px solid #000;
+  }
+  .table {
+border-collapse: collapse !important;
+  }
+  .table-bordered th,
+  .table-bordered td {
+border: 1px solid #ddd !important;
+  }
+}
+* {
+  -webkit-box-sizing: border-box;
+ -moz-box-sizing: border-box;
+  box-sizing: border-box;
+}
+*:before,
+*:after {
+  -webkit-box-sizing: border-box;
+ -moz-box-sizing: border-box;
+  box-sizing: border-box;
+}
+html {
+  font-size: 62.5%;
+
+  -webkit-tap-highlight-color: rgba(0, 0, 0, 0);
+}
+body {
+  font-family: "Helvetica Neue", Helvetica, Arial, sans-serif;
+  font-size: 14px;
+  line-height: 1.42857143;
+  color: #333;
+  background-color: #fff;
+}
+input,
+button,
+select,
+textarea {
+  font-family: inherit;
+  

[31/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/how-to-contribute.html
--
diff --git a/content/how-to-contribute.html b/content/how-to-contribute.html
deleted file mode 100644
index 87ad4a0..000
--- a/content/how-to-contribute.html
+++ /dev/null
@@ -1,345 +0,0 @@
-
-
-  
-
-
-
-
-Apache Flink: How To Contribute
-
-
-
-
-https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
-
-
-
-
-
-
-
-
-https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
-
-
-
-
-  
-
-
-
-
-
-
-
-  
- 
-  
-
-
-
-  
-
-
-
-  
-  
-
-  
-
-  
-
-
-
-
-  
-
-
-Download Flink
-
-
-Home
-
-
-Introduction to Flink
-
-
-Flink Use Cases
-
-
-Powered by Flink
-
-
-Ecosystem
-
-
-Community  Project Info
-
-
-How to 
Contribute
-
-
-Flink 
Blog
-
-
-
-
-
-
-
-
-  Documentation
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
-
-  
-
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
-
-
-
-
-  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
-
-
-
-
-
-
-
-  
-
-
-
-  
-  
-
-
-Project FAQ
-
-
-https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
-
-
-Plan Visualizer 
-
-  
-
-
-
-  
-  
-  
-  
-How To Contribute
-
-   Apache Flink is developed by an open and friendly community. 
Everybody is cordially welcome to join the community and contribute to Apache 
Flink. There are several ways to interact with the community and to contribute 
to Flink including asking questions, filing bug reports, proposing new 
features, joining discussions on the mailing lists, contributing code or 
documentation, improving the website, or testing release candidates.
-
-
-
-  Ask 
questions!
-  File a 
bug report
-  Propose an 
improvement or a new feature
-  Help others and join the 
discussions
-  Test a release candidate
-  Contribute 
code
-  Contribute documentation
-  Improve the website
-  More ways to contribute…
-  Submit a Contributor 
License Agreement
-  How to become a committer
-
-
-
-
-Ask questions!
-
-The Apache Flink community is eager to help and to answer your questions. 
We have a user mailing list, hang 
out in an IRC channel, and watch Stack 
Overflow on the http://stackoverflow.com/questions/tagged/apache-flink;>[apache-flink]
 tag.
-
-
-
-File a bug report
-
-Please let us know if you experienced a problem with Flink and file a bug 
report. Open http://issues.apache.org/jira/browse/FLINK;>Flink’s 
JIRA and click on the blue Create button at the top. 
Please give detailed information about the problem you encountered and, if 
possible, add a description that helps to reproduce the problem. Thank you very 
much.
-
-
-
-Propose an improvement or a 
new feature
-
-Our community is constantly looking for feedback to improve Apache Flink. 
If you have an idea how to improve Flink or have a new feature in mind that 
would be beneficial for Flink users, please open an issue in http://issues.apache.org/jira/browse/FLINK;>Flink’s JIRA. The 
improvement or new feature should be described in appropriate detail and 
include the scope and its requirements if possible. Detailed information is 
important for a few reasons:
-
-
-  It ensures your requirements are met when the improvement or feature is 
implemented.
-  It helps to estimate the effort and to design a solution that addresses 
your needs.
-  It allow for constructive discussions that might arise around this 
issue.
-
-
-Detailed information is also required, if you plan to contribute the 
improvement or feature you proposed yourself. Please read the Contribute code guide in this case as well.
-
-We recommend to first reach consensus with the community on whether a new 
feature is required and how to implement a new feature, before starting with 
the implementation. Some features might be out of scope of the project, and 
it’s best to 

[34/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
an Gates
-PMC, Committer
-gates
-  
-  
-https://avatars0.githubusercontent.com/u/2388347?s=50; />
-Fabian Hueske
-PMC, Committer
-fhueske
-  
-
-https://avatars3.githubusercontent.com/u/498957?v=3s=50; />
-Vasia Kalavri
-PMC, Committer
-vasia
-  
-  /tr
-
-https://avatars0.githubusercontent.com/u/68551?s=50; />
-Aljoscha Krettek
-PMC, Committer
-aljoscha
-  
-  
-https://avatars2.githubusercontent.com/u/2550549?s=50; />
-Andra Lungu
-Committer
-andra
-  
-  
-https://avatars0.githubusercontent.com/u/89049?s=50; />
-Robert Metzger
-PMC, Committer
-rmetzger
-  
-  
-https://avatars2.githubusercontent.com/u/837221?s=50; />
-Maximilian Michels
-PMC, Committer
-mxm
-  
-  
-https://avatars2.githubusercontent.com/u/1941681?s=50; />
-Chiwan Park
-Committer
-chiwanpark
-  
-  
-https://avatars1.githubusercontent.com/u/5756858?s=50; />
-Till Rohrmann
-PMC, Committer
-trohrmann
-  
-  
-https://avatars0.githubusercontent.com/u/105434?s=50; />
-Henry Saputra
-PMC, Committer
-hsaputra
-  
-  
-https://avatars0.githubusercontent.com/u/8959638?s=50; />
-Matthias J. Sax
-Committer
-mjsax
-  
-  
-https://avatars1.githubusercontent.com/u/409707?s=50; />
-Sebastian Schelter
-PMC, Committer
-ssc
-  
-  
-https://avatars2.githubusercontent.com/u/1925554?s=50; />
-Kostas Tzoumas
-PMC, Committer
-ktzoumas
-  
-  
-
-Timo Walther
-PMC, Committer
-twalthr
-  
-  
-https://avatars1.githubusercontent.com/u/1826769?s=50; />
-Daniel Warneke
-PMC, Committer
-warneke
-  
-  
-https://avatars1.githubusercontent.com/u/4425616?s=50; />
-ChengXiang Li
-Committer
-chengxiang
-  
-  
-https://avatars1.githubusercontent.com/u/569655?s=50; />
-Greg Hogan
-Committer
-greg
-  
-  
-https://avatars1.githubusercontent.com/u/5284370?s=50; />
-Tzu-Li (Gordon) Tai
-Committer
-tzulitai
-  
-
-
-You can reach committers directly at 
apache-id@apache.org. A list of all contributors can be 
found https://cwiki.apache.org/confluence/display/FLINK/List+of+contributors;>here.
-
-Former mentors
-
-The following people were very kind to mentor the project while in 
incubation.
-
-
-  
-
-Name
-Role
-Apache ID
-  
-  
-
-Ashutosh Chauhan
-Former PPMC, Mentor
-hashutosh
-  
-  
-
-Ted Dunning
-Former PPMC, Mentor
-tdunning
-  
-  
-
-Alan Gates
-Former PPMC, Mentor
-gates
-  
-  /tr
-
-
-Owen O'Malley
-Former PPMC, Mentor
-omalley
-  
-  
-
-Sean Owen
-Former PPMC, Mentor
-srowen
-  
-  
-
-Henry Saputra
-Former PPMC, Mentor
-hsaputra
-  
-
-
-Slides
-
-Note: Keep in mind that code examples on slides have a 
chance of being incomplete or outdated. Always refer to the http://ci.apache.org/projects/flink/flink-docs-release-1.2;>latest 
documentation for an up to date reference.
-
-2016
-
-
-  Stefan Richter: A look at Apache Flink 1.2 and beyond 
Apache Flink Meetup Berlin, November 2016: http://www.slideshare.net/StefanRichter10/a-look-at-flink-12;>SlideShare
-  Robert Metzger: Apache Flink Community Updates November 
2016 Apache Flink Meetup Berlin, November 2016: http://www.slideshare.net/robertmetzger1/apache-flink-community-updates-november-2016-berlin-meetup;>SlideShare
-  Aljoscha Krettek: Apache Flink for IoT: How Event-Time 
Processing Enables Easy and Accurate Analytics Big Data Spain, 
Madrid November 2016: http://www.slideshare.net/dataArtisans/aljoscha-krettek-apache-flink-for-iot-how-eventtime-processing-enables-easy-and-accurate-analytics;>SlideShare
-  Stephan Ewen: Keynote -The maturing data streaming ecosystem and 
Apache Flink’s accelerated growth Apache Big Data Europe 2016, 
Seville November 2016: http://www.slideshare.net/dataArtisans/keynote-stephan-ewen-stream-processing-as-a-foundational-paradigm-and-apache-flinks-approach-to-it;>SlideShare
-  Kostas Tzoumas: Stream Processing with Apache Flink® 
Apache Flink London Meetup, November 2016: http://www.slideshare.net/dataArtisans/kostas-tzoumas-stream-processing-with-apache-flink;>SlideShare
-  Kostas Tzoumas: Apache Flink®: State of the Union and What’s 
Next Strata + Hadoop World New York, September 2016: http://www.slideshare.net/dataArtisans/kostas-tzoumas-apache-flink-state-of-the-union-and-whats-next;>SlideShare
-  Kostas Tzoumas  Stephan Ewen: Keynote -The maturing data 
streaming ecosystem and Apache Flink’s accelerated growth Flink 
Forward, Berlin September 2016: http://www.slideshare.net/FlinkForward/kostas-tzoumasstpehan-ewen-keynote-the-maturing-data-streaming-ecosystem-and-apache-flinks-accelerated-growth;>SlideShare
-  Robert Metzger: Connecting Apache Flink to the World - Reviewing 
the streaming

[12/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/news/2016/03/08/release-1.0.0.html
--
diff --git a/content/news/2016/03/08/release-1.0.0.html 
b/content/news/2016/03/08/release-1.0.0.html
deleted file mode 100644
index 13313d8..000
--- a/content/news/2016/03/08/release-1.0.0.html
+++ /dev/null
@@ -1,319 +0,0 @@
-
-
-  
-
-
-
-
-Apache Flink: Announcing Apache Flink 1.0.0
-
-
-
-
-https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
-
-
-
-
-
-
-
-
-https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
-
-
-
-
-  
-
-
-
-
-
-
-
-  
- 
-  
-
-
-
-  
-
-
-
-  
-  
-
-  
-
-  
-
-
-
-
-  
-
-
-Download Flink
-
-
-Home
-
-
-Introduction to Flink
-
-
-Flink Use Cases
-
-
-Powered by Flink
-
-
-Ecosystem
-
-
-Community  Project Info
-
-
-How to Contribute
-
-
-Flink 
Blog
-
-
-
-
-
-
-
-
-  Documentation
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
-
-  
-
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
-
-
-
-
-  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
-
-
-
-
-
-
-
-  
-
-
-
-  
-  
-
-
-Project FAQ
-
-
-https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
-
-
-Plan Visualizer 
-
-  
-
-
-
-  
-  
-  
-  
-
-  Announcing Apache Flink 1.0.0
-
-  
-08 Mar 2016
-
-The Apache Flink community is pleased to announce the availability of the 
1.0.0 release. The community put significant effort into improving and 
extending Apache Flink since the last release, focusing on improving the 
experience of writing and executing data stream processing pipelines in 
production.
-
-
-
-
-
-Flink version 1.0.0 marks the beginning of the 1.X.X series of releases, 
which will maintain backwards compatibility with 1.0.0. This means that 
applications written against stable APIs of Flink 1.0.0 will compile and run 
with all Flink versions in the 1. series. This is the first time we are 
formally guaranteeing compatibility in Flink’s history, and we therefore see 
this release as a major milestone of the project, perhaps the most important 
since graduation as a top-level project.
-
-Apart from backwards compatibility, Flink 1.0.0 brings a variety of new 
user-facing features, as well as tons of bug fixes. About 64 contributors 
provided bug fixes, improvements, and new features such that in total more than 
450 JIRA issues could be resolved.
-
-We encourage everyone to http://flink.apache.org/downloads.html;>download the release and https://ci.apache.org/projects/flink/flink-docs-release-1.0/;>check out 
the documentation. Feedback through the Flink http://flink.apache.org/community.html#mailing-lists;>mailing lists 
is, as always, very welcome!
-
-Interface stability annotations
-
-Flink 1.0.0 introduces interface stability annotations for API classes and 
methods. Interfaces defined as @Public are guaranteed to remain 
stable across all releases of the 1.x series. The @PublicEvolving 
annotation marks API features that may be subject to change in future 
versions.
-
-Flink’s stability annotations will help users to implement applications 
that compile and execute unchanged against future versions of Flink 1.x. This 
greatly reduces the complexity for users when upgrading to a newer Flink 
release.
-
-Out-of-core state support
-
-Flink 1.0.0 adds a new state backend that uses RocksDB to store state (both 
windows and user-defined key-value state). http://rocksdb.org/;>RocksDB is an embedded key/value store database, 
originally developed by Facebook.
-When using this backend, active state in streaming programs can grow well 
beyond memory. The RocksDB files are stored in a distributed file system such 
as HDFS or S3 for backups.
-
-Savepoints and version upgrades
-
-Savepoints are checkpoints of the state of a running streaming job that can 
be manually 

[56/78] [abbrv] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/03/02/february-2015-in-flink.html
--
diff --git a/content/news/2015/03/02/february-2015-in-flink.html 
b/content/news/2015/03/02/february-2015-in-flink.html
new file mode 100644
index 000..1e3a09b
--- /dev/null
+++ b/content/news/2015/03/02/february-2015-in-flink.html
@@ -0,0 +1,308 @@
+
+
+  
+
+
+
+
+Apache Flink: February 2015 in the Flink community
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+
+  February 2015 in the Flink community
+
+  
+02 Mar 2015
+
+February might be the shortest month of the year, but this does not
+mean that the Flink community has not been busy adding features to the
+system and fixing bugs. Here’s a rundown of the activity in the Flink
+community last month.
+
+0.8.1 release
+
+Flink 0.8.1 was released. This bugfixing release resolves a total of 22 
issues.
+
+New committer
+
+https://github.com/mxm;>Max Michels has been voted a committer 
by the Flink PMC.
+
+Flink adapter for Apache SAMOA
+
+http://samoa.incubator.apache.org;>Apache SAMOA (incubating) 
is a
+distributed streaming machine learning (ML) framework with a
+programming abstraction for distributed streaming ML algorithms. SAMOA
+runs on a variety of backend engines, currently Apache Storm and
+Apache S4.  A https://github.com/apache/incubator-samoa/pull/11;>pull
+request is
+available at the SAMOA repository that adds a Flink adapter for SAMOA.
+
+Easy Flink deployment 
on Google Compute Cloud
+
+Flink is now integrated in bdutil, Google’s open source tool for
+creating and configuring (Hadoop) clusters in Google Compute
+Engine. Deployment of Flink clusters in now supported starting with
+https://groups.google.com/forum/#!topic/gcp-hadoop-announce/uVJ_6y9cGKM;>bdutil
+1.2.0.
+
+Flink on the Web
+
+A new blog post on http://flink.apache.org/news/2015/02/09/streaming-example.html;>Flink
+Streaming
+was published at the blog. Flink was mentioned in several articles on
+the web. Here are some examples:
+
+
+  
+http://dataconomy.com/how-flink-became-an-apache-top-level-project/;>How 
Flink became an Apache Top-Level Project
+  
+  
+https://www.linkedin.com/pulse/stale-synchronous-parallelism-new-frontier-apache-flink-nam-luc-tran?utm_content=buffer461afutm_medium=socialutm_source=linkedin.comutm_campaign=buffer;>Stale
 Synchronous Parallelism: The new frontier for Apache Flink?
+  
+  
+http://www.hadoopsphere.com/2015/02/distributed-data-processing-with-apache.html;>Distributed
 data processing with Apache Flink
+  
+  
+http://www.hadoopsphere.com/2015/02/ciao-latency-hallo-speed.html;>Ciao 
latency, hello speed
+  
+
+
+In the Flink master
+
+The following features have been now merged in Flink’s master 
repository.
+
+Gelly
+
+Gelly, Flink’s Graph API allows users to manipulate graph-shaped data
+directly. Here’s for example a calculation of shortest paths in a
+graph:
+
+GraphLong, Double, Double graph 
= Graph.fromDataSet(vertices, edges, env);
+
+DataSetVertexLong, 

[54/78] [abbrv] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/08/24/introducing-flink-gelly.html
--
diff --git a/content/news/2015/08/24/introducing-flink-gelly.html 
b/content/news/2015/08/24/introducing-flink-gelly.html
new file mode 100644
index 000..714e75a
--- /dev/null
+++ b/content/news/2015/08/24/introducing-flink-gelly.html
@@ -0,0 +1,649 @@
+
+
+  
+
+
+
+
+Apache Flink: Introducing Gelly: Graph Processing with Apache 
Flink
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+
+  Introducing Gelly: Graph Processing with Apache Flink
+
+  
+24 Aug 2015
+
+This blog post introduces Gelly, Apache Flink’s 
graph-processing API and library. Flink’s native support
+for iterations makes it a suitable platform for large-scale graph analytics.
+By leveraging delta iterations, Gelly is able to map various graph processing 
models such as
+vertex-centric or gather-sum-apply to Flink dataflows.
+
+Gelly allows Flink users to perform end-to-end data analysis in a single 
system.
+Gelly can be seamlessly used with Flink’s DataSet API,
+which means that pre-processing, graph creation, analysis, and post-processing 
can be done
+in the same application. At the end of this post, we will go through a 
step-by-step example
+in order to demonstrate that loading, transformation, filtering, graph 
creation, and analysis
+can be performed in a single Flink program.
+
+Overview
+
+
+  What is Gelly?
+  Graph Representation and 
Creation
+  Transformations and 
Utilities
+  Iterative Graph Processing
+  Library of Graph 
Algorithms
+  Use-Case: Music Profiles
+  Ongoing and Future Work
+
+
+
+
+What is Gelly?
+
+Gelly is a Graph API for Flink. It is currently supported in both Java and 
Scala.
+The Scala methods are implemented as wrappers on top of the basic Java 
operations.
+The API contains a set of utility functions for graph analysis, supports 
iterative graph
+processing and introduces a library of graph algorithms.
+
+
+
+
+
+Back to top
+
+Graph Representation and 
Creation
+
+In Gelly, a graph is represented by a DataSet of vertices and a DataSet of 
edges.
+A vertex is defined by its unique ID and a value, whereas an edge is defined 
by its source ID,
+target ID, and value. A vertex or edge for which a value is not specified will 
simply have the
+value type set to NullValue.
+
+A graph can be created from:
+
+
+  DataSet of edges and an optional DataSet of 
vertices using Graph.fromDataSet()
+  DataSet of Tuple3 and an optional DataSet of 
Tuple2 using Graph.fromTupleDataSet()
+  Collection of edges and an optional Collection 
of vertices using Graph.fromCollection()
+
+
+In all three cases, if the vertices are not provided,
+Gelly will automatically produce the vertex IDs from the edge source and 
target IDs.
+
+Back to top
+
+Transformations and Utilities
+
+These are methods of the Graph class and include common graph metrics, 
transformations
+and mutations as well as neighborhood aggregations.
+
+Common Graph Metrics
+These methods can be used to 

[21/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/svg/white_filled.svg
--
diff --git a/content/img/logo/svg/white_filled.svg 
b/content/img/logo/svg/white_filled.svg
deleted file mode 100755
index 28bd9f9..000
--- a/content/img/logo/svg/white_filled.svg
+++ /dev/null
@@ -1,231 +0,0 @@
-
-
-http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd;>
-http://www.w3.org/2000/svg; 
xmlns:xlink="http://www.w3.org/1999/xlink; x="0px" y="0px"
-width="2000px" height="1280px" viewBox="0 0 2000 1280" 
enable-background="new 0 0 2000 1280" xml:space="preserve">
-
-   
-   
-   
-   
-   
-   
-   
-   
-   
-   
-   
-   
-   
-   
-
-
-   
-   
-   
-   
-   
-   
-   
-   
-   
-   
-   
-
-
-   
-   
-   
-   
-   
-   
-   
-   
-   
-   
-   
-   
-   
-
-
-   
-   
-
-
-   
-   
-   
-   
-   
-   
-   
-   
-   
-
-Filled white
-

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/managed-state.png
--
diff --git a/content/img/managed-state.png b/content/img/managed-state.png
deleted file mode 100755
index 3dbb2fd..000
Binary files a/content/img/managed-state.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/memory_heap_division.png
--
diff --git a/content/img/memory_heap_division.png 
b/content/img/memory_heap_division.png
deleted file mode 100644
index 2b4c2e2..000
Binary files a/content/img/memory_heap_division.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/navbar-brand-logo.jpg
--
diff --git a/content/img/navbar-brand-logo.jpg 
b/content/img/navbar-brand-logo.jpg
deleted file mode 100755
index 5993ee8..000
Binary files a/content/img/navbar-brand-logo.jpg and /dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/navbar-brand-logo.png
--
diff --git a/content/img/navbar-brand-logo.png 
b/content/img/navbar-brand-logo.png
deleted file mode 100644
index 152f74e..000
Binary files a/content/img/navbar-brand-logo.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/one_runtime.png
--
diff --git a/content/img/one_runtime.png b/content/img/one_runtime.png
deleted file mode 100644
index 9cb4363..000
Binary files a/content/img/one_runtime.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/optimizer_choice.png
--
diff --git a/content/img/optimizer_choice.png b/content/img/optimizer_choice.png
deleted file mode 100644
index 1f8004b..000
Binary files a/content/img/optimizer_choice.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/otto-group-logo.jpg
--
diff --git a/content/img/otto-group-logo.jpg b/content/img/otto-group-logo.jpg
deleted file mode 100644
index f578af6..000
Binary files a/content/img/otto-group-logo.jpg and /dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/out_of_order_stream.png
--
diff --git a/content/img/out_of_order_stream.png 
b/content/img/out_of_order_stream.png
deleted file mode 100644
index 20ad09d..000
Binary files a/content/img/out_of_order_stream.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/parallel_dataflows.png
--
diff --git a/content/img/parallel_dataflows.png 
b/content/img/parallel_dataflows.png
deleted file mode 100644
index dbd83b2..000
Binary files a/content/img/parallel_dataflows.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/poweredby/alibaba-logo.png

[04/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/js/graphCreator.js
--
diff --git a/content/visualizer/js/graphCreator.js 
b/content/visualizer/js/graphCreator.js
new file mode 100755
index 000..86c8c81
--- /dev/null
+++ b/content/visualizer/js/graphCreator.js
@@ -0,0 +1,445 @@
+/*!
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+/*
+ * Document initialization.
+ */
+$(document).ready(function () 
+{
+   zoom = d3.behavior.zoom("#svg-main").on("zoom", function() {
+   var ev = d3.event;
+   d3.select("#svg-main g")
+   .attr("transform", "translate(" + ev.translate + ") 
scale(" + ev.scale + ")");
+   });
+   zoom.scaleExtent([0.3, 3])
+});
+
+//The current JSON file
+var JSONData; 
+//The informations for the iterations
+var iterationIds = new Array();
+var iterationGraphs = new Array();
+var iterationWidths = new Array();
+var iterationHeights = new Array();
+
+//The zoom element
+var zoom;
+//informations about the enviroment
+var svgWidth;
+var svgHeight;
+
+//Renders and draws the graph
+function drawGraph(data, svgID){
+   JSONData = data;
+   
+   //First step: precompute all iteration graphs
+   
+   //find all iterations
+   iterationNodes = searchForIterationNodes();
+   
+   //add the graphs of iterations and their sizes + Ids to arrays
+   if (iterationNodes != null) {
+   for (var i in iterationNodes) {
+   var itNode = iterationNodes[i];
+   iterationIds.push(itNode.id);
+   var g0 = loadJsonToDagre(itNode);
+   iterationGraphs.push(g0);
+   var r = new dagreD3.Renderer();
+   var l = dagreD3.layout()
+.nodeSep(20)
+.rankDir("LR");
+   l = r.layout(l).run(g0, d3.select("#svg-main"));
+   
+   iterationWidths.push(l._value.width);
+   iterationHeights.push(l._value.height);
+   
+   //Clean svg
+   $("#svg-main g").empty();
+   }
+   }
+   
+   //Continue normal
+   var g = loadJsonToDagre(data);
+   var selector = svgID + " g";
+   var renderer = new dagreD3.Renderer();
+   var layout = dagreD3.layout()
+   .nodeSep(20)
+   .rankDir("LR");
+   var svgElement = d3.select(selector);
+   layout = renderer.layout(layout).run(g, svgElement);
+   
+   svgHeight = layout._value.height;
+   svgWidth = layout._value.width;
+   
+var svg = d3.select("#svg-main")
+   .attr("width", $(document).width() - 15)
+   .attr("height", $(document).height() - 15 - 110)
+// .attr("viewBox", "0 0 "+ ($(document).width() - 150) +" 
"+($(document).height() - 15 - 110))
+   .call(zoom);
+   
+   // This should now draw the precomputed graphs in the svgs... . 
+   
+   for (var i in iterationIds) {
+   var workset = searchForNode(iterationIds[i]);
+   renderer = new dagreD3.Renderer();
+   layout = dagreD3.layout()
+.nodeSep(20)
+.rankDir("LR");
+   selector = "#svg-"+iterationIds[i]+" g";
+   svgElement = d3.select(selector);
+   layout = renderer.layout(layout).run(iterationGraphs[i], 
svgElement);
+   }
+   
+   //enable Overlays and register function for overlay-infos
+   $("a[rel]").overlay({ 
+   onBeforeLoad: function(){ 
+   var id = this.getTrigger().attr("nodeID")
+   showProperties(id);
+   }
+   });
+   
+   }
+
+//Creates the dagreD3 graph object
+//Responsible for adding nodes and edges
+function loadJsonToDagre(data){
+   
+   //stores all nodes that are in current graph -> no edges to nodes which 
are outside of current 

[17/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/03/02/february-2015-in-flink.html
--
diff --git a/content/news/2015/03/02/february-2015-in-flink.html 
b/content/news/2015/03/02/february-2015-in-flink.html
new file mode 100644
index 000..1e3a09b
--- /dev/null
+++ b/content/news/2015/03/02/february-2015-in-flink.html
@@ -0,0 +1,308 @@
+
+
+  
+
+
+
+
+Apache Flink: February 2015 in the Flink community
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+
+  February 2015 in the Flink community
+
+  
+02 Mar 2015
+
+February might be the shortest month of the year, but this does not
+mean that the Flink community has not been busy adding features to the
+system and fixing bugs. Here’s a rundown of the activity in the Flink
+community last month.
+
+0.8.1 release
+
+Flink 0.8.1 was released. This bugfixing release resolves a total of 22 
issues.
+
+New committer
+
+https://github.com/mxm;>Max Michels has been voted a committer 
by the Flink PMC.
+
+Flink adapter for Apache SAMOA
+
+http://samoa.incubator.apache.org;>Apache SAMOA (incubating) 
is a
+distributed streaming machine learning (ML) framework with a
+programming abstraction for distributed streaming ML algorithms. SAMOA
+runs on a variety of backend engines, currently Apache Storm and
+Apache S4.  A https://github.com/apache/incubator-samoa/pull/11;>pull
+request is
+available at the SAMOA repository that adds a Flink adapter for SAMOA.
+
+Easy Flink deployment 
on Google Compute Cloud
+
+Flink is now integrated in bdutil, Google’s open source tool for
+creating and configuring (Hadoop) clusters in Google Compute
+Engine. Deployment of Flink clusters in now supported starting with
+https://groups.google.com/forum/#!topic/gcp-hadoop-announce/uVJ_6y9cGKM;>bdutil
+1.2.0.
+
+Flink on the Web
+
+A new blog post on http://flink.apache.org/news/2015/02/09/streaming-example.html;>Flink
+Streaming
+was published at the blog. Flink was mentioned in several articles on
+the web. Here are some examples:
+
+
+  
+http://dataconomy.com/how-flink-became-an-apache-top-level-project/;>How 
Flink became an Apache Top-Level Project
+  
+  
+https://www.linkedin.com/pulse/stale-synchronous-parallelism-new-frontier-apache-flink-nam-luc-tran?utm_content=buffer461afutm_medium=socialutm_source=linkedin.comutm_campaign=buffer;>Stale
 Synchronous Parallelism: The new frontier for Apache Flink?
+  
+  
+http://www.hadoopsphere.com/2015/02/distributed-data-processing-with-apache.html;>Distributed
 data processing with Apache Flink
+  
+  
+http://www.hadoopsphere.com/2015/02/ciao-latency-hallo-speed.html;>Ciao 
latency, hello speed
+  
+
+
+In the Flink master
+
+The following features have been now merged in Flink’s master 
repository.
+
+Gelly
+
+Gelly, Flink’s Graph API allows users to manipulate graph-shaped data
+directly. Here’s for example a calculation of shortest paths in a
+graph:
+
+GraphLong, Double, Double graph 
= Graph.fromDataSet(vertices, edges, env);
+
+DataSetVertexLong, 

[01/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
Repository: flink-web
Updated Branches:
  refs/heads/asf-site 16a92b0c1 -> 9ec0a879f


http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/js/program.js
--
diff --git a/content/visualizer/js/program.js b/content/visualizer/js/program.js
new file mode 100755
index 000..683185d
--- /dev/null
+++ b/content/visualizer/js/program.js
@@ -0,0 +1,233 @@
+/*!
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+var maxColumnWidth = 200;
+var minColumnWidth = 100;
+
+// global variable for the currently requested plan
+var pactPlanRequested = 0;
+
+/*
+ * This function toggels the child checkbox on and of, depending on the 
parent's state
+ */
+function toggleShowPlanBox(box)
+{
+  var child = $('#suspendJobDuringPlanCheck');
+  
+  if (box.is(':checked')) {
+child.attr('disabled', false);
+  }
+  else {
+child.attr('disabled', true);
+  }
+}
+
+/*
+ * Shows an error message below the upload field.
+ */
+function showUploadError(message)
+{
+  $('#upload_error_text').fadeOut("fast", function () { 
$('#upload_error_text')[0].innerHTML = "" + message;
+   
$('#upload_error_text').fadeIn("slow"); } );
+}
+
+/*
+ * Checks the selected file and triggers an upload, if all is correct.
+ */
+function processUpload()
+{
+
+  var filename = $('#upload_file_input').val();
+  var len = filename.length;
+  if (len == 0) {
+showUploadError("Please select a file.");
+  }
+  else if (len > 4 && filename.substr(len - 4, len) == ".jar") {
+$('#upload_form')[0].submit();
+  }
+  else {
+showUploadError("Please select a .jar file.");
+  }
+}
+
+/*
+ * This function makes sure only one checkbox is selected.
+ * Upon selection it initializes the drawing of the pact plan.
+ * Upon deselection, it clears the pact plan.
+ */
+function toggleCheckboxes(box)
+{
+
+  if (box.is(':checked')) {
+$('.jobItemCheckbox').attr('checked', false);
+box.attr('checked', true);
+var id = box.parentsUntil('.JobListItems').parent().attr('id').substr(4);
+var assemblerClass = box.attr('id');
+
+$('#mainCanvas').html('');
+pactPlanRequested = id;
+
+$.ajax({
+type: "GET",
+url: "pactPlan",
+data: { job: id, assemblerClass: assemblerClass},
+success: function(response) { showPreviewPlan(response); }
+});
+  }
+  else {
+$('#mainCanvas').html('');
+  }
+}
+
+/*
+ * Function that takes the returned plan and draws it.
+ */
+function showPreviewPlan(data)
+{
+   $("#mainCanvas").empty();
+var svgElement = "";
+$("#mainCanvas").append(svgElement);
+drawGraph(data.plan, "#svg-main");
+pactPlanRequested = 0;
+
+//activate zoom buttons
+activateZoomButtons();
+//  }
+}
+
+/*
+ * Asynchronously loads the jobs and creates a list of them.
+ */
+function loadJobList()
+{
+  $.get("jobs", { action: "list" }, createJobList);
+}
+
+/*
+ * Triggers an AJAX request to delete a job.
+ */
+function deleteJob(id)
+{
+  var name = id.substr(4);
+  $.get("jobs", { action: "delete", filename: name }, loadJobList);
+}
+
+/*
+ * Creates and lists the returned jobs.
+ */
+function createJobList(data)
+{
+  var markup = "";
+   
+  var lines = data.split("\n");
+  for (var i = 0; i < lines.length; i++)
+  {
+if (lines[i] == null || lines[i].length == 0) {
+  continue;
+}
+
+var date = "unknown date";
+var assemblerClass = "no entry class specified";
+
+var tokens = lines[i].split("\t");
+var name = tokens[0];
+if (tokens.length > 1) {
+  date = tokens[1];
+  if (tokens.length > 2) {
+assemblerClass = tokens[2];
+  }
+}
+
+var entries = assemblerClass.split("#");
+var classes = entries[0].split(",");
+
+markup += '';
+markup += '' + name + 
'';
+markup += '' + date + '';
+markup += '';
+
+var j = 0;
+for (var idx in classes) {
+  markup += '';
+  markup += '' + classes[idx] + '';
+}
+markup += '';
+  }
+  
+  // add the contents
+  $('#jobsContents').html(markup); 
+  
+  // register the event handler that triggers the delete when the 

[25/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/svg/color_black.svg
--
diff --git a/content/img/logo/svg/color_black.svg 
b/content/img/logo/svg/color_black.svg
new file mode 100755
index 000..d54707c
--- /dev/null
+++ b/content/img/logo/svg/color_black.svg
@@ -0,0 +1,1561 @@
+
+
+http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd;>
+http://www.w3.org/2000/svg; 
xmlns:xlink="http://www.w3.org/1999/xlink; x="0px" y="0px"
+width="2000px" height="1280px" viewBox="0 0 2000 1280" 
enable-background="new 0 0 2000 1280" xml:space="preserve">
+
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+
+
+   
+   
+
+
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   

[16/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/05/11/Juggling-with-Bits-and-Bytes.html
--
diff --git a/content/news/2015/05/11/Juggling-with-Bits-and-Bytes.html 
b/content/news/2015/05/11/Juggling-with-Bits-and-Bytes.html
new file mode 100644
index 000..a7a5436
--- /dev/null
+++ b/content/news/2015/05/11/Juggling-with-Bits-and-Bytes.html
@@ -0,0 +1,383 @@
+
+
+  
+
+
+
+
+Apache Flink: Juggling with Bits and Bytes
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+
+  Juggling with Bits and Bytes
+
+  
+11 May 2015 by Fabian Hüske (https://twitter.com/fhueske;>@fhueske)
+
+How Apache Flink operates on 
binary data
+
+Nowadays, a lot of open-source systems for analyzing large data sets are 
implemented in Java or other JVM-based programming languages. The most 
well-known example is Apache Hadoop, but also newer frameworks such as Apache 
Spark, Apache Drill, and also Apache Flink run on JVMs. A common challenge that 
JVM-based data analysis engines face is to store large amounts of data in 
memory - both for caching and for efficient processing such as sorting and 
joining of data. Managing the JVM memory well makes the difference between a 
system that is hard to configure and has unpredictable reliability and 
performance and a system that behaves robustly with few configuration knobs.
+
+In this blog post we discuss how Apache Flink manages memory, talk about 
its custom data de/serialization stack, and show how it operates on binary 
data.
+
+Data Objects? Let’s put them 
on the heap!
+
+The most straight-forward approach to process lots of data in a JVM is to 
put it as objects on the heap and operate on these objects. Caching a data set 
as objects would be as simple as maintaining a list containing an object for 
each record. An in-memory sort would simply sort the list of objects.
+However, this approach has a few notable drawbacks. First of all it is not 
trivial to watch and control heap memory usage when a lot of objects are 
created and invalidated constantly. Memory overallocation instantly kills the 
JVM with an OutOfMemoryError. Another aspect is garbage collection 
on multi-GB JVMs which are flooded with new objects. The overhead of garbage 
collection in such environments can easily reach 50% and more. Finally, Java 
objects come with a certain space overhead depending on the JVM and platform. 
For data sets with many small objects this can significantly reduce the 
effectively usable amount of memory. Given proficient system design and 
careful, use-case specific system parameter tuning, heap memory usage can be 
more or less controlled and OutOfMemoryErrors avoided. However, 
such setups are rather fragile especially if data characteristics or the 
execution environment change.
+
+What is Flink doing about that?
+
+Apache Flink has its roots at a research project which aimed to combine the 
best technologies of MapReduce-based systems and parallel database systems. 
Coming from this background, Flink has always had 

[06/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/js/d3.js
--
diff --git a/content/visualizer/js/d3.js b/content/visualizer/js/d3.js
new file mode 100755
index 000..2d09bc9
--- /dev/null
+++ b/content/visualizer/js/d3.js
@@ -0,0 +1,9270 @@
+/**
+ * D3 renderer from http://d3js.org/
+ * Copyright (c) 2010-2014, Michael Bostock
+ * 
+ * Under BSD License (as stated under http://d3js.org/)
+ */
+
+!function() {
+  var d3 = {
+version: "3.4.6"
+  };
+  if (!Date.now) Date.now = function() {
+return +new Date();
+  };
+  var d3_arraySlice = [].slice, d3_array = function(list) {
+return d3_arraySlice.call(list);
+  };
+  var d3_document = document, d3_documentElement = 
d3_document.documentElement, d3_window = window;
+  try {
+d3_array(d3_documentElement.childNodes)[0].nodeType;
+  } catch (e) {
+d3_array = function(list) {
+  var i = list.length, array = new Array(i);
+  while (i--) array[i] = list[i];
+  return array;
+};
+  }
+  try {
+d3_document.createElement("div").style.setProperty("opacity", 0, "");
+  } catch (error) {
+var d3_element_prototype = d3_window.Element.prototype, 
d3_element_setAttribute = d3_element_prototype.setAttribute, 
d3_element_setAttributeNS = d3_element_prototype.setAttributeNS, 
d3_style_prototype = d3_window.CSSStyleDeclaration.prototype, 
d3_style_setProperty = d3_style_prototype.setProperty;
+d3_element_prototype.setAttribute = function(name, value) {
+  d3_element_setAttribute.call(this, name, value + "");
+};
+d3_element_prototype.setAttributeNS = function(space, local, value) {
+  d3_element_setAttributeNS.call(this, space, local, value + "");
+};
+d3_style_prototype.setProperty = function(name, value, priority) {
+  d3_style_setProperty.call(this, name, value + "", priority);
+};
+  }
+  d3.ascending = d3_ascending;
+  function d3_ascending(a, b) {
+return a < b ? -1 : a > b ? 1 : a >= b ? 0 : NaN;
+  }
+  d3.descending = function(a, b) {
+return b < a ? -1 : b > a ? 1 : b >= a ? 0 : NaN;
+  };
+  d3.min = function(array, f) {
+var i = -1, n = array.length, a, b;
+if (arguments.length === 1) {
+  while (++i < n && !((a = array[i]) != null && a <= a)) a = undefined;
+  while (++i < n) if ((b = array[i]) != null && a > b) a = b;
+} else {
+  while (++i < n && !((a = f.call(array, array[i], i)) != null && a <= a)) 
a = undefined;
+  while (++i < n) if ((b = f.call(array, array[i], i)) != null && a > b) a 
= b;
+}
+return a;
+  };
+  d3.max = function(array, f) {
+var i = -1, n = array.length, a, b;
+if (arguments.length === 1) {
+  while (++i < n && !((a = array[i]) != null && a <= a)) a = undefined;
+  while (++i < n) if ((b = array[i]) != null && b > a) a = b;
+} else {
+  while (++i < n && !((a = f.call(array, array[i], i)) != null && a <= a)) 
a = undefined;
+  while (++i < n) if ((b = f.call(array, array[i], i)) != null && b > a) a 
= b;
+}
+return a;
+  };
+  d3.extent = function(array, f) {
+var i = -1, n = array.length, a, b, c;
+if (arguments.length === 1) {
+  while (++i < n && !((a = c = array[i]) != null && a <= a)) a = c = 
undefined;
+  while (++i < n) if ((b = array[i]) != null) {
+if (a > b) a = b;
+if (c < b) c = b;
+  }
+} else {
+  while (++i < n && !((a = c = f.call(array, array[i], i)) != null && a <= 
a)) a = undefined;
+  while (++i < n) if ((b = f.call(array, array[i], i)) != null) {
+if (a > b) a = b;
+if (c < b) c = b;
+  }
+}
+return [ a, c ];
+  };
+  d3.sum = function(array, f) {
+var s = 0, n = array.length, a, i = -1;
+if (arguments.length === 1) {
+  while (++i < n) if (!isNaN(a = +array[i])) s += a;
+} else {
+  while (++i < n) if (!isNaN(a = +f.call(array, array[i], i))) s += a;
+}
+return s;
+  };
+  function d3_number(x) {
+return x != null && !isNaN(x);
+  }
+  d3.mean = function(array, f) {
+var s = 0, n = array.length, a, i = -1, j = n;
+if (arguments.length === 1) {
+  while (++i < n) if (d3_number(a = array[i])) s += a; else --j;
+} else {
+  while (++i < n) if (d3_number(a = f.call(array, array[i], i))) s += a; 
else --j;
+}
+return j ? s / j : undefined;
+  };
+  d3.quantile = function(values, p) {
+var H = (values.length - 1) * p + 1, h = Math.floor(H), v = +values[h - 
1], e = H - h;
+return e ? v + e * (values[h] - v) : v;
+  };
+  d3.median = function(array, f) {
+if (arguments.length > 1) array = array.map(f);
+array = array.filter(d3_number);
+return array.length ? d3.quantile(array.sort(d3_ascending), .5) : 
undefined;
+  };
+  function d3_bisector(compare) {
+return {
+  left: function(a, x, lo, hi) {
+if (arguments.length < 3) lo = 0;
+if (arguments.length < 4) hi = a.length;
+

[38/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/blog/feed.xml
--
diff --git a/content/blog/feed.xml b/content/blog/feed.xml
new file mode 100644
index 000..066ae2f
--- /dev/null
+++ b/content/blog/feed.xml
@@ -0,0 +1,6041 @@
+
+http://www.w3.org/2005/Atom;>
+
+Flink Blog Feed
+Flink Blog
+http://flink.apache.org/blog
+http://flink.apache.org/blog/feed.xml; rel="self" 
type="application/rss+xml" />
+
+
+Apache Flink 1.1.4 Released
+pThe Apache Flink community released the next bugfix 
version of the Apache Flink 1.1 series./p
+
+pThis release includes major robustness improvements for checkpoint 
cleanup on failures and consumption of intermediate streams. We highly 
recommend all users to upgrade to Flink 1.1.4./p
+
+div class=highlightprecode 
class=language-xmlspan 
class=ntlt;dependencygt;/span
+  span 
class=ntlt;groupIdgt;/spanorg.apache.flinkspan
 class=ntlt;/groupIdgt;/span
+  span 
class=ntlt;artifactIdgt;/spanflink-javaspan
 class=ntlt;/artifactIdgt;/span
+  span 
class=ntlt;versiongt;/span1.1.4span 
class=ntlt;/versiongt;/span
+span class=ntlt;/dependencygt;/span
+span class=ntlt;dependencygt;/span
+  span 
class=ntlt;groupIdgt;/spanorg.apache.flinkspan
 class=ntlt;/groupIdgt;/span
+  span 
class=ntlt;artifactIdgt;/spanflink-streaming-java_2.10span
 class=ntlt;/artifactIdgt;/span
+  span 
class=ntlt;versiongt;/span1.1.4span 
class=ntlt;/versiongt;/span
+span class=ntlt;/dependencygt;/span
+span class=ntlt;dependencygt;/span
+  span 
class=ntlt;groupIdgt;/spanorg.apache.flinkspan
 class=ntlt;/groupIdgt;/span
+  span 
class=ntlt;artifactIdgt;/spanflink-clients_2.10span
 class=ntlt;/artifactIdgt;/span
+  span 
class=ntlt;versiongt;/span1.1.4span 
class=ntlt;/versiongt;/span
+span 
class=ntlt;/dependencygt;/span/code/pre/div
+
+pYou can find the binaries on the updated a 
href=http://flink.apache.org/downloads.htmlDownloads 
page/a./p
+
+h2 id=note-for-rocksdb-backend-usersNote for RocksDB 
Backend Users/h2
+
+pWe updated Flink’s RocksDB dependency version from 
code4.5.1/code to code4.11.2/code. Between 
these versions some of RocksDB’s internal configuration defaults changed that 
would affect the memory footprint of running Flink with RocksDB. Therefore, we 
manually reset them to the previous defaults. If you want to run with the new 
Rocks 4.11.2 defaults, you can do this via:/p
+
+div class=highlightprecode 
class=language-javaspan 
class=nRocksDBStateBackend/span span 
class=nbackend/span span 
class=o=/span span 
class=knew/span span 
class=nfRocksDBStateBackend/spanspan 
class=o(/spanspan 
class=squot;...quot;/spanspan 
class=o);/span
+span class=c1// Use the new default options. Otherwise, 
the default for RocksDB 4.5.1/span
+span class=c1// `PredefinedOptions.DEFAULT_ROCKS_4_5_1` 
will be used./span
+span class=nbackend/spanspan 
class=o./spanspan 
class=nasetPredefinedOptions/spanspan 
class=o(/spanspan 
class=nPredefinedOptions/spanspan 
class=o./spanspan 
class=naDEFAULT/spanspan 
class=o);/span/code/pre/div
+
+h2 id=release-notes---flink---version-114Release Notes - 
Flink - Version 1.1.4/h2
+
+h3 id=sub-taskSub-task/h3
+ul
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4510FLINK-4510/a;]
 - Always create CheckpointCoordinator
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4984FLINK-4984/a;]
 - Add Cancellation Barriers to BarrierTracker and BarrierBuffer
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4985FLINK-4985/a;]
 - Report Declined/Canceled Checkpoints to Checkpoint Coordinator
+/li
+/ul
+
+h3 id=bugBug/h3
+ul
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-2662FLINK-2662/a;]
 - CompilerException: quot;Bug: Plan generation for Unions picked 
a ship strategy between binary plan operators.quot;
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-3680FLINK-3680/a;]
 - Remove or improve (not set) text in the Job Plan UI
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-3813FLINK-3813/a;]
 - YARNSessionFIFOITCase.testDetachedMode failed on Travis
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4108FLINK-4108/a;]
 - NPE in Row.productArity
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4506FLINK-4506/a;]
 - CsvOutputFormat defaults allowNullValues to false, even though doc 
and declaration says true
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4581FLINK-4581/a;]
 - Table API throws quot;No suitable driver found for 
jdbc:calcitequot;
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4586FLINK-4586/a;]
 - NumberSequenceIterator and Accumulator threading issue
+/li
+li[a 
href=https://issues.apache.org/jira/browse/FLINK-4619FLINK-4619/a;]
 - JobManager does not answer to client when restore from savepoint 
fails
+/li
+li[a 

[39/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
Rebuild site


Project: http://git-wip-us.apache.org/repos/asf/flink-web/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink-web/commit/9ec0a879
Tree: http://git-wip-us.apache.org/repos/asf/flink-web/tree/9ec0a879
Diff: http://git-wip-us.apache.org/repos/asf/flink-web/diff/9ec0a879

Branch: refs/heads/asf-site
Commit: 9ec0a879f23ebb9c0533bf7e001ebf830a8bdd6f
Parents: 16a92b0
Author: Robert Metzger 
Authored: Wed Jan 18 15:04:07 2017 +0100
Committer: Robert Metzger 
Committed: Wed Jan 18 15:04:07 2017 +0100

--
 content/blog/feed.xml   | 6041 
 content/blog/index.html |  747 ++
 content/blog/page2/index.html   |  743 ++
 content/blog/page3/index.html   |  756 ++
 content/blog/page4/index.html   |  760 ++
 .../release_1.0.0-changelog_known_issues.html   | 1139 +++
 content/blog/release_1.1.0-changelog.html   |  689 ++
 content/community.html  |  743 ++
 content/contribute-code.html|  519 +
 content/contribute-documentation.html   |  248 +
 content/css/flink.css   |  286 +
 content/css/syntax.css  |   60 +
 content/doap_flink.rdf  |   51 +
 content/documentation.html  |  196 +
 content/downloads.html  |  320 +
 content/ecosystem.html  |  278 +
 content/faq.html|  683 ++
 content/favicon.ico |  Bin 0 -> 1150 bytes
 content/features.html   |  511 +
 content/how-to-contribute.html  |  345 +
 content/img/alibaba-logo.png|  Bin 0 -> 22897 bytes
 content/img/assets/WhatIsFlink.png  |  Bin 0 -> 248892 bytes
 content/img/assets/grep.png |  Bin 0 -> 121239 bytes
 content/img/assets/hadoop-img.png   |  Bin 0 -> 67841 bytes
 content/img/assets/optimizer-visual.png |  Bin 0 -> 204452 bytes
 content/img/assets/pagerank.pdf |  Bin 0 -> 21052 bytes
 content/img/assets/pagerank.png |  Bin 0 -> 135851 bytes
 content/img/blog/GSA-plan.png   |  Bin 0 -> 54646 bytes
 content/img/blog/appeared-in.png|  Bin 0 -> 317098 bytes
 content/img/blog/blog_basic_window.png  |  Bin 0 -> 144158 bytes
 content/img/blog/blog_data_driven.png   |  Bin 0 -> 96886 bytes
 content/img/blog/blog_multi_input.png   |  Bin 0 -> 112104 bytes
 content/img/blog/blog_social_media.png  |  Bin 0 -> 98666 bytes
 content/img/blog/blog_stream_join.png   |  Bin 0 -> 112305 bytes
 content/img/blog/cep-monitoring.svg | 2838 ++
 content/img/blog/code-growth.png|  Bin 0 -> 87120 bytes
 content/img/blog/commit-stats.png   |  Bin 0 -> 5 bytes
 content/img/blog/community-growth.png   |  Bin 0 -> 102186 bytes
 content/img/blog/data-serialization.png |  Bin 0 -> 395455 bytes
 content/img/blog/emr-firefoxsettings.png|  Bin 0 -> 176706 bytes
 content/img/blog/emr-hadoopversion.png  |  Bin 0 -> 97799 bytes
 content/img/blog/emr-jobmanager.png |  Bin 0 -> 84576 bytes
 content/img/blog/emr-running.png|  Bin 0 -> 78210 bytes
 content/img/blog/emr-security.png   |  Bin 0 -> 76951 bytes
 content/img/blog/emr-yarnappmaster.png  |  Bin 0 -> 85393 bytes
 content/img/blog/feature-timeline.png   |  Bin 0 -> 182292 bytes
 content/img/blog/ff-speakers.png|  Bin 0 -> 367370 bytes
 content/img/blog/flink-1.0.png  |  Bin 0 -> 114360 bytes
 content/img/blog/flink-dow-2016.png |  Bin 0 -> 3251 bytes
 content/img/blog/flink-forward-banner.png   |  Bin 0 -> 37458 bytes
 content/img/blog/flink-hod-2016.png |  Bin 0 -> 3090 bytes
 content/img/blog/flink-lines-of-code-2016.png   |  Bin 0 -> 3935 bytes
 content/img/blog/flink-meetups-dec2016.png  |  Bin 0 -> 560426 bytes
 content/img/blog/flink-releases-2016.png|  Bin 0 -> 109796 bytes
 content/img/blog/flink-stack.png|  Bin 0 -> 32934 bytes
 content/img/blog/flink-storm.png|  Bin 0 -> 127217 bytes
 content/img/blog/flinkSer-int-gc.png|  Bin 0 -> 12405 bytes
 content/img/blog/flinkSer-int-mem.png   |  Bin 0 -> 15689 bytes
 content/img/blog/github-stats-2016.png  |  Bin 0 -> 93704 bytes
 content/img/blog/hadoop-summit.png  |  Bin 0 -> 330780 bytes
 content/img/blog/hcompat-flow.png   |  Bin 0 -> 102575 bytes
 content/img/blog/hcompat-logos.png  |  Bin 0 -> 320253 bytes
 content/img/blog/iteration.png  |  Bin 0 -> 24531 bytes
 

[02/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/js/jquery.tools.min.js
--
diff --git a/content/visualizer/js/jquery.tools.min.js 
b/content/visualizer/js/jquery.tools.min.js
new file mode 100755
index 000..dc914d8
--- /dev/null
+++ b/content/visualizer/js/jquery.tools.min.js
@@ -0,0 +1,20 @@
+/*!
+ * jQuery Tools v1.2.7 - The missing UI library for the Web
+ * 
+ * overlay/overlay.js
+ * scrollable/scrollable.js
+ * tabs/tabs.js
+ * tooltip/tooltip.js
+ * 
+ * NO COPYRIGHTS OR LICENSES. DO WHAT YOU LIKE.
+ * 
+ * http://flowplayer.org/tools/
+ * 
+ */
+/*! jQuery v1.7.2 jquery.com | jquery.org/license */
+(function(a,b){function cy(a){return 
f.isWindow(a)?a:a.nodeType===9?a.defaultView||a.parentWindow:!1}function 
cu(a){if(!cj[a]){var 
b=c.body,d=f("<"+a+">").appendTo(b),e=d.css("display");d.remove();if(e==="none"||e===""){ck||(ck=c.createElement("iframe"),ck.frameBorder=ck.width=ck.height=0),b.appendChild(ck);if(!cl||!ck.createElement)cl=(ck.contentWindow||ck.contentDocument).document,cl.write((f.support.boxModel?"":"")+""),cl.close();d=cl.createElement(a),cl.body.appendChild(d),e=f.css(d,"display"),b.removeChild(ck)}cj[a]=e}return
 cj[a]}function ct(a,b){var 
c={};f.each(cp.concat.apply([],cp.slice(0,b)),function(){c[this]=a});return 
c}function cs(){cq=b}function cr(){setTimeout(cs,0);return cq=f.now()}function 
ci(){try{return new a.ActiveXObject("Microsoft.XMLHTTP")}catch(b){}}function 
ch(){try{return new a.XMLHttpRequest}catch(b){}}function 
cb(a,c){a.dataFilter&&(c=a.dataFilter(c,a.dataType));var 
d=a.dataTypes,e={},g,h,i=d.length,j,k=d[0],l,m,n,o,p;for(g=1;
 g0){if(c!=="border")for(;e

[33/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/contribute-documentation.html
--
diff --git a/content/contribute-documentation.html 
b/content/contribute-documentation.html
new file mode 100644
index 000..11f9000
--- /dev/null
+++ b/content/contribute-documentation.html
@@ -0,0 +1,248 @@
+
+
+  
+
+
+
+
+Apache Flink: Contribute Documentation
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+Contribute Documentation
+
+   Good documentation is crucial for any kind of software. This is 
especially true for sophisticated software systems such as distributed data 
processing engines like Apache Flink. The Apache Flink community aims to 
provide concise, precise, and complete documentation and welcomes any 
contribution to improve Apache Flink’s documentation.
+
+
+
+  Obtain the documentation 
sources
+  Before you 
start working on the documentation …
+  Update or extend the 
documentation
+  Submit your contribution
+
+
+
+
+Obtain the documentation sources
+
+Apache Flink’s documentation is maintained in the same http://git-scm.com/;>git repository as the code base. This is done to 
ensure that code and documentation can be easily kept in sync.
+
+The easiest way to contribute documentation is to fork https://github.com/apache/flink;>Flink’s mirrored repository on 
Github into your own Github account by clicking on the fork button at the 
top right. If you have no Github account, you can create one for free.
+
+Next, clone your fork to your local machine.
+
+git clone 
https://github.com/your-user-name/flink.git
+
+
+The documentation is located in the docs/ subdirectory of the 
Flink code base.
+
+Before you start 
working on the documentation …
+
+… please make sure there exists a https://issues.apache.org/jira/browse/FLINK;>JIRA issue that 
corresponds to your contribution. We require all documentation changes to refer 
to a JIRA issue, except for trivial fixes such as typos.
+
+Update or extend the 
documentation
+
+The Flink documentation is written in http://daringfireball.net/projects/markdown/;>Markdown. Markdown is a 
lightweight markup language which can be translated to HTML.
+
+In order to update or extend the documentation you have to modify the 
Markdown (.md) files. Please verify your changes by starting the 
build script in preview mode.
+
+cd docs
+./build_docs.sh -p
+
+
+The script compiles the Markdown files into static HTML pages and starts a 
local webserver. Open your browser at http://localhost:4000 to 
view the compiled documentation including your changes. The served 
documentation is automatically re-compiled and updated when you modify and save 
Markdown files and refresh your browser.
+
+Please feel free to ask any questions you have on the developer mailing 
list.
+
+Submit your contribution
+
+The Flink project accepts documentation contributions through the https://github.com/apache/flink;>GitHub Mirror as https://help.github.com/articles/using-pull-requests;>Pull Requests. 
Pull requests are a simple way of offering a 

[36/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/blog/release_1.0.0-changelog_known_issues.html
--
diff --git a/content/blog/release_1.0.0-changelog_known_issues.html 
b/content/blog/release_1.0.0-changelog_known_issues.html
new file mode 100644
index 000..9b7ddc7
--- /dev/null
+++ b/content/blog/release_1.0.0-changelog_known_issues.html
@@ -0,0 +1,1139 @@
+
+
+  
+
+
+
+
+Apache Flink: Release 1.0.0 – Changelog and Known Issues
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+Release 1.0.0 – Changelog and Known Issues
+
+   
+  Changelog
+  Sub-task
+  Bug
+  Improvement
+  Task
+  Test
+  Wish
+
+  
+  Known 
Issues
+
+
+Changelog
+
+Sub-task
+
+
+[https://issues.apache.org/jira/browse/FLINK-7;>FLINK-7] -
 [GitHub] Enable Range Partitioner
+
+[https://issues.apache.org/jira/browse/FLINK-146;>FLINK-146] 
- Sorted output not working
+
+[https://issues.apache.org/jira/browse/FLINK-1982;>FLINK-1982] -   
  Remove dependencies on Record for Flink runtime and core
+
+[https://issues.apache.org/jira/browse/FLINK-2676;>FLINK-2676] -   
  Add abstraction for keyed window state
+
+[https://issues.apache.org/jira/browse/FLINK-2680;>FLINK-2680] -   
  Create a dedicated aligned-event time window operator
+
+[https://issues.apache.org/jira/browse/FLINK-2730;>FLINK-2730] -   
  Add CPU/Network utilization graphs to new web dashboard
+
+[https://issues.apache.org/jira/browse/FLINK-2732;>FLINK-2732] -   
  Add access to the TaskManagers log file and out file in the web 
dashboard.
+
+[https://issues.apache.org/jira/browse/FLINK-2850;>FLINK-2850] -   
  Limit the types of jobs which can run in detached mode
+
+[https://issues.apache.org/jira/browse/FLINK-2853;>FLINK-2853] -   
  Apply JMH on MutableHashTablePerformanceBenchmark class.
+
+[https://issues.apache.org/jira/browse/FLINK-2869;>FLINK-2869] -   
  Apply JMH on IOManagerPerformanceBenchmark class.
+
+[https://issues.apache.org/jira/browse/FLINK-2889;>FLINK-2889] -   
  Apply JMH on LongSerializationSpeedBenchmark class
+
+[https://issues.apache.org/jira/browse/FLINK-2890;>FLINK-2890] -   
  Apply JMH on StringSerializationSpeedBenchmark class.
+
+[https://issues.apache.org/jira/browse/FLINK-2900;>FLINK-2900] -   
  Remove Record-API dependencies from Hadoop Compat module
+
+[https://issues.apache.org/jira/browse/FLINK-2901;>FLINK-2901] -   
  Several flink-test ITCases depend on Record API features
+
+[https://issues.apache.org/jira/browse/FLINK-2906;>FLINK-2906] -   
  Remove Record-API
+
+[https://issues.apache.org/jira/browse/FLINK-2919;>FLINK-2919] -   
  Apply JMH on FieldAccessMinibenchmark class.
+
+[https://issues.apache.org/jira/browse/FLINK-2920;>FLINK-2920] -   
  Apply JMH on KryoVersusAvroMinibenchmark class.
+
+[https://issues.apache.org/jira/browse/FLINK-2933;>FLINK-2933] -   
  Flink scala libraries exposed with maven should carry scala version
+

[05/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/js/dagre-d3.js
--
diff --git a/content/visualizer/js/dagre-d3.js 
b/content/visualizer/js/dagre-d3.js
new file mode 100755
index 000..482ce82
--- /dev/null
+++ b/content/visualizer/js/dagre-d3.js
@@ -0,0 +1,4560 @@
+;(function e(t,n,r){function s(o,u){if(!n[o]){if(!t[o]){var a=typeof 
require=="function"&if(!u&)return a(o,!0);if(i)return i(o,!0);throw 
new Error("Cannot find module '"+o+"'")}var 
f=n[o]={exports:{}};t[o][0].call(f.exports,function(e){var n=t[o][1][e];return 
s(n?n:e)},f,f.exports,e,t,n,r)}return n[o].exports}var i=typeof 
require=="function"&for(var o=0;o

[21/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/svg/white_filled.svg
--
diff --git a/content/img/logo/svg/white_filled.svg 
b/content/img/logo/svg/white_filled.svg
new file mode 100755
index 000..28bd9f9
--- /dev/null
+++ b/content/img/logo/svg/white_filled.svg
@@ -0,0 +1,231 @@
+
+
+http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd;>
+http://www.w3.org/2000/svg; 
xmlns:xlink="http://www.w3.org/1999/xlink; x="0px" y="0px"
+width="2000px" height="1280px" viewBox="0 0 2000 1280" 
enable-background="new 0 0 2000 1280" xml:space="preserve">
+
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+
+
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+
+
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+
+
+   
+   
+
+
+   
+   
+   
+   
+   
+   
+   
+   
+   
+
+Filled white
+

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/managed-state.png
--
diff --git a/content/img/managed-state.png b/content/img/managed-state.png
new file mode 100755
index 000..3dbb2fd
Binary files /dev/null and b/content/img/managed-state.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/memory_heap_division.png
--
diff --git a/content/img/memory_heap_division.png 
b/content/img/memory_heap_division.png
new file mode 100644
index 000..2b4c2e2
Binary files /dev/null and b/content/img/memory_heap_division.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/navbar-brand-logo.jpg
--
diff --git a/content/img/navbar-brand-logo.jpg 
b/content/img/navbar-brand-logo.jpg
new file mode 100755
index 000..5993ee8
Binary files /dev/null and b/content/img/navbar-brand-logo.jpg differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/navbar-brand-logo.png
--
diff --git a/content/img/navbar-brand-logo.png 
b/content/img/navbar-brand-logo.png
new file mode 100644
index 000..152f74e
Binary files /dev/null and b/content/img/navbar-brand-logo.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/one_runtime.png
--
diff --git a/content/img/one_runtime.png b/content/img/one_runtime.png
new file mode 100644
index 000..9cb4363
Binary files /dev/null and b/content/img/one_runtime.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/optimizer_choice.png
--
diff --git a/content/img/optimizer_choice.png b/content/img/optimizer_choice.png
new file mode 100644
index 000..1f8004b
Binary files /dev/null and b/content/img/optimizer_choice.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/otto-group-logo.jpg
--
diff --git a/content/img/otto-group-logo.jpg b/content/img/otto-group-logo.jpg
new file mode 100644
index 000..f578af6
Binary files /dev/null and b/content/img/otto-group-logo.jpg differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/out_of_order_stream.png
--
diff --git a/content/img/out_of_order_stream.png 
b/content/img/out_of_order_stream.png
new file mode 100644
index 000..20ad09d
Binary files /dev/null and b/content/img/out_of_order_stream.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/parallel_dataflows.png
--
diff --git a/content/img/parallel_dataflows.png 
b/content/img/parallel_dataflows.png
new file mode 100644
index 000..dbd83b2
Binary files /dev/null and b/content/img/parallel_dataflows.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/poweredby/alibaba-logo.png
--
diff --git 

[12/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2016/03/08/release-1.0.0.html
--
diff --git a/content/news/2016/03/08/release-1.0.0.html 
b/content/news/2016/03/08/release-1.0.0.html
new file mode 100644
index 000..13313d8
--- /dev/null
+++ b/content/news/2016/03/08/release-1.0.0.html
@@ -0,0 +1,319 @@
+
+
+  
+
+
+
+
+Apache Flink: Announcing Apache Flink 1.0.0
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+
+  Announcing Apache Flink 1.0.0
+
+  
+08 Mar 2016
+
+The Apache Flink community is pleased to announce the availability of the 
1.0.0 release. The community put significant effort into improving and 
extending Apache Flink since the last release, focusing on improving the 
experience of writing and executing data stream processing pipelines in 
production.
+
+
+
+
+
+Flink version 1.0.0 marks the beginning of the 1.X.X series of releases, 
which will maintain backwards compatibility with 1.0.0. This means that 
applications written against stable APIs of Flink 1.0.0 will compile and run 
with all Flink versions in the 1. series. This is the first time we are 
formally guaranteeing compatibility in Flink’s history, and we therefore see 
this release as a major milestone of the project, perhaps the most important 
since graduation as a top-level project.
+
+Apart from backwards compatibility, Flink 1.0.0 brings a variety of new 
user-facing features, as well as tons of bug fixes. About 64 contributors 
provided bug fixes, improvements, and new features such that in total more than 
450 JIRA issues could be resolved.
+
+We encourage everyone to http://flink.apache.org/downloads.html;>download the release and https://ci.apache.org/projects/flink/flink-docs-release-1.0/;>check out 
the documentation. Feedback through the Flink http://flink.apache.org/community.html#mailing-lists;>mailing lists 
is, as always, very welcome!
+
+Interface stability annotations
+
+Flink 1.0.0 introduces interface stability annotations for API classes and 
methods. Interfaces defined as @Public are guaranteed to remain 
stable across all releases of the 1.x series. The @PublicEvolving 
annotation marks API features that may be subject to change in future 
versions.
+
+Flink’s stability annotations will help users to implement applications 
that compile and execute unchanged against future versions of Flink 1.x. This 
greatly reduces the complexity for users when upgrading to a newer Flink 
release.
+
+Out-of-core state support
+
+Flink 1.0.0 adds a new state backend that uses RocksDB to store state (both 
windows and user-defined key-value state). http://rocksdb.org/;>RocksDB is an embedded key/value store database, 
originally developed by Facebook.
+When using this backend, active state in streaming programs can grow well 
beyond memory. The RocksDB files are stored in a distributed file system such 
as HDFS or S3 for backups.
+
+Savepoints and version upgrades
+
+Savepoints are checkpoints of the state of a running streaming job that can 
be manually 

[14/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/09/16/off-heap-memory.html
--
diff --git a/content/news/2015/09/16/off-heap-memory.html 
b/content/news/2015/09/16/off-heap-memory.html
new file mode 100644
index 000..60d7bda
--- /dev/null
+++ b/content/news/2015/09/16/off-heap-memory.html
@@ -0,0 +1,1083 @@
+
+
+  
+
+
+
+
+Apache Flink: Off-heap Memory in Apache Flink and the curious JIT 
compiler
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+
+  Off-heap Memory in Apache Flink and the curious JIT compiler
+
+  
+16 Sep 2015 by Stephan Ewen (https://twitter.com/stephanewen;>@stephanewen)
+
+Running data-intensive code in the JVM and making it well-behaved is 
tricky. Systems that put billions of data objects naively onto the JVM heap 
face unpredictable OutOfMemoryErrors and Garbage Collection stalls. Of course, 
you still want to to keep your data in memory as much as possible, for speed 
and responsiveness of the processing applications. In that context, 
“off-heap” has become almost something like a magic word to solve these 
problems.
+
+In this blog post, we will look at how Flink exploits off-heap memory. The 
feature is part of the upcoming release, but you can try it out with the latest 
nightly builds. We will also give a few interesting insights into the behavior 
for Java’s JIT compiler for highly optimized methods and loops.
+
+Recap: Memory Management in 
Flink
+
+To understand Flink’s approach to off-heap memory, we need to recap 
Flink’s approach to custom managed memory. We have written an earlier blog post 
about how Flink manages JVM memory itself
+
+As a summary, the core part is that Flink implements its algorithms not 
against Java objects, arrays, or lists, but actually against a data structure 
similar to java.nio.ByteBuffer. Flink uses its own specialized 
version, called https://github.com/apache/flink/blob/release-0.9.1-rc1/flink-core/src/main/java/org/apache/flink/core/memory/MemorySegment.java;>MemorySegment
 on which algorithms put and get at specific positions ints, longs, byte 
arrays, etc, and compare and copy memory. The memory segments are held and 
distributed by a central component (called MemoryManager) from 
which algorithms request segments according to their calculated memory 
budgets.
+
+Don’t believe that this can be fast? Have a look at the benchmarks in the 
earlier blogpost, which show that it is actually often much faster than 
working on objects, due to better control over data layout (cache efficiency, 
data size), and reducing the pressure on Java’s Garbage Collector.
+
+This form of memory management has been in Flink for a long time. 
Anecdotally, the first public demo of Flink’s predecessor project 
Stratosphere, at the VLDB conference in 2010, was running its programs 
with custom managed memory (although I believe few attendees were aware of 
that).
+
+Why actually bother with 
off-heap memory?
+
+Given that Flink has a sophisticated level of managing on-heap memory, 

[29/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/commit-stats.png
--
diff --git a/content/img/blog/commit-stats.png 
b/content/img/blog/commit-stats.png
new file mode 100644
index 000..dfeb7b3
Binary files /dev/null and b/content/img/blog/commit-stats.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/community-growth.png
--
diff --git a/content/img/blog/community-growth.png 
b/content/img/blog/community-growth.png
new file mode 100644
index 000..e37df51
Binary files /dev/null and b/content/img/blog/community-growth.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/data-serialization.png
--
diff --git a/content/img/blog/data-serialization.png 
b/content/img/blog/data-serialization.png
new file mode 100755
index 000..80667f6
Binary files /dev/null and b/content/img/blog/data-serialization.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/emr-firefoxsettings.png
--
diff --git a/content/img/blog/emr-firefoxsettings.png 
b/content/img/blog/emr-firefoxsettings.png
new file mode 100755
index 000..cc515a4
Binary files /dev/null and b/content/img/blog/emr-firefoxsettings.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/emr-hadoopversion.png
--
diff --git a/content/img/blog/emr-hadoopversion.png 
b/content/img/blog/emr-hadoopversion.png
new file mode 100755
index 000..7646b75
Binary files /dev/null and b/content/img/blog/emr-hadoopversion.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/emr-jobmanager.png
--
diff --git a/content/img/blog/emr-jobmanager.png 
b/content/img/blog/emr-jobmanager.png
new file mode 100755
index 000..4cbe5cc
Binary files /dev/null and b/content/img/blog/emr-jobmanager.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/emr-running.png
--
diff --git a/content/img/blog/emr-running.png b/content/img/blog/emr-running.png
new file mode 100755
index 000..44d21d2
Binary files /dev/null and b/content/img/blog/emr-running.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/emr-security.png
--
diff --git a/content/img/blog/emr-security.png 
b/content/img/blog/emr-security.png
new file mode 100755
index 000..fee4fed
Binary files /dev/null and b/content/img/blog/emr-security.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/emr-yarnappmaster.png
--
diff --git a/content/img/blog/emr-yarnappmaster.png 
b/content/img/blog/emr-yarnappmaster.png
new file mode 100755
index 000..62be9ed
Binary files /dev/null and b/content/img/blog/emr-yarnappmaster.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/feature-timeline.png
--
diff --git a/content/img/blog/feature-timeline.png 
b/content/img/blog/feature-timeline.png
new file mode 100644
index 000..1f56747
Binary files /dev/null and b/content/img/blog/feature-timeline.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/ff-speakers.png
--
diff --git a/content/img/blog/ff-speakers.png b/content/img/blog/ff-speakers.png
new file mode 100644
index 000..429a2c8
Binary files /dev/null and b/content/img/blog/ff-speakers.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/flink-1.0.png
--
diff --git a/content/img/blog/flink-1.0.png b/content/img/blog/flink-1.0.png
new file mode 100644
index 000..88a006b
Binary files /dev/null and b/content/img/blog/flink-1.0.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/flink-dow-2016.png
--
diff --git a/content/img/blog/flink-dow-2016.png 
b/content/img/blog/flink-dow-2016.png
new file mode 100644
index 000..c90d768
Binary files /dev/null and b/content/img/blog/flink-dow-2016.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/flink-forward-banner.png

[26/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/colors/flink_colors.pptx
--
diff --git a/content/img/logo/colors/flink_colors.pptx 
b/content/img/logo/colors/flink_colors.pptx
new file mode 100644
index 000..8ecd0be
Binary files /dev/null and b/content/img/logo/colors/flink_colors.pptx differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/100/flink_squirrel_100_black.png
--
diff --git a/content/img/logo/png/100/flink_squirrel_100_black.png 
b/content/img/logo/png/100/flink_squirrel_100_black.png
new file mode 100755
index 000..6869f58
Binary files /dev/null and 
b/content/img/logo/png/100/flink_squirrel_100_black.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/100/flink_squirrel_100_color.png
--
diff --git a/content/img/logo/png/100/flink_squirrel_100_color.png 
b/content/img/logo/png/100/flink_squirrel_100_color.png
new file mode 100755
index 000..c508e1e
Binary files /dev/null and 
b/content/img/logo/png/100/flink_squirrel_100_color.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/100/flink_squirrel_100_white.png
--
diff --git a/content/img/logo/png/100/flink_squirrel_100_white.png 
b/content/img/logo/png/100/flink_squirrel_100_white.png
new file mode 100755
index 000..088fb27
Binary files /dev/null and 
b/content/img/logo/png/100/flink_squirrel_100_white.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink1000_black.png
--
diff --git a/content/img/logo/png/1000/flink1000_black.png 
b/content/img/logo/png/1000/flink1000_black.png
new file mode 100755
index 000..31af663
Binary files /dev/null and b/content/img/logo/png/1000/flink1000_black.png 
differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink1000_color_black.png
--
diff --git a/content/img/logo/png/1000/flink1000_color_black.png 
b/content/img/logo/png/1000/flink1000_color_black.png
new file mode 100755
index 000..2f4b991
Binary files /dev/null and 
b/content/img/logo/png/1000/flink1000_color_black.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink1000_color_white.png
--
diff --git a/content/img/logo/png/1000/flink1000_color_white.png 
b/content/img/logo/png/1000/flink1000_color_white.png
new file mode 100755
index 000..cd0b56e
Binary files /dev/null and 
b/content/img/logo/png/1000/flink1000_color_white.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink1000_white.png
--
diff --git a/content/img/logo/png/1000/flink1000_white.png 
b/content/img/logo/png/1000/flink1000_white.png
new file mode 100755
index 000..3b7b253
Binary files /dev/null and b/content/img/logo/png/1000/flink1000_white.png 
differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink_squirrel_1000.png
--
diff --git a/content/img/logo/png/1000/flink_squirrel_1000.png 
b/content/img/logo/png/1000/flink_squirrel_1000.png
new file mode 100755
index 000..37c980e
Binary files /dev/null and b/content/img/logo/png/1000/flink_squirrel_1000.png 
differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink_squirrel_black_1000.png
--
diff --git a/content/img/logo/png/1000/flink_squirrel_black_1000.png 
b/content/img/logo/png/1000/flink_squirrel_black_1000.png
new file mode 100755
index 000..b68621d
Binary files /dev/null and 
b/content/img/logo/png/1000/flink_squirrel_black_1000.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink_squirrel_white_1000.png
--
diff --git a/content/img/logo/png/1000/flink_squirrel_white_1000.png 
b/content/img/logo/png/1000/flink_squirrel_white_1000.png
new file mode 100755
index 000..14c7350
Binary files /dev/null and 
b/content/img/logo/png/1000/flink_squirrel_white_1000.png differ

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/200/flink2_200_black.png
--
diff --git 

[08/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/css/bootstrap.css
--
diff --git a/content/visualizer/css/bootstrap.css 
b/content/visualizer/css/bootstrap.css
new file mode 100755
index 000..fc41022
--- /dev/null
+++ b/content/visualizer/css/bootstrap.css
@@ -0,0 +1,5785 @@
+/*!
+ * Bootstrap v3.1.1 (http://getbootstrap.com)
+ * Copyright 2011-2014 Twitter, Inc.
+ * Licensed under MIT (https://github.com/twbs/bootstrap/blob/master/LICENSE)
+ */
+
+/*! normalize.css v3.0.0 | MIT License | git.io/normalize */
+html {
+  font-family: sans-serif;
+  -webkit-text-size-adjust: 100%;
+  -ms-text-size-adjust: 100%;
+}
+body {
+  margin: 0;
+}
+article,
+aside,
+details,
+figcaption,
+figure,
+footer,
+header,
+hgroup,
+main,
+nav,
+section,
+summary {
+  display: block;
+}
+audio,
+canvas,
+progress,
+video {
+  display: inline-block;
+  vertical-align: baseline;
+}
+audio:not([controls]) {
+  display: none;
+  height: 0;
+}
+[hidden],
+template {
+  display: none;
+}
+a {
+  background: transparent;
+}
+a:active,
+a:hover {
+  outline: 0;
+}
+abbr[title] {
+  border-bottom: 1px dotted;
+}
+b,
+strong {
+  font-weight: bold;
+}
+dfn {
+  font-style: italic;
+}
+h1 {
+  margin: .67em 0;
+  font-size: 2em;
+}
+mark {
+  color: #000;
+  background: #ff0;
+}
+small {
+  font-size: 80%;
+}
+sub,
+sup {
+  position: relative;
+  font-size: 75%;
+  line-height: 0;
+  vertical-align: baseline;
+}
+sup {
+  top: -.5em;
+}
+sub {
+  bottom: -.25em;
+}
+img {
+  border: 0;
+}
+svg:not(:root) {
+  overflow: hidden;
+}
+figure {
+  margin: 1em 40px;
+}
+hr {
+  height: 0;
+  -moz-box-sizing: content-box;
+   box-sizing: content-box;
+}
+pre {
+  overflow: auto;
+}
+code,
+kbd,
+pre,
+samp {
+  font-family: monospace, monospace;
+  font-size: 1em;
+}
+button,
+input,
+optgroup,
+select,
+textarea {
+  margin: 0;
+  font: inherit;
+  color: inherit;
+}
+button {
+  overflow: visible;
+}
+button,
+select {
+  text-transform: none;
+}
+button,
+html input[type="button"],
+input[type="reset"],
+input[type="submit"] {
+  -webkit-appearance: button;
+  cursor: pointer;
+}
+button[disabled],
+html input[disabled] {
+  cursor: default;
+}
+button::-moz-focus-inner,
+input::-moz-focus-inner {
+  padding: 0;
+  border: 0;
+}
+input {
+  line-height: normal;
+}
+input[type="checkbox"],
+input[type="radio"] {
+  box-sizing: border-box;
+  padding: 0;
+}
+input[type="number"]::-webkit-inner-spin-button,
+input[type="number"]::-webkit-outer-spin-button {
+  height: auto;
+}
+input[type="search"] {
+  -webkit-box-sizing: content-box;
+ -moz-box-sizing: content-box;
+  box-sizing: content-box;
+  -webkit-appearance: textfield;
+}
+input[type="search"]::-webkit-search-cancel-button,
+input[type="search"]::-webkit-search-decoration {
+  -webkit-appearance: none;
+}
+fieldset {
+  padding: .35em .625em .75em;
+  margin: 0 2px;
+  border: 1px solid #c0c0c0;
+}
+legend {
+  padding: 0;
+  border: 0;
+}
+textarea {
+  overflow: auto;
+}
+optgroup {
+  font-weight: bold;
+}
+table {
+  border-spacing: 0;
+  border-collapse: collapse;
+}
+td,
+th {
+  padding: 0;
+}
+@media print {
+  * {
+color: #000 !important;
+text-shadow: none !important;
+background: transparent !important;
+box-shadow: none !important;
+  }
+  a,
+  a:visited {
+text-decoration: underline;
+  }
+  a[href]:after {
+content: " (" attr(href) ")";
+  }
+  abbr[title]:after {
+content: " (" attr(title) ")";
+  }
+  a[href^="javascript:"]:after,
+  a[href^="#"]:after {
+content: "";
+  }
+  pre,
+  blockquote {
+border: 1px solid #999;
+
+page-break-inside: avoid;
+  }
+  thead {
+display: table-header-group;
+  }
+  tr,
+  img {
+page-break-inside: avoid;
+  }
+  img {
+max-width: 100% !important;
+  }
+  p,
+  h2,
+  h3 {
+orphans: 3;
+widows: 3;
+  }
+  h2,
+  h3 {
+page-break-after: avoid;
+  }
+  select {
+background: #fff !important;
+  }
+  .navbar {
+display: none;
+  }
+  .table td,
+  .table th {
+background-color: #fff !important;
+  }
+  .btn > .caret,
+  .dropup > .btn > .caret {
+border-top-color: #000 !important;
+  }
+  .label {
+border: 1px solid #000;
+  }
+  .table {
+border-collapse: collapse !important;
+  }
+  .table-bordered th,
+  .table-bordered td {
+border: 1px solid #ddd !important;
+  }
+}
+* {
+  -webkit-box-sizing: border-box;
+ -moz-box-sizing: border-box;
+  box-sizing: border-box;
+}
+*:before,
+*:after {
+  -webkit-box-sizing: border-box;
+ -moz-box-sizing: border-box;
+  box-sizing: border-box;
+}
+html {
+  font-size: 62.5%;
+
+  -webkit-tap-highlight-color: rgba(0, 0, 0, 0);
+}
+body {
+  font-family: "Helvetica Neue", Helvetica, Arial, sans-serif;
+  font-size: 14px;
+  line-height: 1.42857143;
+  color: #333;
+  background-color: #fff;
+}
+input,
+button,
+select,
+textarea {
+  font-family: inherit;
+  

[32/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/faq.html
--
diff --git a/content/faq.html b/content/faq.html
new file mode 100644
index 000..7653ff8
--- /dev/null
+++ b/content/faq.html
@@ -0,0 +1,683 @@
+
+
+  
+
+
+
+
+Apache Flink: Frequently Asked Questions (FAQ)
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project 
FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+Frequently Asked Questions (FAQ)
+
+   
+
+The following questions are frequently asked with regard to the Flink 
project in general. If you have further questions, make sure 
to consult the documentation or ask the 
community.
+
+
+
+  General
+  Is Flink a Hadoop Project?
+  Do I have to 
install Apache Hadoop to use Flink?
+
+  
+  Usage
+  How do I 
assess the progress of a Flink program?
+  How can I figure 
out why a program failed?
+  How do I debug Flink 
programs?
+  What is the 
parallelism? How do I set it?
+
+  
+  Errors
+  Why am I getting 
a “NonSerializableException” ?
+  In
 Scala API, I get an error about implicit values and evidence 
parameters
+  I
 get an error message saying that not enough buffers are available. How do I 
fix this?
+  My
 job fails early with a java.io.EOFException. What could be the cause?
+  My
 job fails with various exceptions from the HDFS/Hadoop code. What can I 
do?
+  In 
Eclipse, I get compilation errors in the Scala projects
+  My
 program does not compute the correct result. Why are my custom key 
types
+  I
 get a java.lang.InstantiationException for my data type, what is 
wrong?
+  I
 can’t stop Flink with the provided stop-scripts. What can I do?
+  I got an 
OutOfMemoryException. What can I do?
+  Why do the 
TaskManager log files become so huge?
+  The
 slot allocated for my task manager has been released. What should I 
do?
+
+  
+  YARN 
Deployment
+  The YARN session 
runs only for a few seconds
+  My
 YARN containers are killed because they use too much memory
+  The
 YARN session crashes with a HDFS permission exception during startup
+  My job is not 
reacting to a job cancellation?
+
+  
+  Features
+  What kind of 
fault-tolerance does Flink provide?
+  Are
 Hadoop-like utilities, such as Counters and the DistributedCache 
supported?
+
+  
+
+
+
+
+General
+
+Is Flink a Hadoop Project?
+
+Flink is a data processing system and an alternative to Hadoop’s
+MapReduce component. It comes with its own runtime rather 
than building on top
+of MapReduce. As such, it can work completely independently of the Hadoop
+ecosystem. However, Flink can also access Hadoop’s distributed file
+system (HDFS) to read and write data, and Hadoop’s next-generation resource
+manager (YARN) to provision cluster resources. Since most Flink users are
+using Hadoop HDFS to store their data, Flink already ships the required 
libraries to
+access HDFS.
+
+Do I have to install 
Apache Hadoop to use Flink?
+
+No. Flink can run without a Hadoop 
installation. However, a very 

[13/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/12/04/Introducing-windows.html
--
diff --git a/content/news/2015/12/04/Introducing-windows.html 
b/content/news/2015/12/04/Introducing-windows.html
new file mode 100644
index 000..aa0834f
--- /dev/null
+++ b/content/news/2015/12/04/Introducing-windows.html
@@ -0,0 +1,349 @@
+
+
+  
+
+
+
+
+Apache Flink: Introducing Stream Windows in Apache Flink
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+
+  Introducing Stream Windows in Apache Flink
+
+  
+04 Dec 2015 by Fabian Hueske (https://twitter.com/fhueske;>@fhueske)
+
+The data analysis space is witnessing an evolution from batch to stream 
processing for many use cases. Although batch can be handled as a special case 
of stream processing, analyzing never-ending streaming data often requires a 
shift in the mindset and comes with its own terminology (for example, 
“windowing” and “at-least-once”/”exactly-once” processing). This 
shift and the new terminology can be quite confusing for people being new to 
the space of stream processing. Apache Flink is a production-ready stream 
processor with an easy-to-use yet very expressive API to define advanced stream 
analysis programs. Flink’s API features very flexible window definitions on 
data streams which let it stand out among other open source stream 
processors.
+
+In this blog post, we discuss the concept of windows for stream processing, 
present Flink’s built-in windows, and explain its support for custom 
windowing semantics.
+
+What are windows and what 
are they good for?
+
+Consider the example of a traffic sensor that counts every 15 seconds the 
number of vehicles passing a certain location. The resulting stream could look 
like:
+
+
+
+
+
+If you would like to know, how many vehicles passed that location, you 
would simply sum the individual counts. However, the nature of a sensor stream 
is that it continuously produces data. Such a stream never ends and it is not 
possible to compute a final sum that can be returned. Instead, it is possible 
to compute rolling sums, i.e., return for each input event an updated sum 
record. This would yield a new stream of partial sums.
+
+
+
+
+
+However, a stream of partial sums might not be what we are looking for, 
because it constantly updates the count and even more important, some 
information such as variation over time is lost. Hence, we might want to 
rephrase our question and ask for the number of cars that pass the location 
every minute. This requires us to group the elements of the stream into finite 
sets, each set corresponding to sixty seconds. This operation is called a 
tumbling windows operation.
+
+
+
+
+
+Tumbling windows discretize a stream into non-overlapping windows. For 
certain applications it is important that windows are not disjunct because an 
application might require smoothed aggregates. For example, we can compute 
every thirty seconds the number of cars passed in the last minute. Such windows 

[24/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/svg/color_white.svg
--
diff --git a/content/img/logo/svg/color_white.svg 
b/content/img/logo/svg/color_white.svg
new file mode 100755
index 000..e24ac1c
--- /dev/null
+++ b/content/img/logo/svg/color_white.svg
@@ -0,0 +1,1563 @@
+
+
+http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd;>
+http://www.w3.org/2000/svg; 
xmlns:xlink="http://www.w3.org/1999/xlink; x="0px" y="0px"
+width="2000px" height="1280px" viewBox="0 0 2000 1280" 
enable-background="new 0 0 2000 1280" xml:space="preserve">
+
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+
+
+   
+   
+
+
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   

[15/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/08/24/introducing-flink-gelly.html
--
diff --git a/content/news/2015/08/24/introducing-flink-gelly.html 
b/content/news/2015/08/24/introducing-flink-gelly.html
new file mode 100644
index 000..714e75a
--- /dev/null
+++ b/content/news/2015/08/24/introducing-flink-gelly.html
@@ -0,0 +1,649 @@
+
+
+  
+
+
+
+
+Apache Flink: Introducing Gelly: Graph Processing with Apache 
Flink
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+
+  Introducing Gelly: Graph Processing with Apache Flink
+
+  
+24 Aug 2015
+
+This blog post introduces Gelly, Apache Flink’s 
graph-processing API and library. Flink’s native support
+for iterations makes it a suitable platform for large-scale graph analytics.
+By leveraging delta iterations, Gelly is able to map various graph processing 
models such as
+vertex-centric or gather-sum-apply to Flink dataflows.
+
+Gelly allows Flink users to perform end-to-end data analysis in a single 
system.
+Gelly can be seamlessly used with Flink’s DataSet API,
+which means that pre-processing, graph creation, analysis, and post-processing 
can be done
+in the same application. At the end of this post, we will go through a 
step-by-step example
+in order to demonstrate that loading, transformation, filtering, graph 
creation, and analysis
+can be performed in a single Flink program.
+
+Overview
+
+
+  What is Gelly?
+  Graph Representation and 
Creation
+  Transformations and 
Utilities
+  Iterative Graph Processing
+  Library of Graph 
Algorithms
+  Use-Case: Music Profiles
+  Ongoing and Future Work
+
+
+
+
+What is Gelly?
+
+Gelly is a Graph API for Flink. It is currently supported in both Java and 
Scala.
+The Scala methods are implemented as wrappers on top of the basic Java 
operations.
+The API contains a set of utility functions for graph analysis, supports 
iterative graph
+processing and introduces a library of graph algorithms.
+
+
+
+
+
+Back to top
+
+Graph Representation and 
Creation
+
+In Gelly, a graph is represented by a DataSet of vertices and a DataSet of 
edges.
+A vertex is defined by its unique ID and a value, whereas an edge is defined 
by its source ID,
+target ID, and value. A vertex or edge for which a value is not specified will 
simply have the
+value type set to NullValue.
+
+A graph can be created from:
+
+
+  DataSet of edges and an optional DataSet of 
vertices using Graph.fromDataSet()
+  DataSet of Tuple3 and an optional DataSet of 
Tuple2 using Graph.fromTupleDataSet()
+  Collection of edges and an optional Collection 
of vertices using Graph.fromCollection()
+
+
+In all three cases, if the vertices are not provided,
+Gelly will automatically produce the vertex IDs from the edge source and 
target IDs.
+
+Back to top
+
+Transformations and Utilities
+
+These are methods of the Graph class and include common graph metrics, 
transformations
+and mutations as well as neighborhood aggregations.
+
+Common Graph Metrics
+These methods can be used to 

[22/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/svg/flink_logotypes.svg
--
diff --git a/content/img/logo/svg/flink_logotypes.svg 
b/content/img/logo/svg/flink_logotypes.svg
new file mode 100755
index 000..006243a
--- /dev/null
+++ b/content/img/logo/svg/flink_logotypes.svg
@@ -0,0 +1,3835 @@
+
+
+http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd;>
+http://www.w3.org/2000/svg; 
xmlns:xlink="http://www.w3.org/1999/xlink; x="0px" y="0px"
+width="799.736px" height="803.173px" viewBox="168.959 68.626 799.736 
803.173"
+enable-background="new 168.959 68.626 799.736 803.173" 
xml:space="preserve">
+
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+
+
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+
+
+   
+   
+
+
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   
+   

[30/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/cep-monitoring.svg
--
diff --git a/content/img/blog/cep-monitoring.svg 
b/content/img/blog/cep-monitoring.svg
new file mode 100644
index 000..02cca81
--- /dev/null
+++ b/content/img/blog/cep-monitoring.svg
@@ -0,0 +1,2838 @@
+
+
+
+http://purl.org/dc/elements/1.1/;
+   xmlns:cc="http://creativecommons.org/ns#;
+   xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#;
+   xmlns:svg="http://www.w3.org/2000/svg;
+   xmlns="http://www.w3.org/2000/svg;
+   xmlns:xlink="http://www.w3.org/1999/xlink;
+   xmlns:sodipodi="http://sodipodi.sourceforge.net/DTD/sodipodi-0.dtd;
+   xmlns:inkscape="http://www.inkscape.org/namespaces/inkscape;
+   width="210mm"
+   height="148mm"
+   viewBox="0 0 744.09449 524.40944"
+   id="svg2"
+   version="1.1"
+   inkscape:version="0.91 r13725"
+   sodipodi:docname="CEP_final.svg">
+  
+
+  
+
+
+  
+
+
+  
+
+
+  
+
+
+  
+
+
+  
+
+
+  
+
+
+  
+
+
+  
+
+
+  
+
+
+  
+
+
+  
+
+
+  
+
+
+  
+
+
+  
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+  
+  
+
+  
+  
+
+  
+image/svg+xml
+http://purl.org/dc/dcmitype/StillImage; />
+
+  
+
+  
+  
+
+  
+
+
+
+
+  
+
+
+  
+  
+  
+  
+  
+  
+  
+  
+  
+  
+
+
+
+
+  
+  
+  
+  
+  
+  
+  
+  
+  
+  
+
+
+
+
+
+
+
+  
+  
+  
+
+
+
+
+
+
+
+  
+  
+
+
+
+  
+  
+
+
+
+
+  
+  
+
+
+
+
+  
+  
+
+
+
+
+
+
+  
+
+
+  
+  
+
+
+
+  
+  
+  
+
+
+
+  
+  
+  
+
+
+
+
+
+
+
+  
+  
+
+
+
+
+  
+
+
+  
+  
+  
+
+
+  
+  
+
+
+
+
+  
+  
+
+
+
+  
+
+
+  
+  
+  
+
+
+
+
+  
+  
+  
+
+
+
+
+
+  
+  
+  
+
+
+
+  
+  
+  
+
+
+
+
+
+
+  
+  
+  
+
+
+
+  
+
+
+  
+
+
+
+http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/code-growth.png
--
diff --git a/content/img/blog/code-growth.png b/content/img/blog/code-growth.png
new file mode 100644
index 000..196c907
Binary files /dev/null and b/content/img/blog/code-growth.png differ



[flink-web] Git Push Summary [forced push!] [Forced Update!]

2017-01-18 Thread rmetzger
Repository: flink-web
Updated Branches:
  refs/heads/asf-site 9ec0a879f -> 4bb81d3f1 (forced update)


[31/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/how-to-contribute.html
--
diff --git a/content/how-to-contribute.html b/content/how-to-contribute.html
new file mode 100644
index 000..87ad4a0
--- /dev/null
+++ b/content/how-to-contribute.html
@@ -0,0 +1,345 @@
+
+
+  
+
+
+
+
+Apache Flink: How To Contribute
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to 
Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+How To Contribute
+
+   Apache Flink is developed by an open and friendly community. 
Everybody is cordially welcome to join the community and contribute to Apache 
Flink. There are several ways to interact with the community and to contribute 
to Flink including asking questions, filing bug reports, proposing new 
features, joining discussions on the mailing lists, contributing code or 
documentation, improving the website, or testing release candidates.
+
+
+
+  Ask 
questions!
+  File a 
bug report
+  Propose an 
improvement or a new feature
+  Help others and join the 
discussions
+  Test a release candidate
+  Contribute 
code
+  Contribute documentation
+  Improve the website
+  More ways to contribute…
+  Submit a Contributor 
License Agreement
+  How to become a committer
+
+
+
+
+Ask questions!
+
+The Apache Flink community is eager to help and to answer your questions. 
We have a user mailing list, hang 
out in an IRC channel, and watch Stack 
Overflow on the http://stackoverflow.com/questions/tagged/apache-flink;>[apache-flink]
 tag.
+
+
+
+File a bug report
+
+Please let us know if you experienced a problem with Flink and file a bug 
report. Open http://issues.apache.org/jira/browse/FLINK;>Flink’s 
JIRA and click on the blue Create button at the top. 
Please give detailed information about the problem you encountered and, if 
possible, add a description that helps to reproduce the problem. Thank you very 
much.
+
+
+
+Propose an improvement or a 
new feature
+
+Our community is constantly looking for feedback to improve Apache Flink. 
If you have an idea how to improve Flink or have a new feature in mind that 
would be beneficial for Flink users, please open an issue in http://issues.apache.org/jira/browse/FLINK;>Flink’s JIRA. The 
improvement or new feature should be described in appropriate detail and 
include the scope and its requirements if possible. Detailed information is 
important for a few reasons:
+
+
+  It ensures your requirements are met when the improvement or feature is 
implemented.
+  It helps to estimate the effort and to design a solution that addresses 
your needs.
+  It allow for constructive discussions that might arise around this 
issue.
+
+
+Detailed information is also required, if you plan to contribute the 
improvement or feature you proposed yourself. Please read the Contribute code guide in this case as well.
+
+We recommend to first reach consensus with the community on whether a new 
feature is required and how to implement a new feature, before starting with 
the implementation. Some features might be out of scope of the project, and 
it’s best to 

[11/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2016/05/24/stream-sql.html
--
diff --git a/content/news/2016/05/24/stream-sql.html 
b/content/news/2016/05/24/stream-sql.html
new file mode 100644
index 000..f22a063
--- /dev/null
+++ b/content/news/2016/05/24/stream-sql.html
@@ -0,0 +1,321 @@
+
+
+  
+
+
+
+
+Apache Flink: Stream Processing for Everyone with SQL and Apache 
Flink
+
+
+
+
+https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
+
+
+
+
+
+
+
+
+https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
+
+
+
+
+  
+
+
+
+
+
+
+
+  
+ 
+  
+
+
+
+  
+
+
+
+  
+  
+
+  
+
+  
+
+
+
+
+  
+
+
+Download Flink
+
+
+Home
+
+
+Introduction to Flink
+
+
+Flink Use Cases
+
+
+Powered by Flink
+
+
+Ecosystem
+
+
+Community  Project Info
+
+
+How to Contribute
+
+
+Flink 
Blog
+
+
+
+
+
+
+
+
+  Documentation
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
+  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
+
+  
+
+
+
+  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
+
+
+
+
+  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
+
+
+
+
+
+
+
+  
+
+
+
+  
+  
+
+
+Project FAQ
+
+
+https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
+
+
+Plan Visualizer 
+
+  
+
+
+
+  
+  
+  
+  
+
+  Stream Processing for Everyone with SQL and Apache Flink
+
+  
+24 May 2016 by Fabian Hueske (https://twitter.com/fhueske;>@fhueske)
+
+The capabilities of open source systems for distributed stream processing 
have evolved significantly over the last years. Initially, the first systems in 
the field (notably https://storm.apache.org;>Apache Storm) 
provided low latency processing, but were limited to at-least-once guarantees, 
processing-time semantics, and rather low-level APIs. Since then, several new 
systems emerged and pushed the state of the art of open source stream 
processing in several dimensions. Today, users of Apache Flink or https://beam.incubator.apache.org;>Apache Beam can use fluent Scala 
and Java APIs to implement stream processing jobs that operate in event-time 
with exactly-once semantics at high throughput and low latency.
+
+In the meantime, stream processing has taken off in the industry. We are 
witnessing a rapidly growing interest in stream processing which is reflected 
by prevalent deployments of streaming processing infrastructure such as https://kafka.apache.org;>Apache Kafka and Apache Flink. The 
increasing number of available data streams results in a demand for people that 
can analyze streaming data and turn it into real-time insights. However, stream 
data analysis requires a special skill set including knowledge of streaming 
concepts such as the characteristics of unbounded streams, windows, time, and 
state as well as the skills to implement stream analysis jobs usually against 
Java or Scala APIs. People with this skill set are rare and hard to find.
+
+About six months ago, the Apache Flink community started an effort to add a 
SQL interface for stream data analysis. SQL is the standard language 
to access and process data. Everybody who occasionally analyzes data is 
familiar with SQL. Consequently, a SQL interface for stream data processing 
will make this technology accessible to a much wider audience. Moreover, SQL 
support for streaming data will also enable new use cases such as interactive 
and ad-hoc stream analysis and significantly simplify many applications 
including stream ingestion and simple transformations. In this blog post, we 
report on the current status, architectural design, and future plans of the 
Apache Flink community to implement support for SQL as a language for analyzing 
data streams.
+
+Where did we come from?
+
+With the http://flink.apache.org/news/2015/04/13/release-0.9.0-milestone1.html;>0.9.0-milestone1
 release, Apache Flink added an API 

[07/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/css/graph.css
--
diff --git a/content/visualizer/css/graph.css b/content/visualizer/css/graph.css
new file mode 100755
index 000..f4365cc
--- /dev/null
+++ b/content/visualizer/css/graph.css
@@ -0,0 +1,48 @@
+/*!
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+g.type-TK > rect {
+  fill: #00ffd0;
+}
+
+svg {
+  border: 0px solid #999;
+  overflow: hidden;
+}
+
+text {
+  font-weight: 300;
+  font-family: "Helvetica Neue", Helvetica, Arial, sans-serf;
+  font-size: 14px;
+}
+
+.node rect {
+  stroke: #999;
+  stroke-width: 0px;
+  fill: #fff;
+}
+
+.edgeLabel rect {
+  fill: #fff;
+}
+
+.edgePath path {
+  stroke: #333;
+  stroke-width: 3px;
+  fill: none;
+}

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/css/nephelefrontend.css
--
diff --git a/content/visualizer/css/nephelefrontend.css 
b/content/visualizer/css/nephelefrontend.css
new file mode 100755
index 000..3ca1837
--- /dev/null
+++ b/content/visualizer/css/nephelefrontend.css
@@ -0,0 +1,198 @@
+/*!
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+*
+{
+   padding: 0px;
+   margin: 0px;
+}
+
+html, body
+{
+   height: 100%;
+   position: relative;
+   overflow: auto;
+}
+
+h1
+{
+   font-family: verdana;
+   font-size: 36px;
+   font-weight: normal;
+   font-decoration: none;
+   font-style: normal;
+   font-variant: small-caps;
+}
+
+h1 img {
+   vertical-align: middle;
+   margin-right: 50px;
+}
+
+h3
+{
+   font-family: verdana;
+   font-size: 20px;
+   font-weight: normal;
+   font-decoration: none;
+   font-style: italic;
+   font-variant: normal;
+   text-align: center;
+   color: #66;
+   margin: 10px 10px 20px 10px;
+}
+
+h4
+{
+   font-family: Verdana;
+   font-size: 16px;
+   font-weight: normal;
+   font-decoration: none;
+   font-style: italic;
+   font-variant: normal;
+   text-align: left;
+   color: #00;
+   margin: 5px 5px 10px 5px;
+   padding: 2px 2px 5px 2px;
+   border-bottom: 1px dashed #44;
+}
+
+input
+{
+   font-family: verdana, Sans-Serif;
+   font-size: 13px;
+
+   border: solid 1.5px #33;
+   background-color: #ee;
+   color: #33;
+
+   padding: 5px;
+   margin: 5px;
+}
+
+.fadedPropertiesText
+{
+   font-family: verdana;
+   font-size: 18px;
+   color: #AA;
+   font-style: italic;
+}
+
+
+.mainHeading
+{
+   border: 1px solid #262A37;
+   margin: 5px;
+   text-align: left;
+   background-color: white;
+   background-repeat: no-repeat;
+   background-position: right;
+   height: 100px;
+   overflow: hidden;
+}
+
+
+.boxed
+{
+   border: 1px solid #262A37;
+   margin: 5px;
+}
+
+.spaced
+{
+   margin: 5px;
+}
+
+.footer
+{
+   position: absolute;
+   left: 0px;
+   right: 0px;
+   bottom: 0px;
+}
+
+.error_text
+{
+   font-family: verdana;
+   font-size: 14px;
+   font-weight: normal;
+   font-decoration: none;
+   font-style: italic;
+   font-variant: normal;
+   text-align: center;
+   color: #DF0101;
+}
+
+.translucent
+{
+   -moz-opacity:0 ;
+   

[20/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/js/jquery.jcarousel.min.js
--
diff --git a/content/js/jquery.jcarousel.min.js 
b/content/js/jquery.jcarousel.min.js
new file mode 100644
index 000..7421030
--- /dev/null
+++ b/content/js/jquery.jcarousel.min.js
@@ -0,0 +1,4 @@
+/*! jCarousel - v0.3.4 - 2015-09-23
+* http://sorgalla.com/jcarousel/
+* Copyright (c) 2006-2015 Jan Sorgalla; Licensed MIT */
+!function(a){"use strict";var b=a.jCarousel={};b.version="0.3.4";var 
c=/^([+\-]=)?(.+)$/;b.parseTarget=function(a){var b=!1,d="object"!=typeof 
a?c.exec(a):null;return 
d?(a=parseInt(d[2],10)||0,d[1]&&(b=!0,"-="===d[1]&&(a*=-1))):"object"!=typeof 
a&&(a=parseInt(a,10)||0),{target:a,relative:b}},b.detectCarousel=function(a){for(var
 b;a.length>0;){if(b=a.filter("[data-jcarousel]"),b.length>0)return 
b;if(b=a.find("[data-jcarousel]"),b.length>0)return b;a=a.parent()}return 
null},b.base=function(c){return{version:b.version,_options:{},_element:null,_carousel:null,_init:a.noop,_create:a.noop,_destroy:a.noop,_reload:a.noop,create:function(){return
 
this._element.attr("data-"+c.toLowerCase(),!0).data(c,this),!1===this._trigger("create")?this:(this._create(),this._trigger("createend"),this)},destroy:function(){return!1===this._trigger("destroy")?this:(this._destroy(),this._trigger("destroyend"),this._element.removeData(c).removeAttr("data-"+c.toLowerCase()),this)},reload:function(a){return!1===t
 
his._trigger("reload")?this:(a&(a),this._reload(),this._trigger("reloadend"),this)},element:function(){return
 this._element},options:function(b,c){if(0===arguments.length)return 
a.extend({},this._options);if("string"==typeof b){if("undefined"==typeof 
c)return"undefined"==typeof 
this._options[b]?null:this._options[b];this._options[b]=c}else 
this._options=a.extend({},this._options,b);return 
this},carousel:function(){return 
this._carousel||(this._carousel=b.detectCarousel(this.options("carousel")||this._element),this._carousel||a.error('Could
 not detect carousel for plugin 
"'+c+'"')),this._carousel},_trigger:function(b,d,e){var f,g=!1;return 
e=[this].concat(e||[]),(d||this._element).each(function(){f=a.Event((c+":"+b).toLowerCase()),a(this).trigger(f,e),f.isDefaultPrevented()&&(g=!0)}),!g}}},b.plugin=function(c,d){var
 
e=a[c]=function(b,c){this._element=a(b),this.options(c),this._init(),this.create()};return
 e.fn=e.prototype=a.extend({},b.base(c),d),a.fn[c]=function(b){var 
 d=Array.prototype.slice.call(arguments,1),f=this;return 
this.each("string"==typeof b?function(){var e=a(this).data(c);if(!e)return 
a.error("Cannot call methods on "+c+' prior to initialization; attempted to 
call method "'+b+'"');if(!a.isFunction(e[b])||"_"===b.charAt(0))return 
a.error('No such method "'+b+'" for '+c+" instance");var 
g=e[b].apply(e,d);return g!==e&&"undefined"!=typeof g?(f=g,!1):void 
0}:function(){var d=a(this).data(c);d instanceof e?d.reload(b):new 
e(this,b)}),f},e}}(jQuery),function(a,b){"use strict";var c=function(a){return 
parseFloat(a)||0};a.jCarousel.plugin("jcarousel",{animating:!1,tail:0,inTail:!1,resizeTimer:null,lt:null,vertical:!1,rtl:!1,circular:!1,underflow:!1,relative:!1,_options:{list:function(){return
 this.element().children().eq(0)},items:function(){return 
this.list().children()},animation:400,transitions:!1,wrap:null,vertical:null,rtl:null,center:!1},_list:null,_items:null,_target:a(),_first:a(),_last:a(),_visible:a(),_fullyvisible:a(),_init:functio
 n(){var a=this;return 
this.onWindowResize=function(){a.resizeTimer&(a.resizeTimer),a.resizeTimer=setTimeout(function(){a.reload()},100)},this},_create:function(){this._reload(),a(b).on("resize.jcarousel",this.onWindowResize)},_destroy:function(){a(b).off("resize.jcarousel",this.onWindowResize)},_reload:function(){this.vertical=this.options("vertical"),null==this.vertical&&(this.vertical=this.list().height()>this.list().width()),this.rtl=this.options("rtl"),null==this.rtl&&(this.rtl=function(b){if("rtl"===(""+b.attr("dir")).toLowerCase())return!0;var
 c=!1;return 
b.parents("[dir]").each(function(){return/rtl/i.test(a(this).attr("dir"))?(c=!0,!1):void
 
0}),c}(this._element)),this.lt=this.vertical?"top":"left",this.relative="relative"===this.list().css("position"),this._list=null,this._items=null;var
 
b=this.index(this._target)>=0?this._target:this.closest();this.circular="circular"===this.options("wrap"),this.underflow=!1;var
 c={left:0,top:0};return b.length>0&&(this._prepar
 
e(b),this.list().find("[data-jcarousel-clone]").remove(),this._items=null,this.underflow=this._fullyvisible.length>=this.items().length,this.circular=this.circular&&!this.underflow,c[this.lt]=this._position(b)+"px"),this.move(c),this},list:function(){if(null===this._list){var
 
b=this.options("list");this._list=a.isFunction(b)?b.call(this):this._element.find(b)}return
 this._list},items:function(){if(null===this._items){var 

[34/39] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
an Gates
+PMC, Committer
+gates
+  
+  
+https://avatars0.githubusercontent.com/u/2388347?s=50; />
+Fabian Hueske
+PMC, Committer
+fhueske
+  
+
+https://avatars3.githubusercontent.com/u/498957?v=3s=50; />
+Vasia Kalavri
+PMC, Committer
+vasia
+  
+  
+
+https://avatars0.githubusercontent.com/u/68551?s=50; />
+Aljoscha Krettek
+PMC, Committer
+aljoscha
+  
+  
+https://avatars2.githubusercontent.com/u/2550549?s=50; />
+Andra Lungu
+Committer
+andra
+  
+  
+https://avatars0.githubusercontent.com/u/89049?s=50; />
+Robert Metzger
+PMC, Committer
+rmetzger
+  
+  
+https://avatars2.githubusercontent.com/u/837221?s=50; />
+Maximilian Michels
+PMC, Committer
+mxm
+  
+  
+https://avatars2.githubusercontent.com/u/1941681?s=50; />
+Chiwan Park
+Committer
+chiwanpark
+  
+  
+https://avatars1.githubusercontent.com/u/5756858?s=50; />
+Till Rohrmann
+PMC, Committer
+trohrmann
+  
+  
+https://avatars0.githubusercontent.com/u/105434?s=50; />
+Henry Saputra
+PMC, Committer
+hsaputra
+  
+  
+https://avatars0.githubusercontent.com/u/8959638?s=50; />
+Matthias J. Sax
+Committer
+mjsax
+  
+  
+https://avatars1.githubusercontent.com/u/409707?s=50; />
+Sebastian Schelter
+PMC, Committer
+ssc
+  
+  
+https://avatars2.githubusercontent.com/u/1925554?s=50; />
+Kostas Tzoumas
+PMC, Committer
+ktzoumas
+  
+  
+
+Timo Walther
+PMC, Committer
+twalthr
+  
+  
+https://avatars1.githubusercontent.com/u/1826769?s=50; />
+Daniel Warneke
+PMC, Committer
+warneke
+  
+  
+https://avatars1.githubusercontent.com/u/4425616?s=50; />
+ChengXiang Li
+Committer
+chengxiang
+  
+  
+https://avatars1.githubusercontent.com/u/569655?s=50; />
+Greg Hogan
+Committer
+greg
+  
+  
+https://avatars1.githubusercontent.com/u/5284370?s=50; />
+Tzu-Li (Gordon) Tai
+Committer
+tzulitai
+  
+
+
+You can reach committers directly at 
apache-id@apache.org. A list of all contributors can be 
found https://cwiki.apache.org/confluence/display/FLINK/List+of+contributors;>here.
+
+Former mentors
+
+The following people were very kind to mentor the project while in 
incubation.
+
+
+  
+
+Name
+Role
+Apache ID
+  
+  
+
+Ashutosh Chauhan
+Former PPMC, Mentor
+hashutosh
+  
+  
+
+Ted Dunning
+Former PPMC, Mentor
+tdunning
+  
+  
+
+Alan Gates
+Former PPMC, Mentor
+gates
+  
+  
+
+
+Owen O'Malley
+Former PPMC, Mentor
+omalley
+  
+  
+
+Sean Owen
+Former PPMC, Mentor
+srowen
+  
+  
+
+Henry Saputra
+Former PPMC, Mentor
+hsaputra
+  
+
+
+Slides
+
+Note: Keep in mind that code examples on slides have a 
chance of being incomplete or outdated. Always refer to the http://ci.apache.org/projects/flink/flink-docs-release-1.2;>latest 
documentation for an up to date reference.
+
+2016
+
+
+  Stefan Richter: A look at Apache Flink 1.2 and beyond 
Apache Flink Meetup Berlin, November 2016: http://www.slideshare.net/StefanRichter10/a-look-at-flink-12;>SlideShare
+  Robert Metzger: Apache Flink Community Updates November 
2016 Apache Flink Meetup Berlin, November 2016: http://www.slideshare.net/robertmetzger1/apache-flink-community-updates-november-2016-berlin-meetup;>SlideShare
+  Aljoscha Krettek: Apache Flink for IoT: How Event-Time 
Processing Enables Easy and Accurate Analytics Big Data Spain, 
Madrid November 2016: http://www.slideshare.net/dataArtisans/aljoscha-krettek-apache-flink-for-iot-how-eventtime-processing-enables-easy-and-accurate-analytics;>SlideShare
+  Stephan Ewen: Keynote -The maturing data streaming ecosystem and 
Apache Flink’s accelerated growth Apache Big Data Europe 2016, 
Seville November 2016: http://www.slideshare.net/dataArtisans/keynote-stephan-ewen-stream-processing-as-a-foundational-paradigm-and-apache-flinks-approach-to-it;>SlideShare
+  Kostas Tzoumas: Stream Processing with Apache Flink® 
Apache Flink London Meetup, November 2016: http://www.slideshare.net/dataArtisans/kostas-tzoumas-stream-processing-with-apache-flink;>SlideShare
+  Kostas Tzoumas: Apache Flink®: State of the Union and What’s 
Next Strata + Hadoop World New York, September 2016: http://www.slideshare.net/dataArtisans/kostas-tzoumas-apache-flink-state-of-the-union-and-whats-next;>SlideShare
+  Kostas Tzoumas  Stephan Ewen: Keynote -The maturing data 
streaming ecosystem and Apache Flink’s accelerated growth Flink 
Forward, Berlin September 2016: http://www.slideshare.net/FlinkForward/kostas-tzoumasstpehan-ewen-keynote-the-maturing-data-streaming-ecosystem-and-apache-flinks-accelerated-growth;>SlideShare
+  Robert Metzger: Connecting Apache Flink to the World - Reviewing 
the streaming connectors F

[35/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/blog/release_1.1.0-changelog.html
--
diff --git a/content/blog/release_1.1.0-changelog.html 
b/content/blog/release_1.1.0-changelog.html
deleted file mode 100644
index 15d6ddc..000
--- a/content/blog/release_1.1.0-changelog.html
+++ /dev/null
@@ -1,689 +0,0 @@
-
-
-  
-
-
-
-
-Apache Flink: Release 1.1.0 – Changelog
-
-
-
-
-https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
-
-
-
-
-
-
-
-
-https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
-
-
-
-
-  
-
-
-
-
-
-
-
-  
- 
-  
-
-
-
-  
-
-
-
-  
-  
-
-  
-
-  
-
-
-
-
-  
-
-
-Download Flink
-
-
-Home
-
-
-Introduction to Flink
-
-
-Flink Use Cases
-
-
-Powered by Flink
-
-
-Ecosystem
-
-
-Community  Project Info
-
-
-How to Contribute
-
-
-Flink 
Blog
-
-
-
-
-
-
-
-
-  Documentation
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
-
-  
-
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
-
-
-
-
-  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
-
-
-
-
-
-
-
-  
-
-
-
-  
-  
-
-
-Project FAQ
-
-
-https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
-
-
-Plan Visualizer 
-
-  
-
-
-
-  
-  
-  
-  
-Release 1.1.0 – Changelog
-
-   
-  Changelog
-  Sub-task
-  Bug
-  Improvement
-  New 
Feature
-  Task
-  Test
-  Wish
-
-  
-
-
-Changelog
-
-The 1.1.0 release https://issues.apache.org/jira/issues/?jql=project+%3D+FLINK+AND+fixVersion+%3D+1.1.0;>resolved
 457 JIRA issues in total.
-
-Sub-task
-
-
-  https://issues.apache.org/jira/browse/FLINK-1502;>FLINK-1502: Expose 
metrics to graphite, ganglia and JMX.
-  https://issues.apache.org/jira/browse/FLINK-1550;>FLINK-1550: Show 
JVM Metrics for JobManager
-  https://issues.apache.org/jira/browse/FLINK-3129;>FLINK-3129: Add 
tooling to ensure interface stability
-  https://issues.apache.org/jira/browse/FLINK-3141;>FLINK-3141: Design 
of NULL values handling in operation
-  https://issues.apache.org/jira/browse/FLINK-3223;>FLINK-3223: 
Translate Table API query into logical relational plans for Apache Calcite
-  https://issues.apache.org/jira/browse/FLINK-3225;>FLINK-3225: 
Optimize logical Table API plans in Calcite
-  https://issues.apache.org/jira/browse/FLINK-3226;>FLINK-3226: 
Translate optimized logical Table API plans into physical plans representing 
DataSet programs
-  https://issues.apache.org/jira/browse/FLINK-3229;>FLINK-3229: Kinesis 
streaming consumer with integration of Flink's checkpointing mechanics
-  https://issues.apache.org/jira/browse/FLINK-3230;>FLINK-3230: Kinesis 
streaming producer
-  https://issues.apache.org/jira/browse/FLINK-3231;>FLINK-3231: Handle 
Kinesis-side resharding in Kinesis streaming consumer
-  https://issues.apache.org/jira/browse/FLINK-3327;>FLINK-3327: Attach 
the ExecutionConfig to the JobGraph and make it accessible to the 
AbstractInvocable.
-  https://issues.apache.org/jira/browse/FLINK-3489;>FLINK-3489: 
Refactor Table API before merging into master
-  https://issues.apache.org/jira/browse/FLINK-3544;>FLINK-3544: 
ResourceManager runtime components
-  https://issues.apache.org/jira/browse/FLINK-3545;>FLINK-3545: 
ResourceManager: YARN integration
-  https://issues.apache.org/jira/browse/FLINK-3547;>FLINK-3547: Add 
support for streaming projection, selection, and union
-  https://issues.apache.org/jira/browse/FLINK-3550;>FLINK-3550: Rework 
stream join example
-  https://issues.apache.org/jira/browse/FLINK-3552;>FLINK-3552: Change 
socket WordCount to be properly windowed
-  https://issues.apache.org/jira/browse/FLINK-3573;>FLINK-3573: 
Implement more String functions for Table API
-  https://issues.apache.org/jira/browse/FLINK-3574;>FLINK-3574: 
Implement math functions for Table API
-  

[48/78] [abbrv] flink-web git commit: Rebuild site

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/q/quickstart-SNAPSHOT.sh
--
diff --git a/content/q/quickstart-SNAPSHOT.sh b/content/q/quickstart-SNAPSHOT.sh
new file mode 100755
index 000..1b1f2d2
--- /dev/null
+++ b/content/q/quickstart-SNAPSHOT.sh
@@ -0,0 +1,47 @@
+#!/usr/bin/env bash
+
+
+#  Licensed to the Apache Software Foundation (ASF) under one
+#  or more contributor license agreements.  See the NOTICE file
+#  distributed with this work for additional information
+#  regarding copyright ownership.  The ASF licenses this file
+#  to you under the Apache License, Version 2.0 (the
+#  "License"); you may not use this file except in compliance
+#  with the License.  You may obtain a copy of the License at
+#
+#  http://www.apache.org/licenses/LICENSE-2.0
+#
+#  Unless required by applicable law or agreed to in writing, software
+#  distributed under the License is distributed on an "AS IS" BASIS,
+#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+#  See the License for the specific language governing permissions and
+# limitations under the License.
+
+
+
+PACKAGE=quickstart
+
+mvn archetype:generate 
\
+  -DarchetypeGroupId=org.apache.flink  \
+  -DarchetypeArtifactId=flink-quickstart-java  \
+  -DarchetypeVersion=1.2-SNAPSHOT  \
+  -DgroupId=org.myorg.quickstart   \
+  -DartifactId=$PACKAGE
\
+  -Dversion=0.1
\
+  -Dpackage=org.myorg.quickstart   \
+  -DinteractiveMode=false  
\
+  
-DarchetypeCatalog=https://repository.apache.org/content/repositories/snapshots/
+
+#
+# Give some guidance
+#
+echo -e "\\n\\n"
+echo -e "\\tA sample quickstart Flink Job has been created."
+echo -e "\\tSwitch into the directory using"
+echo -e "\\t\\t cd $PACKAGE"
+echo -e "\\tImport the project there using your favorite IDE (Import it as a 
maven project)"
+echo -e "\\tBuild a jar inside the directory using"
+echo -e "\\t\\t mvn clean package"
+echo -e "\\tYou will find the runnable jar in $PACKAGE/target"
+echo -e "\\tConsult our website if you have any troubles: 
http://flink.apache.org/community.html#mailing-lists;
+echo -e "\\n\\n"

http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/q/quickstart-scala-SNAPSHOT.sh
--
diff --git a/content/q/quickstart-scala-SNAPSHOT.sh 
b/content/q/quickstart-scala-SNAPSHOT.sh
new file mode 100755
index 000..9cfeb2e
--- /dev/null
+++ b/content/q/quickstart-scala-SNAPSHOT.sh
@@ -0,0 +1,47 @@
+#!/usr/bin/env bash
+
+
+#  Licensed to the Apache Software Foundation (ASF) under one
+#  or more contributor license agreements.  See the NOTICE file
+#  distributed with this work for additional information
+#  regarding copyright ownership.  The ASF licenses this file
+#  to you under the Apache License, Version 2.0 (the
+#  "License"); you may not use this file except in compliance
+#  with the License.  You may obtain a copy of the License at
+#
+#  http://www.apache.org/licenses/LICENSE-2.0
+#
+#  Unless required by applicable law or agreed to in writing, software
+#  distributed under the License is distributed on an "AS IS" BASIS,
+#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+#  See the License for the specific language governing permissions and
+# limitations under the License.
+
+
+
+PACKAGE=quickstart
+
+mvn archetype:generate 
\
+  -DarchetypeGroupId=org.apache.flink  \
+  -DarchetypeArtifactId=flink-quickstart-scala \
+  -DarchetypeVersion=1.2-SNAPSHOT  \
+  -DgroupId=org.myorg.quickstart   \
+  -DartifactId=$PACKAGE
\
+  -Dversion=0.1
\
+  -Dpackage=org.myorg.quickstart   \
+  -DinteractiveMode=false  
\
+  
-DarchetypeCatalog=https://repository.apache.org/content/repositories/snapshots/
+
+#
+# Give some guidance
+#
+echo -e "\\n\\n"
+echo -e 

[15/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/news/2015/08/24/introducing-flink-gelly.html
--
diff --git a/content/news/2015/08/24/introducing-flink-gelly.html 
b/content/news/2015/08/24/introducing-flink-gelly.html
deleted file mode 100644
index 714e75a..000
--- a/content/news/2015/08/24/introducing-flink-gelly.html
+++ /dev/null
@@ -1,649 +0,0 @@
-
-
-  
-
-
-
-
-Apache Flink: Introducing Gelly: Graph Processing with Apache 
Flink
-
-
-
-
-https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;>
-
-
-
-
-
-
-
-
-https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js";>
-
-
-
-
-  
-
-
-
-
-
-
-
-  
- 
-  
-
-
-
-  
-
-
-
-  
-  
-
-  
-
-  
-
-
-
-
-  
-
-
-Download Flink
-
-
-Home
-
-
-Introduction to Flink
-
-
-Flink Use Cases
-
-
-Powered by Flink
-
-
-Ecosystem
-
-
-Community  Project Info
-
-
-How to Contribute
-
-
-Flink 
Blog
-
-
-
-
-
-
-
-
-  Documentation
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1; 
target="_blank">1.1 (Latest stable release) 
-  http://ci.apache.org/projects/flink/flink-docs-release-1.2; 
target="_blank">1.2 (Snapshot) 
-
-  
-
-
-
-  http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html;
 target="_blank">Quickstart 
-
-
-
-
-  https://github.com/apache/flink; target="_blank">Flink 
on GitHub 
-
-
-
-
-
-
-
-  
-
-
-
-  
-  
-
-
-Project FAQ
-
-
-https://twitter.com/apacheflink; 
target="_blank">@ApacheFlink 
-
-
-Plan Visualizer 
-
-  
-
-
-
-  
-  
-  
-  
-
-  Introducing Gelly: Graph Processing with Apache Flink
-
-  
-24 Aug 2015
-
-This blog post introduces Gelly, Apache Flink’s 
graph-processing API and library. Flink’s native support
-for iterations makes it a suitable platform for large-scale graph analytics.
-By leveraging delta iterations, Gelly is able to map various graph processing 
models such as
-vertex-centric or gather-sum-apply to Flink dataflows.
-
-Gelly allows Flink users to perform end-to-end data analysis in a single 
system.
-Gelly can be seamlessly used with Flink’s DataSet API,
-which means that pre-processing, graph creation, analysis, and post-processing 
can be done
-in the same application. At the end of this post, we will go through a 
step-by-step example
-in order to demonstrate that loading, transformation, filtering, graph 
creation, and analysis
-can be performed in a single Flink program.
-
-Overview
-
-
-  What is Gelly?
-  Graph Representation and 
Creation
-  Transformations and 
Utilities
-  Iterative Graph Processing
-  Library of Graph 
Algorithms
-  Use-Case: Music Profiles
-  Ongoing and Future Work
-
-
-
-
-What is Gelly?
-
-Gelly is a Graph API for Flink. It is currently supported in both Java and 
Scala.
-The Scala methods are implemented as wrappers on top of the basic Java 
operations.
-The API contains a set of utility functions for graph analysis, supports 
iterative graph
-processing and introduces a library of graph algorithms.
-
-
-
-
-
-Back to top
-
-Graph Representation and 
Creation
-
-In Gelly, a graph is represented by a DataSet of vertices and a DataSet of 
edges.
-A vertex is defined by its unique ID and a value, whereas an edge is defined 
by its source ID,
-target ID, and value. A vertex or edge for which a value is not specified will 
simply have the
-value type set to NullValue.
-
-A graph can be created from:
-
-
-  DataSet of edges and an optional DataSet of 
vertices using Graph.fromDataSet()
-  DataSet of Tuple3 and an optional DataSet of 
Tuple2 using Graph.fromTupleDataSet()
-  Collection of edges and an optional Collection 
of vertices using Graph.fromCollection()
-
-
-In all three cases, if the vertices are not provided,
-Gelly will automatically produce the vertex IDs from the edge source and 
target IDs.
-
-Back to top
-
-Transformations and Utilities
-
-These are methods of the Graph class and include common graph metrics, 
transformations
-and mutations as well as neighborhood aggregations.
-
-Common Graph Metrics
-These methods can be used 

[02/78] [abbrv] flink-web git commit: Revert "Rebuild website"

2017-01-18 Thread rmetzger
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/js/jquery.tools.min.js
--
diff --git a/content/visualizer/js/jquery.tools.min.js 
b/content/visualizer/js/jquery.tools.min.js
deleted file mode 100755
index dc914d8..000
--- a/content/visualizer/js/jquery.tools.min.js
+++ /dev/null
@@ -1,20 +0,0 @@
-/*!
- * jQuery Tools v1.2.7 - The missing UI library for the Web
- * 
- * overlay/overlay.js
- * scrollable/scrollable.js
- * tabs/tabs.js
- * tooltip/tooltip.js
- * 
- * NO COPYRIGHTS OR LICENSES. DO WHAT YOU LIKE.
- * 
- * http://flowplayer.org/tools/
- * 
- */
-/*! jQuery v1.7.2 jquery.com | jquery.org/license */
-(function(a,b){function cy(a){return 
f.isWindow(a)?a:a.nodeType===9?a.defaultView||a.parentWindow:!1}function 
cu(a){if(!cj[a]){var 
b=c.body,d=f("<"+a+">").appendTo(b),e=d.css("display");d.remove();if(e==="none"||e===""){ck||(ck=c.createElement("iframe"),ck.frameBorder=ck.width=ck.height=0),b.appendChild(ck);if(!cl||!ck.createElement)cl=(ck.contentWindow||ck.contentDocument).document,cl.write((f.support.boxModel?"":"")+""),cl.close();d=cl.createElement(a),cl.body.appendChild(d),e=f.css(d,"display"),b.removeChild(ck)}cj[a]=e}return
 cj[a]}function ct(a,b){var 
c={};f.each(cp.concat.apply([],cp.slice(0,b)),function(){c[this]=a});return 
c}function cs(){cq=b}function cr(){setTimeout(cs,0);return cq=f.now()}function 
ci(){try{return new a.ActiveXObject("Microsoft.XMLHTTP")}catch(b){}}function 
ch(){try{return new a.XMLHttpRequest}catch(b){}}function 
cb(a,c){a.dataFilter&&(c=a.dataFilter(c,a.dataType));var 
d=a.dataTypes,e={},g,h,i=d.length,j,k=d[0],l,m,n,o,p;for(g=1;
 g0){if(c!=="border")for(;e

<    1   2   3   4   5   6   7   8   9   10   >