flink git commit: [hotfix][kinesis] Fix build instructions
Repository: flink Updated Branches: refs/heads/release-1.1 e4f62d3b3 -> b054da46e [hotfix][kinesis] Fix build instructions Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/b054da46 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/b054da46 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/b054da46 Branch: refs/heads/release-1.1 Commit: b054da46ea00e958b1c055f6e4f8b5dd51d3a390 Parents: e4f62d3 Author: Robert MetzgerAuthored: Mon Aug 8 12:19:03 2016 +0200 Committer: Robert Metzger Committed: Mon Aug 8 12:23:51 2016 +0200 -- docs/apis/streaming/connectors/kinesis.md | 3 +++ 1 file changed, 3 insertions(+) -- http://git-wip-us.apache.org/repos/asf/flink/blob/b054da46/docs/apis/streaming/connectors/kinesis.md -- diff --git a/docs/apis/streaming/connectors/kinesis.md b/docs/apis/streaming/connectors/kinesis.md index e4ded62..63683e4 100644 --- a/docs/apis/streaming/connectors/kinesis.md +++ b/docs/apis/streaming/connectors/kinesis.md @@ -47,6 +47,9 @@ Flink releases because of the licensing issue. Therefore, you need to build the Download the Flink source or check it out from the git repository. Then, use the following Maven command to build the module: {% highlight bash %} mvn clean install -Pinclude-kinesis -DskipTests +# In Maven 3.3 the shading of flink-dist doesn't work properly in one run, so we need to run mvn for flink-dist again. +cd flink-dist +mvn clean install -Pinclude-kinesis -DskipTests {% endhighlight %}
flink git commit: [hotfix][kinesis] Fix build instructions
Repository: flink Updated Branches: refs/heads/master 27232fe32 -> 6a883776d [hotfix][kinesis] Fix build instructions Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/6a883776 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/6a883776 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/6a883776 Branch: refs/heads/master Commit: 6a883776da244a7dcc37ee5c24c5705cc4f55a4d Parents: 27232fe Author: Robert MetzgerAuthored: Mon Aug 8 12:19:03 2016 +0200 Committer: Robert Metzger Committed: Mon Aug 8 12:19:03 2016 +0200 -- docs/apis/streaming/connectors/kinesis.md | 3 +++ 1 file changed, 3 insertions(+) -- http://git-wip-us.apache.org/repos/asf/flink/blob/6a883776/docs/apis/streaming/connectors/kinesis.md -- diff --git a/docs/apis/streaming/connectors/kinesis.md b/docs/apis/streaming/connectors/kinesis.md index e4ded62..63683e4 100644 --- a/docs/apis/streaming/connectors/kinesis.md +++ b/docs/apis/streaming/connectors/kinesis.md @@ -47,6 +47,9 @@ Flink releases because of the licensing issue. Therefore, you need to build the Download the Flink source or check it out from the git repository. Then, use the following Maven command to build the module: {% highlight bash %} mvn clean install -Pinclude-kinesis -DskipTests +# In Maven 3.3 the shading of flink-dist doesn't work properly in one run, so we need to run mvn for flink-dist again. +cd flink-dist +mvn clean install -Pinclude-kinesis -DskipTests {% endhighlight %}
flink git commit: [FLINK-4080] Guarantee exactly-once for Kinesis consumer for failures in the middle of aggregated records
Repository: flink Updated Branches: refs/heads/master dbe41f487 -> fa42cdabf [FLINK-4080] Guarantee exactly-once for Kinesis consumer for failures in the middle of aggregated records This closes #2180 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/fa42cdab Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/fa42cdab Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/fa42cdab Branch: refs/heads/master Commit: fa42cdabfdd37fa3c2f86198ca04d37113ddda00 Parents: dbe41f4 Author: Gordon TaiAuthored: Wed Jun 29 15:46:35 2016 +0800 Committer: Robert Metzger Committed: Wed Jun 29 13:36:54 2016 +0200 -- .../kinesis/FlinkKinesisConsumer.java | 23 ++-- .../kinesis/internals/KinesisDataFetcher.java | 11 +- .../kinesis/internals/ShardConsumerThread.java | 110 +-- .../kinesis/model/SentinelSequenceNumber.java | 18 ++- .../kinesis/model/SequenceNumber.java | 104 ++ .../kinesis/FlinkKinesisConsumerTest.java | 15 +-- .../internals/KinesisDataFetcherTest.java | 3 +- .../internals/ShardConsumerThreadTest.java | 47 +--- 8 files changed, 254 insertions(+), 77 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/fa42cdab/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java -- diff --git a/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java b/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java index 9cbc9d9..19b36e4 100644 --- a/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java +++ b/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java @@ -28,6 +28,7 @@ import org.apache.flink.streaming.connectors.kinesis.config.KinesisConfigConstan import org.apache.flink.streaming.connectors.kinesis.internals.KinesisDataFetcher; import org.apache.flink.streaming.connectors.kinesis.model.KinesisStreamShard; import org.apache.flink.streaming.connectors.kinesis.model.SentinelSequenceNumber; +import org.apache.flink.streaming.connectors.kinesis.model.SequenceNumber; import org.apache.flink.streaming.connectors.kinesis.proxy.KinesisProxy; import org.apache.flink.streaming.connectors.kinesis.serialization.KinesisDeserializationSchema; import org.apache.flink.streaming.connectors.kinesis.serialization.KinesisDeserializationSchemaWrapper; @@ -64,7 +65,7 @@ import static org.apache.flink.util.Preconditions.checkNotNull; * @param the type of data emitted */ public class FlinkKinesisConsumer extends RichParallelSourceFunction - implements CheckpointedAsynchronously >, ResultTypeQueryable { + implements CheckpointedAsynchronously >, ResultTypeQueryable { private static final long serialVersionUID = 4724006128720664870L; @@ -92,10 +93,10 @@ public class FlinkKinesisConsumer extends RichParallelSourceFunction private transient KinesisDataFetcher fetcher; /** The sequence numbers of the last fetched data records from Kinesis by this task */ - private transient HashMap lastSequenceNums; + private transient HashMap lastSequenceNums; /** The sequence numbers to restore to upon restore from failure */ - private transient HashMap sequenceNumsToRestore; + private transient HashMap sequenceNumsToRestore; private volatile boolean hasAssignedShards; @@ -227,14 +228,14 @@ public class FlinkKinesisConsumer extends RichParallelSourceFunction LOG.info("Consumer task {} is restoring sequence numbers from previous checkpointed state", thisConsumerTaskIndex); } - for (Map.Entry restoreSequenceNum : sequenceNumsToRestore.entrySet()) { + for (Map.Entry restoreSequenceNum : sequenceNumsToRestore.entrySet()) { // advance the corresponding shard to the last known sequence number fetcher.advanceSequenceNumberTo(restoreSequenceNum.getKey(), restoreSequenceNum.getValue());
flink git commit: [FLINK-4033] Polish up Kinesis connector documentation
Repository: flink Updated Branches: refs/heads/master fa42cdabf -> 256c9c4da [FLINK-4033] Polish up Kinesis connector documentation Includes: 1. Scala examples for consumer and producer 2. Add information about AWS Kinesis service usage 3. Add Kinesis connecter to the fault tolerance guarantees table 4. Minor typo fix in Kafka documentation This closes #2181 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/256c9c4d Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/256c9c4d Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/256c9c4d Branch: refs/heads/master Commit: 256c9c4daf11349128eaa2e71a434f609e57053c Parents: fa42cda Author: Gordon TaiAuthored: Wed Jun 29 18:29:08 2016 +0800 Committer: Robert Metzger Committed: Wed Jun 29 13:41:35 2016 +0200 -- docs/apis/streaming/connectors/kafka.md | 2 +- docs/apis/streaming/connectors/kinesis.md | 63 ++ docs/apis/streaming/fault_tolerance.md| 16 +-- 3 files changed, 68 insertions(+), 13 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/256c9c4d/docs/apis/streaming/connectors/kafka.md -- diff --git a/docs/apis/streaming/connectors/kafka.md b/docs/apis/streaming/connectors/kafka.md index 9bd70be..e7cd05b 100644 --- a/docs/apis/streaming/connectors/kafka.md +++ b/docs/apis/streaming/connectors/kafka.md @@ -236,7 +236,7 @@ properties.setProperty("bootstrap.servers", "localhost:9092"); properties.setProperty("zookeeper.connect", "localhost:2181"); properties.setProperty("group.id", "test"); -val myConsumer = new FlinkKafkaConsumer08[Stirng]("topic", new SimpleStringSchema(), properties); +val myConsumer = new FlinkKafkaConsumer08[String]("topic", new SimpleStringSchema(), properties); myConsumer.assignTimestampsAndWatermarks(new CustomWatermarkEmitter()); stream = env .addSource(myConsumer) http://git-wip-us.apache.org/repos/asf/flink/blob/256c9c4d/docs/apis/streaming/connectors/kinesis.md -- diff --git a/docs/apis/streaming/connectors/kinesis.md b/docs/apis/streaming/connectors/kinesis.md index 66c078a..db3a9c4 100644 --- a/docs/apis/streaming/connectors/kinesis.md +++ b/docs/apis/streaming/connectors/kinesis.md @@ -52,12 +52,16 @@ mvn clean install -Pinclude-kinesis -DskipTests Note that the streaming connectors are not part of the binary distribution. -See linking with them for cluster execution [here]({{site.baseurl}}/apis/cluster_execution.html#linking-with-modules-not-contained-in-the-binary-distribution). +See how to link with them for cluster execution [here]({{site.baseurl}}/apis/cluster_execution.html#linking-with-modules-not-contained-in-the-binary-distribution). - Usage of Consumer +### Using the Amazon Kinesis Streams Service +Follow the instructions from the [Amazon Kinesis Streams Developer Guide](https://docs.aws.amazon.com/streams/latest/dev/learning-kinesis-module-one-create-stream.html) +to setup Kinesis streams. Make sure to create the appropriate IAM policy and user to read / write to the Kinesis streams. + +### Kinesis Consumer The `FlinkKinesisConsumer` can be used to pull data from multiple Kinesis streams within the same AWS region in parallel. -It participates in Flink's distributed snapshot checkpointing and provides exactly-once processing guarantees. Note +It participates in Flink's distributed snapshot checkpointing and provides exactly-once user-defined state update guarantees. Note that the current version can not handle resharding of Kinesis streams. When Kinesis streams are resharded, the consumer will fail and the Flink streaming job must be resubmitted. @@ -78,10 +82,28 @@ kinesisConsumerConfig.put(KinesisConfigConstants.CONFIG_STREAM_INIT_POSITION_TYP StreamExecutionEnvironment env = StreamExecutionEnvironment.getEnvironment(); -DataStream kinesisRecords = env.addSource(new FlinkKinesisConsumer<>( +DataStream kinesis = env.addSource(new FlinkKinesisConsumer<>( "kinesis_stream_name", new SimpleStringSchema(), kinesisConsumerConfig)) {% endhighlight %} + +{% highlight scala %} +val kinesisConsumerConfig = new Properties(); +kinesisConsumerConfig.put(KinesisConfigConstants.CONFIG_AWS_REGION, "us-east-1"); +kinesisConsumerConfig.put( +KinesisConfigConstants.CONFIG_AWS_CREDENTIALS_PROVIDER_BASIC_ACCESSKEYID, +"aws_access_key_id_here"); +kinesisConsumerConfig.put( +KinesisConfigConstants.CONFIG_AWS_CREDENTIALS_PROVIDER_BASIC_SECRETKEY, +"aws_secret_key_here"); +kinesisConsumerConfig.put(KinesisConfigConstants.CONFIG_STREAM_INIT_POSITION_TYPE, "LATEST"); + +val env = StreamExecutionEnvironment.getEnvironment +
flink git commit: [FLINK-4085][Kinesis] Set Flink-specific user agent
Repository: flink Updated Branches: refs/heads/master 256c9c4da -> 123be2276 [FLINK-4085][Kinesis] Set Flink-specific user agent This closes #2175 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/123be227 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/123be227 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/123be227 Branch: refs/heads/master Commit: 123be22765ebf32a26e3b74dbbd00a36999a5862 Parents: 256c9c4 Author: Robert MetzgerAuthored: Tue Jun 28 13:54:12 2016 +0200 Committer: Robert Metzger Committed: Wed Jun 29 13:43:04 2016 +0200 -- .../connectors/kinesis/proxy/KinesisProxy.java | 17 +++-- 1 file changed, 11 insertions(+), 6 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/123be227/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/proxy/KinesisProxy.java -- diff --git a/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/proxy/KinesisProxy.java b/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/proxy/KinesisProxy.java index b4ae3f5..d035c03 100644 --- a/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/proxy/KinesisProxy.java +++ b/flink-streaming-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/proxy/KinesisProxy.java @@ -17,6 +17,8 @@ package org.apache.flink.streaming.connectors.kinesis.proxy; +import com.amazonaws.ClientConfiguration; +import com.amazonaws.ClientConfigurationFactory; import com.amazonaws.regions.Region; import com.amazonaws.regions.Regions; import com.amazonaws.services.kinesis.AmazonKinesisClient; @@ -29,6 +31,7 @@ import com.amazonaws.services.kinesis.model.LimitExceededException; import com.amazonaws.services.kinesis.model.ResourceNotFoundException; import com.amazonaws.services.kinesis.model.StreamStatus; import com.amazonaws.services.kinesis.model.Shard; +import org.apache.flink.runtime.util.EnvironmentInformation; import org.apache.flink.streaming.connectors.kinesis.config.KinesisConfigConstants; import org.apache.flink.streaming.connectors.kinesis.model.KinesisStreamShard; import org.apache.flink.streaming.connectors.kinesis.util.AWSUtil; @@ -58,9 +61,6 @@ public class KinesisProxy { /** The actual Kinesis client from the AWS SDK that we will be using to make calls */ private final AmazonKinesisClient kinesisClient; - /** The AWS region that this proxy will be making calls to */ - private final String regionId; - /** Configuration properties of this Flink Kinesis Connector */ private final Properties configProps; @@ -72,9 +72,14 @@ public class KinesisProxy { public KinesisProxy(Properties configProps) { this.configProps = checkNotNull(configProps); - this.regionId = configProps.getProperty(KinesisConfigConstants.CONFIG_AWS_REGION); - AmazonKinesisClient client = new AmazonKinesisClient(AWSUtil.getCredentialsProvider(configProps).getCredentials()); - client.setRegion(Region.getRegion(Regions.fromName(this.regionId))); + /* The AWS region that this proxy will be making calls to */ + String regionId = configProps.getProperty(KinesisConfigConstants.CONFIG_AWS_REGION); + // set Flink as a user agent + ClientConfiguration config = new ClientConfigurationFactory().getConfig(); + config.setUserAgent("Apache Flink " + EnvironmentInformation.getVersion() + " (" + EnvironmentInformation.getRevisionInformation().commitId + ") Kinesis Connector"); + AmazonKinesisClient client = new AmazonKinesisClient(AWSUtil.getCredentialsProvider(configProps).getCredentials(), config); + + client.setRegion(Region.getRegion(Regions.fromName(regionId))); this.kinesisClient = client; }
flink git commit: [FLINK-4027] Flush FlinkKafkaProducer on checkpoints
Repository: flink Updated Branches: refs/heads/master ffaf10d22 -> 7206b0ed2 [FLINK-4027] Flush FlinkKafkaProducer on checkpoints This closes #2108 This closes #2058 because its an invalid pull request. Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/7206b0ed Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/7206b0ed Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/7206b0ed Branch: refs/heads/master Commit: 7206b0ed2adb10c94e1ffd3dbe851250b44edcf4 Parents: ffaf10d Author: Robert MetzgerAuthored: Wed Jun 15 17:50:38 2016 +0200 Committer: Robert Metzger Committed: Mon Jul 4 11:55:09 2016 +0200 -- .../connectors/kafka/FlinkKafkaProducer08.java | 18 ++ .../kafka/KafkaTestEnvironmentImpl.java | 4 +- .../connectors/kafka/FlinkKafkaProducer09.java | 7 + .../kafka/KafkaTestEnvironmentImpl.java | 4 +- .../kafka/FlinkKafkaProducerBase.java | 120 -- .../kafka/AtLeastOnceProducerTest.java | 218 +++ .../connectors/kafka/KafkaConsumerTestBase.java | 3 +- 7 files changed, 358 insertions(+), 16 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/7206b0ed/flink-streaming-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer08.java -- diff --git a/flink-streaming-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer08.java b/flink-streaming-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer08.java index 4975f9a..e509d2f 100644 --- a/flink-streaming-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer08.java +++ b/flink-streaming-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer08.java @@ -125,4 +125,22 @@ public class FlinkKafkaProducer08 extends FlinkKafkaProducerBase { super(topicId, serializationSchema, producerConfig, customPartitioner); } + @Override + protected void flush() { + // The Kafka 0.8 producer doesn't support flushing, we wait here + // until all pending records are confirmed + //noinspection SynchronizeOnNonFinalField + synchronized (pendingRecordsLock) { + while (pendingRecords > 0) { + try { + pendingRecordsLock.wait(); + } catch (InterruptedException e) { + // this can be interrupted when the Task has been cancelled. + // by throwing an exception, we ensure that this checkpoint doesn't get confirmed + throw new RuntimeException("Flushing got interrupted while checkpointing", e); + } + } + } + } + } http://git-wip-us.apache.org/repos/asf/flink/blob/7206b0ed/flink-streaming-connectors/flink-connector-kafka-0.8/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTestEnvironmentImpl.java -- diff --git a/flink-streaming-connectors/flink-connector-kafka-0.8/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTestEnvironmentImpl.java b/flink-streaming-connectors/flink-connector-kafka-0.8/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTestEnvironmentImpl.java index 77d41ac..75ca9ed 100644 --- a/flink-streaming-connectors/flink-connector-kafka-0.8/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTestEnvironmentImpl.java +++ b/flink-streaming-connectors/flink-connector-kafka-0.8/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTestEnvironmentImpl.java @@ -97,7 +97,9 @@ public class KafkaTestEnvironmentImpl extends KafkaTestEnvironment { @Override public FlinkKafkaProducerBase getProducer(String topic, KeyedSerializationSchema serSchema, Properties props, KafkaPartitioner partitioner) { - return new FlinkKafkaProducer08(topic, serSchema, props, partitioner); + FlinkKafkaProducer08 prod = new FlinkKafkaProducer08(topic, serSchema, props, partitioner); + prod.setFlushOnCheckpoint(true); + return prod; } @Override
flink git commit: [hotfix] Fix NPE in NetworkStackThroughputITCase
Repository: flink Updated Branches: refs/heads/master 7206b0ed2 -> 7ee0425a3 [hotfix] Fix NPE in NetworkStackThroughputITCase Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/7ee0425a Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/7ee0425a Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/7ee0425a Branch: refs/heads/master Commit: 7ee0425a3467f9cf8d1dc967c99b5bb6c29c1745 Parents: 7206b0e Author: Robert MetzgerAuthored: Mon Jul 4 12:32:10 2016 +0200 Committer: Robert Metzger Committed: Mon Jul 4 12:32:10 2016 +0200 -- .../apache/flink/test/runtime/NetworkStackThroughputITCase.java | 3 +++ 1 file changed, 3 insertions(+) -- http://git-wip-us.apache.org/repos/asf/flink/blob/7ee0425a/flink-tests/src/test/java/org/apache/flink/test/runtime/NetworkStackThroughputITCase.java -- diff --git a/flink-tests/src/test/java/org/apache/flink/test/runtime/NetworkStackThroughputITCase.java b/flink-tests/src/test/java/org/apache/flink/test/runtime/NetworkStackThroughputITCase.java index 5506f55..3d0e5ab 100644 --- a/flink-tests/src/test/java/org/apache/flink/test/runtime/NetworkStackThroughputITCase.java +++ b/flink-tests/src/test/java/org/apache/flink/test/runtime/NetworkStackThroughputITCase.java @@ -287,9 +287,12 @@ public class NetworkStackThroughputITCase { config.setInteger(NUM_SLOTS_PER_TM_CONFIG_KEY, (Integer) p[5]); TestBaseWrapper test = new TestBaseWrapper(config); + test.startCluster(); System.out.println(Arrays.toString(p)); test.testProgram(); + + test.stopCluster(); } }
flink git commit: [FLINK-4394] RMQSource: QueueName accessible for subclasses
Repository: flink Updated Branches: refs/heads/master 90fdae452 -> 5ccd90715 [FLINK-4394] RMQSource: QueueName accessible for subclasses The queueName is needed if the subclasses override `setupQueue`. This closes #2373 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/5ccd9071 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/5ccd9071 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/5ccd9071 Branch: refs/heads/master Commit: 5ccd9071580e196d150905b2d05eef71e399a24c Parents: 90fdae4 Author: DominikAuthored: Mon Aug 15 09:41:57 2016 +0200 Committer: Robert Metzger Committed: Mon Aug 15 16:19:49 2016 +0200 -- .../org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/5ccd9071/flink-streaming-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java -- diff --git a/flink-streaming-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java b/flink-streaming-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java index 33cf52c..ee9c3b9 100644 --- a/flink-streaming-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java +++ b/flink-streaming-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java @@ -74,7 +74,7 @@ public class RMQSource extends MultipleIdsMessageAcknowledgingSourceBase schema;
flink git commit: [hotfix] Fix Redis Sink to fail at opening if Redis is not initialized.
Repository: flink Updated Branches: refs/heads/master 47acdeadf -> e9a067229 [hotfix] Fix Redis Sink to fail at opening if Redis is not initialized. This closes #2245 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/e9a06722 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/e9a06722 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/e9a06722 Branch: refs/heads/master Commit: e9a067229d787c3e874dc40303ac8661045cc54f Parents: 47acdea Author: kl0uAuthored: Wed Jul 13 18:40:17 2016 +0200 Committer: Robert Metzger Committed: Thu Aug 18 16:41:53 2016 +0200 -- .../streaming/connectors/redis/RedisSink.java | 8 +- .../common/container/RedisClusterContainer.java | 12 ++- .../container/RedisCommandsContainer.java | 7 ++ .../redis/common/container/RedisContainer.java | 17 - .../connectors/redis/RedisSinkITCase.java | 2 - .../connectors/redis/RedisSinkTest.java | 79 6 files changed, 117 insertions(+), 8 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/e9a06722/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/RedisSink.java -- diff --git a/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/RedisSink.java b/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/RedisSink.java index 43518e8..f6b0fd7 100644 --- a/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/RedisSink.java +++ b/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/RedisSink.java @@ -166,7 +166,13 @@ public class RedisSink extends RichSinkFunction { */ @Override public void open(Configuration parameters) throws Exception { - this.redisCommandsContainer = RedisCommandsContainerBuilder.build(this.flinkJedisConfigBase); + try { + this.redisCommandsContainer = RedisCommandsContainerBuilder.build(this.flinkJedisConfigBase); + this.redisCommandsContainer.open(); + } catch (Exception e) { + LOG.error("Redis has not been properly initialized: ", e); + throw e; + } } /** http://git-wip-us.apache.org/repos/asf/flink/blob/e9a06722/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/common/container/RedisClusterContainer.java -- diff --git a/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/common/container/RedisClusterContainer.java b/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/common/container/RedisClusterContainer.java index 7551c9e..d6621d6 100644 --- a/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/common/container/RedisClusterContainer.java +++ b/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/common/container/RedisClusterContainer.java @@ -33,7 +33,7 @@ public class RedisClusterContainer implements RedisCommandsContainer, Closeable private static final Logger LOG = LoggerFactory.getLogger(RedisClusterContainer.class); - private JedisCluster jedisCluster; + private transient JedisCluster jedisCluster; /** * Initialize Redis command container for Redis cluster. @@ -47,6 +47,16 @@ public class RedisClusterContainer implements RedisCommandsContainer, Closeable } @Override + public void open() throws Exception { + + // echo() tries to open a connection and echos back the + // message passed as argument. Here we use it to monitor + // if we can communicate with the cluster. + + jedisCluster.echo("Test"); + } + + @Override public void hset(final String key, final String hashField, final String value) { try { jedisCluster.hset(key, hashField, value); http://git-wip-us.apache.org/repos/asf/flink/blob/e9a06722/flink-streaming-connectors/flink-connector-redis/src/main/java/org/apache/flink/streaming/connectors/redis/common/container/RedisCommandsContainer.java -- diff --git
svn commit: r18075 - /release/flink/flink-1.2.0/
Author: rmetzger Date: Thu Feb 2 11:02:16 2017 New Revision: 18075 Log: Add files for Flink 1.2.0 release Added: release/flink/flink-1.2.0/ release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz (with props) release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.asc release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.md5 release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.sha release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.11.tgz (with props) release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.11.tgz.asc release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.11.tgz.md5 release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.11.tgz.sha release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.10.tgz (with props) release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.10.tgz.asc release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.10.tgz.md5 release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.10.tgz.sha release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.11.tgz (with props) release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.11.tgz.asc release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.11.tgz.md5 release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop24-scala_2.11.tgz.sha release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.10.tgz (with props) release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.10.tgz.asc release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.10.tgz.md5 release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.10.tgz.sha release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.11.tgz (with props) release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.11.tgz.asc release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.11.tgz.md5 release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop26-scala_2.11.tgz.sha release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.10.tgz (with props) release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.10.tgz.asc release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.10.tgz.md5 release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.10.tgz.sha release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.11.tgz (with props) release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.11.tgz.asc release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.11.tgz.md5 release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop27-scala_2.11.tgz.sha release/flink/flink-1.2.0/flink-1.2.0-src.tgz (with props) release/flink/flink-1.2.0/flink-1.2.0-src.tgz.asc release/flink/flink-1.2.0/flink-1.2.0-src.tgz.md5 release/flink/flink-1.2.0/flink-1.2.0-src.tgz.sha Added: release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz == Binary file - no diff available. Propchange: release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz -- svn:mime-type = application/octet-stream Added: release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.asc == --- release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.asc (added) +++ release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.asc Thu Feb 2 11:02:16 2017 @@ -0,0 +1,17 @@ +-BEGIN PGP SIGNATURE- +Version: GnuPG v1 + +iQIcBAABAgAGBQJYjl9uAAoJEBg/aUTZg5FZ/zcP/2oADsw2nizGwCpgjBnozG6D +84oGwELWAEFVP74VvN8a94UmEt/G0SkQdIPKDFc9SlxjTwwd36q8OkSPEQpkjc2z +EHyVjIcglygng5bQA2li7WQXhowlW+ZLG78Knktay/E+Pql4056JNZ2lBgAZqTvr +H1ajHjBZe1jNR7NxioOqa4yibR56otBsIgK1X34goT1MuW8xe2E9UWnxNbDRSgle +nMCp+HB92DRq8vXF6JEseVs9LwzIXR6VRd/0H8WXc/QIL5/4bpZ3flKBKB8xV4Fo +MSelQYryfIQ0Vx1UMUqyjrmDXyOElD6DsYRXsKoOBet05FUuPGCqRn1TRSj4dOcu +hIVohbsQW203rVylGNYI6mWrjhY/F3ayTDkVcJ0PI8yUcX5SneM0Ug3y5Sp5ckRv +Z9hMrS52/26YqpPyvWXa8sTX0L+9csEy1+HWZVT/8+Ng7hmlxL2y2Vkc3vMOvcGP +SIoLL1+KrVo7B4lZ1NaS2cDu2jNN2kpe5O7DfWlo+atDxhENa0AcToZNXrmt2lw3 +znXAdzrLJdljln7qjSeLQpzmTF4j9YfGDSKn3T6vd9A1CUP+c3di8rnhKZ1WKlRq +AqW50MRjJy/YAasgkSBR8EOtJjghmqZzC5PMfVPjQA79u7BckVb4OzT3XiIOF9jf +NNLcdmxmPYUsqsww5ybO +=NyCB +-END PGP SIGNATURE- Added: release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.md5 == --- release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.md5 (added) +++ release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.md5 Thu Feb 2 11:02:16 2017 @@ -0,0 +1 @@ +792225bf4256b8ae2b2d9582a505c41c flink-1.2.0-bin-hadoop2-scala_2.10.tgz Added: release/flink/flink-1.2.0/flink-1.2.0-bin-hadoop2-scala_2.10.tgz.sha == --- release
svn commit: r18076 - in /release/flink: flink-0.10.2/ flink-1.0.3/ flink-1.1.1/ flink-1.1.2/ flink-1.1.3/
Author: rmetzger Date: Thu Feb 2 11:08:41 2017 New Revision: 18076 Log: Delete old release files (they are still in the archive) Removed: release/flink/flink-0.10.2/ release/flink/flink-1.0.3/ release/flink/flink-1.1.1/ release/flink/flink-1.1.2/ release/flink/flink-1.1.3/
[flink] Git Push Summary
Repository: flink Updated Branches: refs/heads/release-1.2.0-rc0 [deleted] f3c59ceda
[flink] Git Push Summary
Repository: flink Updated Branches: refs/heads/release-1.2.0-rc2 [deleted] 8b5b6a8b2
[flink] Git Push Summary
Repository: flink Updated Branches: refs/heads/release-1.2.0-rc1 [deleted] 1bca12088
[flink] Git Push Summary
Repository: flink Updated Tags: refs/tags/release-1.2.0 [created] 1c659cf4a
[2/2] flink git commit: [FLINK-5153] Support YARN application tags
[FLINK-5153] Support YARN application tags Adds a new config option `yarn.tags`, a comma-separated list of strings passed to YARN as application tags. Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/d9c116e5 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/d9c116e5 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/d9c116e5 Branch: refs/heads/master Commit: d9c116e542889b6ad00485529b97652dc2d59cad Parents: e24a866 Author: Patrick LucasAuthored: Fri Feb 3 19:17:55 2017 -0500 Committer: Robert Metzger Committed: Wed Feb 8 18:39:47 2017 +0100 -- docs/setup/config.md| 2 + .../flink/configuration/ConfigConstants.java| 5 + .../yarn/AbstractYarnClusterDescriptor.java | 114 ++- 3 files changed, 93 insertions(+), 28 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/d9c116e5/docs/setup/config.md -- diff --git a/docs/setup/config.md b/docs/setup/config.md index 2872cfa..2accdc2 100644 --- a/docs/setup/config.md +++ b/docs/setup/config.md @@ -431,6 +431,8 @@ use the `env.java.opts` setting, which is the `%jvmopts%` variable in the String For example when running Flink on YARN on an environment with a restrictive firewall, this option allows specifying a range of allowed ports. +- `yarn.tags` A comma-separated list of tags to apply to the Flink YARN application. + ### Mesos http://git-wip-us.apache.org/repos/asf/flink/blob/d9c116e5/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java -- diff --git a/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java b/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java index 14ba9dd..c608fde 100644 --- a/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java +++ b/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java @@ -442,6 +442,11 @@ public final class ConfigConstants { */ public static final String YARN_APPLICATION_MASTER_PORT = "yarn.application-master.port"; + /** +* A comma-separated list of strings to use as YARN application tags. +*/ + public static final String YARN_APPLICATION_TAGS = "yarn.tags"; + // Mesos Configuration http://git-wip-us.apache.org/repos/asf/flink/blob/d9c116e5/flink-yarn/src/main/java/org/apache/flink/yarn/AbstractYarnClusterDescriptor.java -- diff --git a/flink-yarn/src/main/java/org/apache/flink/yarn/AbstractYarnClusterDescriptor.java b/flink-yarn/src/main/java/org/apache/flink/yarn/AbstractYarnClusterDescriptor.java index b537e09..21599c1 100644 --- a/flink-yarn/src/main/java/org/apache/flink/yarn/AbstractYarnClusterDescriptor.java +++ b/flink-yarn/src/main/java/org/apache/flink/yarn/AbstractYarnClusterDescriptor.java @@ -855,7 +855,7 @@ public abstract class AbstractYarnClusterDescriptor implements ClusterDescriptor name = customName; } - appContext.setApplicationName(name); // application name + appContext.setApplicationName(name); appContext.setApplicationType("Apache Flink"); appContext.setAMContainerSpec(amContainer); appContext.setResource(capability); @@ -863,6 +863,8 @@ public abstract class AbstractYarnClusterDescriptor implements ClusterDescriptor appContext.setQueue(yarnQueue); } + setApplicationTags(appContext); + // add a hook to clean up in case deployment fails Thread deploymentFailureHook = new DeploymentFailureHook(yarnClient, yarnApplication); Runtime.getRuntime().addShutdownHook(deploymentFailureHook); @@ -1024,75 +1026,117 @@ public abstract class AbstractYarnClusterDescriptor implements ClusterDescriptor customName = name; } - private void activateHighAvailabilitySupport(ApplicationSubmissionContext appContext) throws InvocationTargetException, IllegalAccessException { + private void activateHighAvailabilitySupport(ApplicationSubmissionContext appContext) throws + InvocationTargetException, IllegalAccessException { + ApplicationSubmissionContextReflector reflector = ApplicationSubmissionContextReflector.getInstance(); reflector.setKeepContainersAcrossApplicationAttempts(appContext, true);
[1/2] flink git commit: [FLINK-5153] Add test for YARN application tags
Repository: flink Updated Branches: refs/heads/master e24a866bf -> 95765b6d8 [FLINK-5153] Add test for YARN application tags Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/95765b6d Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/95765b6d Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/95765b6d Branch: refs/heads/master Commit: 95765b6d8fdda9c601ad61dc39ce02043ecefa05 Parents: d9c116e Author: Patrick LucasAuthored: Tue Feb 7 12:47:21 2017 -0500 Committer: Robert Metzger Committed: Wed Feb 8 18:39:47 2017 +0100 -- .../YARNSessionCapacitySchedulerITCase.java | 43 1 file changed, 35 insertions(+), 8 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/95765b6d/flink-yarn-tests/src/test/java/org/apache/flink/yarn/YARNSessionCapacitySchedulerITCase.java -- diff --git a/flink-yarn-tests/src/test/java/org/apache/flink/yarn/YARNSessionCapacitySchedulerITCase.java b/flink-yarn-tests/src/test/java/org/apache/flink/yarn/YARNSessionCapacitySchedulerITCase.java index ec66eb2..2a3b6c6 100644 --- a/flink-yarn-tests/src/test/java/org/apache/flink/yarn/YARNSessionCapacitySchedulerITCase.java +++ b/flink-yarn-tests/src/test/java/org/apache/flink/yarn/YARNSessionCapacitySchedulerITCase.java @@ -21,6 +21,7 @@ import com.fasterxml.jackson.databind.JsonNode; import com.fasterxml.jackson.databind.ObjectMapper; import com.fasterxml.jackson.databind.node.ArrayNode; import com.google.common.base.Joiner; +import com.google.common.collect.Sets; import org.apache.commons.io.FileUtils; import org.apache.flink.configuration.ConfigConstants; import org.apache.flink.configuration.GlobalConfiguration; @@ -53,13 +54,9 @@ import org.slf4j.LoggerFactory; import java.io.File; import java.io.FilenameFilter; import java.io.IOException; -import java.util.EnumSet; -import java.util.LinkedList; -import java.util.List; -import java.util.Map; -import java.util.Collections; -import java.util.Comparator; -import java.util.Arrays; +import java.lang.reflect.InvocationTargetException; +import java.lang.reflect.Method; +import java.util.*; import java.util.concurrent.ConcurrentMap; import java.util.regex.Matcher; import java.util.regex.Pattern; @@ -413,11 +410,14 @@ public class YARNSessionCapacitySchedulerITCase extends YarnTestBase { throw new RuntimeException(e); } - Runner runner = startWithArgs(new String[]{"run", "-m", "yarn-cluster", "-yj", flinkUberjar.getAbsolutePath(), + Runner runner = startWithArgs(new String[]{ + "run", "-m", "yarn-cluster", + "-yj", flinkUberjar.getAbsolutePath(), "-yt", flinkLibFolder.getAbsolutePath(), "-yn", "1", "-yjm", "768", "-yD", "yarn.heap-cutoff-ratio=0.5", // test if the cutoff is passed correctly + "-yD", "yarn.tags=test-tag", "-ytm", "1024", "-ys", "2", // test requesting slots from YARN. "--yarndetached", job, @@ -516,6 +516,7 @@ public class YARNSessionCapacitySchedulerITCase extends YarnTestBase { LOG.info("Got report {}", rep); } while(rep.getYarnApplicationState() == YarnApplicationState.RUNNING); + verifyApplicationTags(rep); } catch(Throwable t) { LOG.warn("Error while detached yarn session was running", t); Assert.fail(t.getMessage()); @@ -543,6 +544,32 @@ public class YARNSessionCapacitySchedulerITCase extends YarnTestBase { } } + /** +* Ensures that the YARN application tags were set properly. +* +* Since YARN application tags were only added in Hadoop 2.4, but Flink still supports Hadoop 2.3, reflection is +* required to invoke the methods. If the method does not exist, this test passes. +*/ + private void verifyApplicationTags(final ApplicationReport report) throws InvocationTargetException, + IllegalAccessException { + + final Method applicationTagsMethod; + + Class clazz = ApplicationReport.class; + try { + // this method is only supported by Hadoop 2.4.0 onwards + applicationTagsMethod = clazz.getMethod("getApplicationTags"); + } catch (NoSuchMethodException e) { +
[2/4] flink-web git commit: Add changelog for 1.2.0 release
http://git-wip-us.apache.org/repos/asf/flink-web/blob/5b25f84e/content/blog/release_1.2.0-changelog.html -- diff --git a/content/blog/release_1.2.0-changelog.html b/content/blog/release_1.2.0-changelog.html new file mode 100644 index 000..01ab822 --- /dev/null +++ b/content/blog/release_1.2.0-changelog.html @@ -0,0 +1,1545 @@ + + + + + + + +Apache Flink: Release 1.2.0 â Changelog + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + +Release 1.2.0 â Changelog + + + Changelog + + +Changelog + +The 1.2.0 release https://issues.apache.org/jira/issues/?jql=project+%3D+FLINK+AND+fixVersion+%3D+1.2.0;>resolved 650 JIRA issues in total. + +Sub-task + + +[https://issues.apache.org/jira/browse/FLINK-3202;>FLINK-3202] - Make Timer/Trigger Service Scoped to Key/Namespace + +[https://issues.apache.org/jira/browse/FLINK-3580;>FLINK-3580] - Reintroduce Date/Time and implement scalar functions for it + +[https://issues.apache.org/jira/browse/FLINK-3660;>FLINK-3660] - Measure latency of elements and expose it as a metric + +[https://issues.apache.org/jira/browse/FLINK-3761;>FLINK-3761] - Refactor State Backends/Make Keyed State Key-Group Aware + +[https://issues.apache.org/jira/browse/FLINK-3950;>FLINK-3950] - Add Meter Metric Type + +[https://issues.apache.org/jira/browse/FLINK-4174;>FLINK-4174] - Enhance Window Evictor + +[https://issues.apache.org/jira/browse/FLINK-4221;>FLINK-4221] - Show metrics in WebFrontend + +[https://issues.apache.org/jira/browse/FLINK-4273;>FLINK-4273] - Refactor JobClientActor to watch already submitted jobs + +[https://issues.apache.org/jira/browse/FLINK-4312;>FLINK-4312] - Remove Serializabiliy of ExecutionGraph + +[https://issues.apache.org/jira/browse/FLINK-4317;>FLINK-4317] - Restructure documentation layout + +[https://issues.apache.org/jira/browse/FLINK-4320;>FLINK-4320] - Fix misleading ScheduleMode names + +[https://issues.apache.org/jira/browse/FLINK-4324;>FLINK-4324] - Enable Akka SSL + +[https://issues.apache.org/jira/browse/FLINK-4325;>FLINK-4325] - Implement Web UI HTTPS + +[https://issues.apache.org/jira/browse/FLINK-4380;>FLINK-4380] - Introduce KeyGroupAssigner and Max-Parallelism Parameter + +[https://issues.apache.org/jira/browse/FLINK-4381;>FLINK-4381] - Refactor State to Prepare For Key-Group State Backends + +[https://issues.apache.org/jira/browse/FLINK-4384;>FLINK-4384] - Add a scheduleRunAsync() feature to the RpcEndpoint + +[https://issues.apache.org/jira/browse/FLINK-4386;>FLINK-4386] - Add as way to assert that code runs in the RpcEndpoints Main Thread + +[https://issues.apache.org/jira/browse/FLINK-4389;>FLINK-4389] - Expose metrics to Webfrontend + +[https://issues.apache.org/jira/browse/FLINK-4392;>FLINK-4392] - Make RPC Service Thread Safe + +[https://issues.apache.org/jira/browse/FLINK-4403;>FLINK-4403] - RPC proxy classloading should use Flink
[1/4] flink-web git commit: Add changelog for 1.2.0 release
Repository: flink-web Updated Branches: refs/heads/asf-site 49439c679 -> 5b25f84e4 http://git-wip-us.apache.org/repos/asf/flink-web/blob/5b25f84e/content/q/quickstart-scala-SNAPSHOT.sh -- diff --git a/content/q/quickstart-scala-SNAPSHOT.sh b/content/q/quickstart-scala-SNAPSHOT.sh index 9cfeb2e..95a6ac2 100755 --- a/content/q/quickstart-scala-SNAPSHOT.sh +++ b/content/q/quickstart-scala-SNAPSHOT.sh @@ -24,7 +24,7 @@ PACKAGE=quickstart mvn archetype:generate \ -DarchetypeGroupId=org.apache.flink \ -DarchetypeArtifactId=flink-quickstart-scala \ - -DarchetypeVersion=1.2-SNAPSHOT \ + -DarchetypeVersion=1.3-SNAPSHOT \ -DgroupId=org.myorg.quickstart \ -DartifactId=$PACKAGE \ -Dversion=0.1 \ http://git-wip-us.apache.org/repos/asf/flink-web/blob/5b25f84e/content/q/quickstart-scala.sh -- diff --git a/content/q/quickstart-scala.sh b/content/q/quickstart-scala.sh index 7dcbe34..a561efe 100755 --- a/content/q/quickstart-scala.sh +++ b/content/q/quickstart-scala.sh @@ -24,7 +24,7 @@ PACKAGE=quickstart mvn archetype:generate \ -DarchetypeGroupId=org.apache.flink \ -DarchetypeArtifactId=flink-quickstart-scala \ - -DarchetypeVersion=1.1.4 \ + -DarchetypeVersion=1.2.0 \ -DgroupId=org.myorg.quickstart \ -DartifactId=$PACKAGE \ -Dversion=0.1 \
[3/4] flink-web git commit: Add changelog for 1.2.0 release
http://git-wip-us.apache.org/repos/asf/flink-web/blob/5b25f84e/blog/release_1.2.0-changelog.md -- diff --git a/blog/release_1.2.0-changelog.md b/blog/release_1.2.0-changelog.md new file mode 100755 index 000..b42e0d5 --- /dev/null +++ b/blog/release_1.2.0-changelog.md @@ -0,0 +1,1370 @@ +--- +title: "Release 1.2.0 â Changelog" +--- + +* toc +{:toc} + +## Changelog + +The 1.2.0 release [resolved 650 JIRA issues](https://issues.apache.org/jira/issues/?jql=project+%3D+FLINK+AND+fixVersion+%3D+1.2.0) in total. + + + + +Sub-task + + +[FLINK-3202] - Make Timer/Trigger Service Scoped to Key/Namespace + +[FLINK-3580] - Reintroduce Date/Time and implement scalar functions for it + +[FLINK-3660] - Measure latency of elements and expose it as a metric + +[FLINK-3761] - Refactor State Backends/Make Keyed State Key-Group Aware + +[FLINK-3950] - Add Meter Metric Type + +[FLINK-4174] - Enhance Window Evictor + +[FLINK-4221] - Show metrics in WebFrontend + +[FLINK-4273] - Refactor JobClientActor to watch already submitted jobs + +[FLINK-4312] - Remove Serializabiliy of ExecutionGraph + +[FLINK-4317] - Restructure documentation layout + +[FLINK-4320] - Fix misleading ScheduleMode names + +[FLINK-4324] - Enable Akka SSL + +[FLINK-4325] - Implement Web UI HTTPS + +[FLINK-4380] - Introduce KeyGroupAssigner and Max-Parallelism Parameter + +[FLINK-4381] - Refactor State to Prepare For Key-Group State Backends + +[FLINK-4384] - Add a scheduleRunAsync() feature to the RpcEndpoint + +[FLINK-4386] - Add as way to assert that code runs in the RpcEndpoints Main Thread + +[FLINK-4389] - Expose metrics to Webfrontend + +[FLINK-4392] - Make RPC Service Thread Safe + +[FLINK-4403] - RPC proxy classloading should use Flink class classloader + +[FLINK-4404] - Implement Data Transfer SSL + +[FLINK-4405] - Implement Blob Server SSL + +[FLINK-4434] - Add a testing RPC service + +[FLINK-4490] - Decouple Slot and Instance + +[FLINK-4507] - Deprecate savepoint backend config + +[FLINK-4509] - Specify savepoint directory per savepoint + +[FLINK-4510] - Always create CheckpointCoordinator + +[FLINK-4512] - Add option for persistent checkpoints + +[FLINK-4670] - Add watch mechanism on current RPC framework + +[FLINK-4685] - Gather operator checkpoint durations data sizes from the runtime + +[FLINK-4691] - Add group-windows for streaming tables + +[FLINK-4697] - Gather more detailed checkpoint stats in CheckpointStatsTracker + +[FLINK-4698] - Visualize additional checkpoint information + +[FLINK-4748] - Fix shutdown of automatic watermark context + +[FLINK-4749] - Remove redundant processing time timers and futures in the window operator + +[FLINK-4750] - Ensure that active processing time triggers complete before closing operators + +[FLINK-4768] - Migrate High Availability configuration options + +[FLINK-4826] - Add keytab based kerberos support for Mesos environment + +[FLINK-4877] - Refactorings around FLINK-3674 (User Function Timers) + +[FLINK-4884] - Eagerly Store MergingWindowSet in State in WindowOperator + +[FLINK-4899] - Implement DCOS package + +[FLINK-4900] - Implement Docker image support + +[FLINK-4901] - Build DCOS Docker image + +[FLINK-4918] - Add SSL support to Mesos artifact server + +[FLINK-4921] - Upgrade to Mesos 1.0.1 + +[FLINK-4922] - Write documentation + +[FLINK-4937] - Add incremental group window aggregation for streaming Table API + +[FLINK-4959] - Write Documentation for ProcessFunction + +[FLINK-4984] - Add Cancellation Barriers to BarrierTracker and BarrierBuffer + +[FLINK-4985] - Report Declined/Canceled Checkpoints to Checkpoint Coordinator + +[FLINK-4987] - Harden slot pool logic + +[FLINK-4993] - Dont Allow Trigger.onMerge() to return TriggerResult + +[FLINK-4994] - Dont Clear Trigger State and Merging Window Set When Purging + +[FLINK-5042] - Convert old savepoints to new savepoints + +[FLINK-5043] - Converting keyed state from Flink 1.1 backend implementations to their new counterparts in 1.2 + +[FLINK-5044] - Converting operator and function state from Flink 1.1 for all changed operators in 1.2 + +[FLINK-5091] - Formalize the AppMaster environment for docker compability + +[FLINK-5290] - Ensure backwards compatibility of the hashes used to generate JobVertexIds + +[FLINK-5293] - Make the Kafka consumer backwards compatible. + +[FLINK-5294] - Make the WindowOperator backwards
[4/4] flink-web git commit: Add changelog for 1.2.0 release
Add changelog for 1.2.0 release Project: http://git-wip-us.apache.org/repos/asf/flink-web/repo Commit: http://git-wip-us.apache.org/repos/asf/flink-web/commit/5b25f84e Tree: http://git-wip-us.apache.org/repos/asf/flink-web/tree/5b25f84e Diff: http://git-wip-us.apache.org/repos/asf/flink-web/diff/5b25f84e Branch: refs/heads/asf-site Commit: 5b25f84e46253d3f2976c8db415fa48155ac57e8 Parents: 49439c6 Author: Robert MetzgerAuthored: Thu Feb 2 13:33:30 2017 +0100 Committer: Robert Metzger Committed: Thu Feb 2 13:33:30 2017 +0100 -- blog/release_1.2.0-changelog.md | 1370 + content/blog/release_1.2.0-changelog.html | 1545 content/q/quickstart-scala-SNAPSHOT.sh|2 +- content/q/quickstart-scala.sh |2 +- 4 files changed, 2917 insertions(+), 2 deletions(-) --
[1/4] flink-web git commit: Put Flink 1.2.0 on website
Repository: flink-web Updated Branches: refs/heads/asf-site 5b25f84e4 -> 5c98e415d http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/content/usecases.html -- diff --git a/content/usecases.html b/content/usecases.html index 0b4407b..132b2f3 100644 --- a/content/usecases.html +++ b/content/usecases.html @@ -90,20 +90,20 @@ Documentation - http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) - http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.3; target="_blank">1.3 (Snapshot) - http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + http://ci.apache.org/projects/flink/flink-docs-release-1.2/quickstart/setup_quickstart.html; target="_blank">Quickstart http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/downloads.md -- diff --git a/downloads.md b/downloads.md index e1f8efb..e47107e 100644 --- a/downloads.md +++ b/downloads.md @@ -36,12 +36,6 @@ pick the Hadoop 1 version. -Hadoop® 1.2.1 -Download - - - - Hadoop® 2.3.0 Download Download @@ -83,7 +77,6 @@ pick the Hadoop 1 version. You can add the following dependencies to your `pom.xml` to include Apache Flink in your project. These dependencies include a local execution environment and thus support local testing. -- **Hadoop 1**: If you want to interact with Hadoop 1, use `{{ site.FLINK_VERSION_HADOOP_1_STABLE }}` as the version. - **Scala API**: To use the Scala API, replace the `flink-java` artifact id with `flink-scala_2.10` and `flink-streaming-java_2.10` with `flink-streaming-scala_2.10`. For Scala 2.11 dependencies, use the suffix `_2.11` instead of `_2.10`. ```xml http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/q/quickstart-SNAPSHOT.sh -- diff --git a/q/quickstart-SNAPSHOT.sh b/q/quickstart-SNAPSHOT.sh index 1b1f2d2..7670328 100755 --- a/q/quickstart-SNAPSHOT.sh +++ b/q/quickstart-SNAPSHOT.sh @@ -24,7 +24,7 @@ PACKAGE=quickstart mvn archetype:generate \ -DarchetypeGroupId=org.apache.flink \ -DarchetypeArtifactId=flink-quickstart-java \ - -DarchetypeVersion=1.2-SNAPSHOT \ + -DarchetypeVersion=1.3-SNAPSHOT \ -DgroupId=org.myorg.quickstart \ -DartifactId=$PACKAGE \ -Dversion=0.1 \ http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/q/quickstart.sh -- diff --git a/q/quickstart.sh b/q/quickstart.sh index 972400a..b1322e3 100755 --- a/q/quickstart.sh +++ b/q/quickstart.sh @@ -24,7 +24,7 @@ PACKAGE=quickstart mvn archetype:generate \ -DarchetypeGroupId=org.apache.flink \ -DarchetypeArtifactId=flink-quickstart-java \ - -DarchetypeVersion=1.1.3 \ + -DarchetypeVersion=1.2.0 \ -DgroupId=org.myorg.quickstart \ -DartifactId=$PACKAGE \ -Dversion=0.1 \ http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/q/sbt-quickstart.sh -- diff --git a/q/sbt-quickstart.sh b/q/sbt-quickstart.sh index 78abcb8..48df995 100755 --- a/q/sbt-quickstart.sh +++ b/q/sbt-quickstart.sh @@ -42,7 +42,7 @@ defaultProjectName="Flink Project" defaultOrganization="org.example" defaultVersion="0.1-SNAPSHOT" defaultScalaVersion="2.11.7" -defaultFlinkVersion="1.1.3" +defaultFlinkVersion="1.2.0" echo "This script creates a Flink project using Scala and SBT." http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/q/stable-docs.html
[4/4] flink-web git commit: Put Flink 1.2.0 on website
Put Flink 1.2.0 on website Project: http://git-wip-us.apache.org/repos/asf/flink-web/repo Commit: http://git-wip-us.apache.org/repos/asf/flink-web/commit/5c98e415 Tree: http://git-wip-us.apache.org/repos/asf/flink-web/tree/5c98e415 Diff: http://git-wip-us.apache.org/repos/asf/flink-web/diff/5c98e415 Branch: refs/heads/asf-site Commit: 5c98e415dd22e1bb90c00a3341e67e255b51b13f Parents: 5b25f84 Author: Robert MetzgerAuthored: Mon Feb 6 16:50:45 2017 +0100 Committer: Robert Metzger Committed: Mon Feb 6 16:50:45 2017 +0100 -- _config.yml | 46 +- _includes/navbar.html | 4 +- _posts/2017-02-06-release-1.2.0.md | 244 +++ content/blog/feed.xml | 276 +++- content/blog/index.html | 51 +- content/blog/page2/index.html | 53 +- content/blog/page3/index.html | 51 +- content/blog/page4/index.html | 56 +- content/blog/page5/index.html | 651 +++ .../release_1.0.0-changelog_known_issues.html | 8 +- content/blog/release_1.1.0-changelog.html | 8 +- content/blog/release_1.2.0-changelog.html | 8 +- content/community.html | 12 +- content/contribute-code.html| 16 +- content/contribute-documentation.html | 8 +- content/documentation.html | 8 +- content/downloads.html | 47 +- content/ecosystem.html | 8 +- content/faq.html| 28 +- content/features.html | 8 +- content/how-to-contribute.html | 8 +- content/improve-website.html| 8 +- content/index.html | 16 +- content/introduction.html | 8 +- content/material.html | 8 +- content/news/2014/08/26/release-0.6.html| 8 +- content/news/2014/09/26/release-0.6.1.html | 8 +- content/news/2014/10/03/upcoming_events.html| 8 +- content/news/2014/11/04/release-0.7.0.html | 8 +- .../news/2014/11/18/hadoop-compatibility.html | 8 +- content/news/2015/01/06/december-in-flink.html | 8 +- content/news/2015/01/21/release-0.8.html| 8 +- content/news/2015/02/04/january-in-flink.html | 8 +- content/news/2015/02/09/streaming-example.html | 8 +- .../news/2015/03/02/february-2015-in-flink.html | 8 +- .../peeking-into-Apache-Flinks-Engine-Room.html | 8 +- content/news/2015/04/07/march-in-flink.html | 8 +- .../2015/04/13/release-0.9.0-milestone1.html| 8 +- .../05/11/Juggling-with-Bits-and-Bytes.html | 8 +- .../news/2015/05/14/Community-update-April.html | 8 +- .../announcing-apache-flink-0.9.0-release.html | 8 +- .../2015/08/24/introducing-flink-gelly.html | 8 +- content/news/2015/09/01/release-0.9.1.html | 10 +- content/news/2015/09/03/flink-forward.html | 8 +- content/news/2015/09/16/off-heap-memory.html| 8 +- content/news/2015/11/16/release-0.10.0.html | 8 +- content/news/2015/11/27/release-0.10.1.html | 8 +- .../news/2015/12/04/Introducing-windows.html| 8 +- .../news/2015/12/11/storm-compatibility.html| 8 +- content/news/2015/12/18/a-year-in-review.html | 8 +- content/news/2016/02/11/release-0.10.2.html | 8 +- content/news/2016/03/08/release-1.0.0.html | 8 +- content/news/2016/04/06/cep-monitoring.html | 8 +- content/news/2016/04/06/release-1.0.1.html | 8 +- .../news/2016/04/14/flink-forward-announce.html | 8 +- content/news/2016/04/22/release-1.0.2.html | 8 +- content/news/2016/05/11/release-1.0.3.html | 8 +- content/news/2016/05/24/stream-sql.html | 8 +- content/news/2016/08/08/release-1.1.0.html | 8 +- content/news/2016/08/11/release-1.1.1.html | 8 +- .../news/2016/08/24/ff16-keynotes-panels.html | 8 +- content/news/2016/09/05/release-1.1.2.html | 8 +- content/news/2016/10/12/release-1.1.3.html | 8 +- .../news/2016/12/19/2016-year-in-review.html| 8 +- content/news/2016/12/21/release-1.1.4.html | 8 +- content/news/2017/02/06/release-1.2.0.html | 466 + content/poweredby.html | 8 +- content/privacy-policy.html | 8 +- content/project.html| 8 +- content/q/quickstart-SNAPSHOT.sh| 2 +- content/q/quickstart.sh | 2 +- content/q/sbt-quickstart.sh | 2 +- content/q/stable-docs.html | 2 +- content/slides.html | 10 +- content/usecases.html
[2/4] flink-web git commit: Put Flink 1.2.0 on website
http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/content/news/2015/09/03/flink-forward.html -- diff --git a/content/news/2015/09/03/flink-forward.html b/content/news/2015/09/03/flink-forward.html index 37e7c8c..83c8cf2 100644 --- a/content/news/2015/09/03/flink-forward.html +++ b/content/news/2015/09/03/flink-forward.html @@ -90,20 +90,20 @@ Documentation - http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) - http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.3; target="_blank">1.3 (Snapshot) - http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + http://ci.apache.org/projects/flink/flink-docs-release-1.2/quickstart/setup_quickstart.html; target="_blank">Quickstart http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/content/news/2015/09/16/off-heap-memory.html -- diff --git a/content/news/2015/09/16/off-heap-memory.html b/content/news/2015/09/16/off-heap-memory.html index 60d7bda..7d72a7d 100644 --- a/content/news/2015/09/16/off-heap-memory.html +++ b/content/news/2015/09/16/off-heap-memory.html @@ -90,20 +90,20 @@ Documentation - http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) - http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.3; target="_blank">1.3 (Snapshot) - http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + http://ci.apache.org/projects/flink/flink-docs-release-1.2/quickstart/setup_quickstart.html; target="_blank">Quickstart http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/content/news/2015/11/16/release-0.10.0.html -- diff --git a/content/news/2015/11/16/release-0.10.0.html b/content/news/2015/11/16/release-0.10.0.html index 93bad1f..cbd11a1 100644 --- a/content/news/2015/11/16/release-0.10.0.html +++ b/content/news/2015/11/16/release-0.10.0.html @@ -90,20 +90,20 @@ Documentation - http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) - http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.3; target="_blank">1.3 (Snapshot) - http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + http://ci.apache.org/projects/flink/flink-docs-release-1.2/quickstart/setup_quickstart.html; target="_blank">Quickstart http://git-wip-us.apache.org/repos/asf/flink-web/blob/5c98e415/content/news/2015/11/27/release-0.10.1.html -- diff --git a/content/news/2015/11/27/release-0.10.1.html b/content/news/2015/11/27/release-0.10.1.html index 01e6b7f..d4ef924 100644 --- a/content/news/2015/11/27/release-0.10.1.html +++ b/content/news/2015/11/27/release-0.10.1.html @@ -90,20 +90,20 @@ Documentation - http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) - http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) +
flink git commit: [FLINK-5450] Fix restore from legacy log message
Repository: flink Updated Branches: refs/heads/release-1.2 085288e7e -> f4cd7d0ec [FLINK-5450] Fix restore from legacy log message This closes #3134 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/f4cd7d0e Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/f4cd7d0e Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/f4cd7d0e Branch: refs/heads/release-1.2 Commit: f4cd7d0ec3dd70e51e1db27c8336481838d50e63 Parents: 085288e Author: kl0uAuthored: Mon Jan 16 15:27:38 2017 +0100 Committer: Robert Metzger Committed: Mon Jan 23 14:32:30 2017 +0100 -- .../operators/windowing/WindowOperator.java | 27 +--- 1 file changed, 18 insertions(+), 9 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/f4cd7d0e/flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/operators/windowing/WindowOperator.java -- diff --git a/flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/operators/windowing/WindowOperator.java b/flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/operators/windowing/WindowOperator.java index 628d663..5ed5a4e 100644 --- a/flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/operators/windowing/WindowOperator.java +++ b/flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/operators/windowing/WindowOperator.java @@ -804,9 +804,6 @@ public class WindowOperator public void registerRestoredLegacyStateState() throws Exception { - LOG.info("{} (taskIdx={}) re-registering state from an older Flink version.", - getClass().getSimpleName(), legacyWindowOperatorType, getRuntimeContext().getIndexOfThisSubtask()); - switch (legacyWindowOperatorType) { case NONE: reregisterStateFromLegacyWindowOperator(); @@ -983,14 +980,22 @@ public class WindowOperator // if we restore from an older version, // we have to re-register the recovered state. - if (restoredFromLegacyEventTimeTimers != null) { + if (restoredFromLegacyEventTimeTimers != null && !restoredFromLegacyEventTimeTimers.isEmpty()) { + + LOG.info("{} (taskIdx={}) re-registering event-time timers from an older Flink version.", + getClass().getSimpleName(), getRuntimeContext().getIndexOfThisSubtask()); + for (Timer timer : restoredFromLegacyEventTimeTimers) { setCurrentKey(timer.key); internalTimerService.registerEventTimeTimer(timer.window, timer.timestamp); } } - if (restoredFromLegacyProcessingTimeTimers != null) { + if (restoredFromLegacyProcessingTimeTimers != null && !restoredFromLegacyProcessingTimeTimers.isEmpty()) { + + LOG.info("{} (taskIdx={}) re-registering processing-time timers from an older Flink version.", + getClass().getSimpleName(), getRuntimeContext().getIndexOfThisSubtask()); + for (Timer timer : restoredFromLegacyProcessingTimeTimers) { setCurrentKey(timer.key); internalTimerService.registerProcessingTimeTimer(timer.window, timer.timestamp); @@ -1003,16 +1008,20 @@ public class WindowOperator } public void reregisterStateFromLegacyAlignedWindowOperator() throws Exception { - if (restoredFromLegacyAlignedOpRecords != null) { + if (restoredFromLegacyAlignedOpRecords != null && !restoredFromLegacyAlignedOpRecords.isEmpty()) { + + LOG.info("{} (taskIdx={}) re-registering timers from legacy {} from an older Flink version.", + getClass().getSimpleName(), getRuntimeContext().getIndexOfThisSubtask(), legacyWindowOperatorType); + while (!restoredFromLegacyAlignedOpRecords.isEmpty()) { StreamRecord record = restoredFromLegacyAlignedOpRecords.poll(); setCurrentKey(keySelector.getKey(record.getValue())); processElement(record); } - - // gc friendliness - restoredFromLegacyAlignedOpRecords = null; } + + // gc friendliness +
[1/2] flink git commit: [FLINK-5613][query] querying a non-existing key is inconsistent among state backends
Repository: flink Updated Branches: refs/heads/release-1.2 b323f66a6 -> 44f7098d7 [FLINK-5613][query] querying a non-existing key is inconsistent among state backends Querying for a non-existing key for a state that has a default value set currently results in an UnknownKeyOrNamespace exception when the MemoryStateBackend or FsStateBackend is used but results in the default value if RocksDBStateBackend is set. This removes the special handling from the RocksDBStateBackend and makes it consistent with the other two back-ends, i.e. returning null which results in the mentioned UnknownKeyOrNamespace exception. Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/44f7098d Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/44f7098d Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/44f7098d Branch: refs/heads/release-1.2 Commit: 44f7098d74476ab561f38c8ced5f3a0b5f9bd0c2 Parents: 2e175fb Author: Nico KruberAuthored: Tue Jan 17 14:26:16 2017 +0100 Committer: Robert Metzger Committed: Tue Jan 24 08:44:08 2017 +0100 -- .../streaming/state/RocksDBValueState.java | 12 -- .../flink/test/query/QueryableStateITCase.java | 160 ++- 2 files changed, 153 insertions(+), 19 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/44f7098d/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java -- diff --git a/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java b/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java index 9563ed8..d8d89ba 100644 --- a/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java +++ b/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java @@ -23,7 +23,6 @@ import org.apache.flink.api.common.state.ValueStateDescriptor; import org.apache.flink.api.common.typeutils.TypeSerializer; import org.apache.flink.core.memory.DataInputViewStreamWrapper; import org.apache.flink.core.memory.DataOutputViewStreamWrapper; -import org.apache.flink.runtime.query.netty.message.KvStateRequestSerializer; import org.rocksdb.ColumnFamilyHandle; import org.rocksdb.RocksDBException; import org.rocksdb.WriteOptions; @@ -102,15 +101,4 @@ public class RocksDBValueState throw new RuntimeException("Error while adding data to RocksDB", e); } } - - @Override - public byte[] getSerializedValue(byte[] serializedKeyAndNamespace) throws Exception { - byte[] value = super.getSerializedValue(serializedKeyAndNamespace); - - if (value != null) { - return value; - } else { - return KvStateRequestSerializer.serializeValue(stateDesc.getDefaultValue(), stateDesc.getSerializer()); - } - } } http://git-wip-us.apache.org/repos/asf/flink/blob/44f7098d/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java -- diff --git a/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java b/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java index 327a715..c2df6ae 100644 --- a/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java +++ b/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java @@ -39,6 +39,7 @@ import org.apache.flink.api.java.tuple.Tuple2; import org.apache.flink.configuration.ConfigConstants; import org.apache.flink.configuration.Configuration; import org.apache.flink.configuration.QueryableStateOptions; +import org.apache.flink.contrib.streaming.state.RocksDBStateBackend; import org.apache.flink.runtime.akka.AkkaUtils; import org.apache.flink.runtime.executiongraph.ExecutionAttemptID; import org.apache.flink.runtime.executiongraph.ExecutionGraph; @@ -50,10 +51,14 @@ import org.apache.flink.runtime.messages.JobManagerMessages.JobFound; import org.apache.flink.runtime.query.KvStateLocation; import org.apache.flink.runtime.query.KvStateMessage; import org.apache.flink.runtime.query.QueryableStateClient; +import org.apache.flink.runtime.query.netty.UnknownKeyOrNamespace; import org.apache.flink.runtime.query.netty.message.KvStateRequestSerializer; +import org.apache.flink.runtime.state.AbstractStateBackend; import org.apache.flink.runtime.state.CheckpointListener;
[2/2] flink git commit: [FLINK-5395] [Build System] support locally build distribution by script create_release_files.sh
[FLINK-5395] [Build System] support locally build distribution by script create_release_files.sh This closes #3049 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/2e175fb3 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/2e175fb3 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/2e175fb3 Branch: refs/heads/release-1.2 Commit: 2e175fb301cc84ca0c3195d1fe325319b34fd57a Parents: b323f66 Author: shijinkuiAuthored: Wed Jan 11 17:42:16 2017 +0800 Committer: Robert Metzger Committed: Tue Jan 24 08:44:08 2017 +0100 -- .gitignore| 2 + tools/create_release_files.sh | 156 + 2 files changed, 124 insertions(+), 34 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/2e175fb3/.gitignore -- diff --git a/.gitignore b/.gitignore index 9012d0a..3c9e4e8 100644 --- a/.gitignore +++ b/.gitignore @@ -32,3 +32,5 @@ out/ /docs/.jekyll-metadata *.ipr *.iws +tools/flink +tools/flink-* http://git-wip-us.apache.org/repos/asf/flink/blob/2e175fb3/tools/create_release_files.sh -- diff --git a/tools/create_release_files.sh b/tools/create_release_files.sh index fdf50a5..ed01d97 100755 --- a/tools/create_release_files.sh +++ b/tools/create_release_files.sh @@ -66,16 +66,20 @@ fi GPG_PASSPHRASE=${GPG_PASSPHRASE:-XXX} GPG_KEY=${GPG_KEY:-XXX} GIT_AUTHOR=${GIT_AUTHOR:-"Your name "} -OLD_VERSION=${OLD_VERSION:-1.1-SNAPSHOT} -RELEASE_VERSION=${NEW_VERSION} -RELEASE_CANDIDATE=${RELEASE_CANDIDATE:-rc1} +OLD_VERSION=${OLD_VERSION:-1.2-SNAPSHOT} +RELEASE_VERSION=${NEW_VERSION:-1.3-SNAPSHOT} +RELEASE_CANDIDATE=${RELEASE_CANDIDATE:-none} RELEASE_BRANCH=${RELEASE_BRANCH:-master} USER_NAME=${USER_NAME:-yourapacheidhere} MVN=${MVN:-mvn} GPG=${GPG:-gpg} sonatype_user=${sonatype_user:-yourapacheidhere} sonatype_pw=${sonatype_pw:-XXX} - +# whether only build the dist local and don't release to apache +IS_LOCAL_DIST=${IS_LOCAL_DIST:-false} +GIT_REPO=${GIT_REPO:-git-wip-us.apache.org/repos/asf/flink.git} +SCALA_VERSION=none +HADOOP_VERSION=none if [ "$(uname)" == "Darwin" ]; then SHASUM="shasum -a 512" @@ -85,18 +89,81 @@ else MD5SUM="md5sum" fi +usage() { + set +x + echo "./create_release_files.sh --scala-version 2.11 --hadoop-version 2.7.2" + echo "" + echo "usage:" + echo "[--scala-version ] [--hadoop-version ]" + echo "" + echo "example 1: build apache release" + echo " sonatype_user=APACHEID sonatype_pw=APACHEIDPASSWORD \ " + echo " NEW_VERSION=1.2.0 RELEASE_CANDIDATE="rc1" RELEASE_BRANCH=release-1.2.0 OLD_VERSION=1.1-SNAPSHOT \ " + echo " USER_NAME=APACHEID GPG_PASSPHRASE=XXX GPG_KEY=KEYID \ " + echo " GIT_AUTHOR=\"`git config --get user.name` <`git config --get user.email`>\" \ " + echo " GIT_REPO=github.com/apache/flink.git \ " + echo " ./create_release_files.sh --scala-version 2.11 --hadoop-version 2.7.2" + echo "" + echo "example 2: build local release" + echo " NEW_VERSION=1.2.0 RELEASE_BRANCH=master OLD_VERSION=1.2-SNAPSHOT \ " + echo " GPG_PASSPHRASE=XXX GPG_KEY=XXX IS_LOCAL_DIST=true \ " + echo " ./create_release_files.sh --scala-version 2.11 --hadoop-version 2.7.2" + + exit 1 +} + +# Parse arguments +while (( "$#" )); do + case $1 in +--scala-version) + SCALA_VERSION="$2" + shift + ;; +--hadoop-version) + HADOOP_VERSION="$2" + shift + ;; +--help) + usage + ;; +*) + break + ;; + esac + shift +done + +### prepare() { # prepare - git clone http://git-wip-us.apache.org/repos/asf/flink.git flink + target_branch=release-$RELEASE_VERSION + if [ "$RELEASE_CANDIDATE" != "none" ]; then +target_branch=$target_branch-$RELEASE_CANDIDATE + fi + + if [ ! -d ./flink ]; then +git clone http://$GIT_REPO flink + else +# if flink git repo exist, delete target branch, delete builded distribution +rm -rf flink-*.tgz +cd flink +# try-catch +{ + git pull --all + git checkout master + git branch -D $target_branch -f +} || { + echo "branch $target_branch not found" +} +cd .. + fi + cd flink - git checkout -b "release-$RELEASE_VERSION-$RELEASE_CANDIDATE" origin/$RELEASE_BRANCH - rm -f .gitignore - rm -f .gitattributes - rm -f .travis.yml - rm -f deploysettings.xml - rm -f CHANGELOG - rm -rf .github + + git checkout -b $target_branch origin/$RELEASE_BRANCH + rm -rf .gitignore .gitattributes .travis.yml deploysettings.xml CHANGELOG .github + cd .. } @@ -117,10 +184,13 @@ make_source_release() { perl -pi -e "s#^version_short:
flink git commit: [FLINK-5395] [Build System] support locally build distribution by script create_release_files.sh
Repository: flink Updated Branches: refs/heads/master 00d1ad86a -> 6b55e2ca3 [FLINK-5395] [Build System] support locally build distribution by script create_release_files.sh This closes #3049 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/6b55e2ca Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/6b55e2ca Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/6b55e2ca Branch: refs/heads/master Commit: 6b55e2ca30c5ecbf9998e44b2638a7f2d94a5963 Parents: 00d1ad8 Author: shijinkuiAuthored: Wed Jan 11 17:42:16 2017 +0800 Committer: Robert Metzger Committed: Tue Jan 24 09:11:56 2017 +0100 -- .gitignore| 2 + tools/create_release_files.sh | 156 + 2 files changed, 124 insertions(+), 34 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/6b55e2ca/.gitignore -- diff --git a/.gitignore b/.gitignore index 9012d0a..3c9e4e8 100644 --- a/.gitignore +++ b/.gitignore @@ -32,3 +32,5 @@ out/ /docs/.jekyll-metadata *.ipr *.iws +tools/flink +tools/flink-* http://git-wip-us.apache.org/repos/asf/flink/blob/6b55e2ca/tools/create_release_files.sh -- diff --git a/tools/create_release_files.sh b/tools/create_release_files.sh index fdf50a5..ed01d97 100755 --- a/tools/create_release_files.sh +++ b/tools/create_release_files.sh @@ -66,16 +66,20 @@ fi GPG_PASSPHRASE=${GPG_PASSPHRASE:-XXX} GPG_KEY=${GPG_KEY:-XXX} GIT_AUTHOR=${GIT_AUTHOR:-"Your name "} -OLD_VERSION=${OLD_VERSION:-1.1-SNAPSHOT} -RELEASE_VERSION=${NEW_VERSION} -RELEASE_CANDIDATE=${RELEASE_CANDIDATE:-rc1} +OLD_VERSION=${OLD_VERSION:-1.2-SNAPSHOT} +RELEASE_VERSION=${NEW_VERSION:-1.3-SNAPSHOT} +RELEASE_CANDIDATE=${RELEASE_CANDIDATE:-none} RELEASE_BRANCH=${RELEASE_BRANCH:-master} USER_NAME=${USER_NAME:-yourapacheidhere} MVN=${MVN:-mvn} GPG=${GPG:-gpg} sonatype_user=${sonatype_user:-yourapacheidhere} sonatype_pw=${sonatype_pw:-XXX} - +# whether only build the dist local and don't release to apache +IS_LOCAL_DIST=${IS_LOCAL_DIST:-false} +GIT_REPO=${GIT_REPO:-git-wip-us.apache.org/repos/asf/flink.git} +SCALA_VERSION=none +HADOOP_VERSION=none if [ "$(uname)" == "Darwin" ]; then SHASUM="shasum -a 512" @@ -85,18 +89,81 @@ else MD5SUM="md5sum" fi +usage() { + set +x + echo "./create_release_files.sh --scala-version 2.11 --hadoop-version 2.7.2" + echo "" + echo "usage:" + echo "[--scala-version ] [--hadoop-version ]" + echo "" + echo "example 1: build apache release" + echo " sonatype_user=APACHEID sonatype_pw=APACHEIDPASSWORD \ " + echo " NEW_VERSION=1.2.0 RELEASE_CANDIDATE="rc1" RELEASE_BRANCH=release-1.2.0 OLD_VERSION=1.1-SNAPSHOT \ " + echo " USER_NAME=APACHEID GPG_PASSPHRASE=XXX GPG_KEY=KEYID \ " + echo " GIT_AUTHOR=\"`git config --get user.name` <`git config --get user.email`>\" \ " + echo " GIT_REPO=github.com/apache/flink.git \ " + echo " ./create_release_files.sh --scala-version 2.11 --hadoop-version 2.7.2" + echo "" + echo "example 2: build local release" + echo " NEW_VERSION=1.2.0 RELEASE_BRANCH=master OLD_VERSION=1.2-SNAPSHOT \ " + echo " GPG_PASSPHRASE=XXX GPG_KEY=XXX IS_LOCAL_DIST=true \ " + echo " ./create_release_files.sh --scala-version 2.11 --hadoop-version 2.7.2" + + exit 1 +} + +# Parse arguments +while (( "$#" )); do + case $1 in +--scala-version) + SCALA_VERSION="$2" + shift + ;; +--hadoop-version) + HADOOP_VERSION="$2" + shift + ;; +--help) + usage + ;; +*) + break + ;; + esac + shift +done + +### prepare() { # prepare - git clone http://git-wip-us.apache.org/repos/asf/flink.git flink + target_branch=release-$RELEASE_VERSION + if [ "$RELEASE_CANDIDATE" != "none" ]; then +target_branch=$target_branch-$RELEASE_CANDIDATE + fi + + if [ ! -d ./flink ]; then +git clone http://$GIT_REPO flink + else +# if flink git repo exist, delete target branch, delete builded distribution +rm -rf flink-*.tgz +cd flink +# try-catch +{ + git pull --all + git checkout master + git branch -D $target_branch -f +} || { + echo "branch $target_branch not found" +} +cd .. + fi + cd flink - git checkout -b "release-$RELEASE_VERSION-$RELEASE_CANDIDATE" origin/$RELEASE_BRANCH - rm -f .gitignore - rm -f .gitattributes - rm -f .travis.yml - rm -f deploysettings.xml - rm -f CHANGELOG - rm -rf .github + + git checkout -b $target_branch origin/$RELEASE_BRANCH + rm -rf .gitignore .gitattributes .travis.yml deploysettings.xml CHANGELOG .github + cd .. } @@ -117,10
[1/2] flink git commit: [FLINK-5473] Limit max parallelism to 1 for non-parallel operators
Repository: flink Updated Branches: refs/heads/release-1.2 908376ba9 -> 993a2e2fa http://git-wip-us.apache.org/repos/asf/flink/blob/993a2e2f/flink-streaming-java/src/test/java/org/apache/flink/streaming/api/StreamExecutionEnvironmentTest.java -- diff --git a/flink-streaming-java/src/test/java/org/apache/flink/streaming/api/StreamExecutionEnvironmentTest.java b/flink-streaming-java/src/test/java/org/apache/flink/streaming/api/StreamExecutionEnvironmentTest.java index 3fb4513..3fc1344 100644 --- a/flink-streaming-java/src/test/java/org/apache/flink/streaming/api/StreamExecutionEnvironmentTest.java +++ b/flink-streaming-java/src/test/java/org/apache/flink/streaming/api/StreamExecutionEnvironmentTest.java @@ -17,19 +17,12 @@ package org.apache.flink.streaming.api; -import static org.junit.Assert.assertEquals; -import static org.junit.Assert.assertTrue; -import static org.junit.Assert.fail; - -import java.util.Arrays; -import java.util.Iterator; -import java.util.List; -import java.util.NoSuchElementException; - +import org.apache.flink.api.common.functions.FlatMapFunction; import org.apache.flink.api.common.typeinfo.BasicTypeInfo; import org.apache.flink.api.common.typeinfo.TypeInformation; import org.apache.flink.streaming.api.datastream.DataStream; import org.apache.flink.streaming.api.datastream.DataStreamSource; +import org.apache.flink.streaming.api.datastream.SingleOutputStreamOperator; import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment; import org.apache.flink.streaming.api.functions.sink.DiscardingSink; import org.apache.flink.streaming.api.functions.source.FromElementsFunction; @@ -38,10 +31,20 @@ import org.apache.flink.streaming.api.functions.source.StatefulSequenceSource; import org.apache.flink.streaming.api.graph.StreamGraph; import org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator; import org.apache.flink.streaming.api.operators.StreamOperator; +import org.apache.flink.util.Collector; import org.apache.flink.util.SplittableIterator; - +import org.junit.Assert; import org.junit.Test; +import java.util.Arrays; +import java.util.Iterator; +import java.util.List; +import java.util.NoSuchElementException; + +import static org.junit.Assert.assertEquals; +import static org.junit.Assert.assertTrue; +import static org.junit.Assert.fail; + public class StreamExecutionEnvironmentTest { @Test @@ -124,6 +127,102 @@ public class StreamExecutionEnvironmentTest { assertTrue(getFunctionFromDataSource(src4) instanceof FromElementsFunction); } + @Test + public void testParallelismBounds() { + StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); + + SourceFunction srcFun = new SourceFunction() { + private static final long serialVersionUID = 1L; + + @Override + public void run(SourceContext ctx) throws Exception { + } + + @Override + public void cancel() { + } + }; + + + SingleOutputStreamOperator operator = + env.addSource(srcFun).flatMap(new FlatMapFunction() { + + private static final long serialVersionUID = 1L; + + @Override + public void flatMap(Integer value, Collector out) throws Exception { + + } + }); + + // default value for max parallelism + Assert.assertEquals(-1, operator.getTransformation().getMaxParallelism()); + + // bounds for parallelism 1 + try { + operator.setParallelism(0); + Assert.fail(); + } catch (IllegalArgumentException expected) { + } + + // bounds for parallelism 2 + operator.setParallelism(1); + Assert.assertEquals(1, operator.getParallelism()); + + // bounds for parallelism 3 + operator.setParallelism(1 << 15); + Assert.assertEquals(1 << 15, operator.getParallelism()); + + // default value after generating + env.getStreamGraph().getJobGraph(); + Assert.assertEquals(-1, operator.getTransformation().getMaxParallelism()); + + // configured value after generating + env.setMaxParallelism(42); + env.getStreamGraph().getJobGraph(); + Assert.assertEquals(42, operator.getTransformation().getMaxParallelism()); + + // bounds configured parallelism 1 + try { + env.setMaxParallelism(0); + Assert.fail(); + } catch
[2/2] flink git commit: [FLINK-5473] Limit max parallelism to 1 for non-parallel operators
[FLINK-5473] Limit max parallelism to 1 for non-parallel operators [FLINK-5473] Better default behaviours for unspecified maximum parallelism This closes #3182. Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/993a2e2f Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/993a2e2f Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/993a2e2f Branch: refs/heads/release-1.2 Commit: 993a2e2fa0ceecff0979a267ace7cd7b8e05d359 Parents: 908376b Author: Stefan RichterAuthored: Mon Jan 16 14:31:22 2017 +0100 Committer: Till Rohrmann Committed: Tue Jan 24 09:49:52 2017 +0100 -- .../checkpoint/CheckpointCoordinator.java | 4 +- .../checkpoint/StateAssignmentOperation.java| 285 ++- .../checkpoint/savepoint/SavepointLoader.java | 8 +- .../ResultPartitionDeploymentDescriptor.java| 18 +- .../runtime/executiongraph/ExecutionGraph.java | 8 +- .../executiongraph/ExecutionJobVertex.java | 115 ++-- .../runtime/executiongraph/ExecutionVertex.java | 31 +- .../api/writer/ResultPartitionWriter.java | 4 + .../io/network/partition/ResultPartition.java | 8 + .../flink/runtime/jobgraph/JobVertex.java | 7 +- .../runtime/state/KeyGroupRangeAssignment.java | 42 ++- .../apache/flink/runtime/taskmanager/Task.java | 1 + .../checkpoint/CheckpointCoordinatorTest.java | 1 + .../savepoint/SavepointLoaderTest.java | 1 + ...ResultPartitionDeploymentDescriptorTest.java | 1 + .../executiongraph/ExecutionJobVertexTest.java | 140 + .../ExecutionVertexDeploymentTest.java | 27 +- .../network/partition/ResultPartitionTest.java | 1 + .../consumer/LocalInputChannelTest.java | 1 + .../runtime/jobmanager/JobManagerTest.java | 2 + .../runtime/taskmanager/TaskManagerTest.java| 5 +- .../streaming/api/datastream/KeyedStream.java | 13 +- .../datastream/SingleOutputStreamOperator.java | 29 +- .../environment/StreamExecutionEnvironment.java | 12 +- .../flink/streaming/api/graph/StreamGraph.java | 37 +-- .../api/graph/StreamGraphGenerator.java | 26 +- .../flink/streaming/api/graph/StreamNode.java | 2 - .../api/graph/StreamingJobGraphGenerator.java | 13 +- .../transformations/StreamTransformation.java | 5 + .../partitioner/KeyGroupStreamPartitioner.java | 1 + .../streaming/runtime/tasks/OperatorChain.java | 10 + .../api/StreamExecutionEnvironmentTest.java | 119 +++- .../api/graph/StreamGraphGeneratorTest.java | 9 +- .../test/checkpointing/RescalingITCase.java | 25 +- 34 files changed, 723 insertions(+), 288 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/993a2e2f/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CheckpointCoordinator.java -- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CheckpointCoordinator.java b/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CheckpointCoordinator.java index 9132897..78cad91 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CheckpointCoordinator.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CheckpointCoordinator.java @@ -885,8 +885,10 @@ public class CheckpointCoordinator { LOG.info("Restoring from latest valid checkpoint: {}.", latest); + final Map taskStates = latest.getTaskStates(); + StateAssignmentOperation stateAssignmentOperation = - new StateAssignmentOperation(LOG, tasks, latest, allowNonRestoredState); + new StateAssignmentOperation(LOG, tasks, taskStates, allowNonRestoredState); stateAssignmentOperation.assignStates(); http://git-wip-us.apache.org/repos/asf/flink/blob/993a2e2f/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperation.java -- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperation.java b/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperation.java index 6c23f02..6d075db 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperation.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperation.java @@ -45,29 +45,34 @@ public class StateAssignmentOperation { private final Logger logger; private final Map
[2/2] flink git commit: [FLINK-5617] Change reference version for API stability checks to 1.1.4
[FLINK-5617] Change reference version for API stability checks to 1.1.4 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/20c76a98 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/20c76a98 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/20c76a98 Branch: refs/heads/master Commit: 20c76a98b0b52d3fc164fbc0b8808a93cb50804f Parents: f0d96e3 Author: Robert MetzgerAuthored: Mon Jan 23 14:31:27 2017 +0100 Committer: Robert Metzger Committed: Tue Jan 24 12:01:35 2017 +0100 -- flink-core/pom.xml | 9 + flink-streaming-scala/pom.xml| 2 +- pom.xml | 2 +- tools/collect_japicmp_reports.sh | 31 +++ 4 files changed, 42 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/20c76a98/flink-core/pom.xml -- diff --git a/flink-core/pom.xml b/flink-core/pom.xml index 3debc72..e9738a2 100644 --- a/flink-core/pom.xml +++ b/flink-core/pom.xml @@ -156,6 +156,15 @@ under the License. org.apache.flink.api.common.ExecutionConfig#CONFIG_KEY org.apache.flink.core.fs.FileSystem$FSKey org.apache.flink.api.java.typeutils.WritableTypeInfo + + org.apache.flink.configuration.ConfigConstants#DEFAULT_JOB_MANAGER_MAX_ATTEMPTS_HISTORY_SIZE + org.apache.flink.configuration.ConfigConstants#DEFAULT_NETWORK_REQUEST_BACKOFF_INITIAL + org.apache.flink.configuration.ConfigConstants#DEFAULT_NETWORK_REQUEST_BACKOFF_MAX + org.apache.flink.configuration.ConfigConstants#DEFAULT_TASK_CANCELLATION_TIMEOUT_MILLIS + org.apache.flink.configuration.ConfigConstants#ENABLE_QUARANTINE_MONITOR + org.apache.flink.configuration.ConfigConstants#NETWORK_REQUEST_BACKOFF_INITIAL_KEY + org.apache.flink.configuration.ConfigConstants#NETWORK_REQUEST_BACKOFF_MAX_KEY http://git-wip-us.apache.org/repos/asf/flink/blob/20c76a98/flink-streaming-scala/pom.xml -- diff --git a/flink-streaming-scala/pom.xml b/flink-streaming-scala/pom.xml index d73ed38..f95ad79 100644 --- a/flink-streaming-scala/pom.xml +++ b/flink-streaming-scala/pom.xml @@ -237,7 +237,7 @@ under the License. japicmp-maven-plugin - + *\$\$anon\$* http://git-wip-us.apache.org/repos/asf/flink/blob/20c76a98/pom.xml -- diff --git a/pom.xml b/pom.xml index 389af84..2279b0b 100644 --- a/pom.xml +++ b/pom.xml @@ -1242,7 +1242,7 @@ under the License. org.apache.flink ${project.artifactId} - 1.0.0 + 1.1.4 ${project.packaging} http://git-wip-us.apache.org/repos/asf/flink/blob/20c76a98/tools/collect_japicmp_reports.sh -- diff --git a/tools/collect_japicmp_reports.sh b/tools/collect_japicmp_reports.sh new file mode 100755 index 000..6ff1238 --- /dev/null +++ b/tools/collect_japicmp_reports.sh @@ -0,0 +1,31 @@ +#!/bin/sh + +# +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright
[1/2] flink git commit: [FLINK-5617] Mark some methods as PublicEvolving or Internal
Repository: flink Updated Branches: refs/heads/master f0d96e30d -> 19ef82cf3 [FLINK-5617] Mark some methods as PublicEvolving or Internal This closes #3195 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/19ef82cf Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/19ef82cf Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/19ef82cf Branch: refs/heads/master Commit: 19ef82cf358da54bf204e678bf203f25147b11f9 Parents: 20c76a9 Author: Robert MetzgerAuthored: Mon Jan 23 16:28:48 2017 +0100 Committer: Robert Metzger Committed: Tue Jan 24 12:01:35 2017 +0100 -- .../main/java/org/apache/flink/api/common/ExecutionConfig.java | 2 ++ .../src/main/java/org/apache/flink/core/fs/FileSystem.java | 4 .../flink/streaming/api/datastream/AllWindowedStream.java | 5 - .../apache/flink/streaming/api/datastream/WindowedStream.java | 4 4 files changed, 14 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/19ef82cf/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java -- diff --git a/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java b/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java index 3cde5e7..32ea0a3 100644 --- a/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java +++ b/flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java @@ -19,6 +19,7 @@ package org.apache.flink.api.common; import com.esotericsoftware.kryo.Serializer; +import org.apache.flink.annotation.Internal; import org.apache.flink.annotation.Public; import org.apache.flink.annotation.PublicEvolving; import org.apache.flink.api.common.restartstrategy.RestartStrategies; @@ -848,6 +849,7 @@ public class ExecutionConfig implements Serializable, Archiveablehttp://git-wip-us.apache.org/repos/asf/flink/blob/19ef82cf/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java -- diff --git a/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java b/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java index e6313aa..33addbb 100644 --- a/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java +++ b/flink-core/src/main/java/org/apache/flink/core/fs/FileSystem.java @@ -25,6 +25,7 @@ package org.apache.flink.core.fs; +import org.apache.flink.annotation.Internal; import org.apache.flink.annotation.Public; import org.apache.flink.configuration.ConfigConstants; import org.apache.flink.configuration.Configuration; @@ -97,6 +98,7 @@ public abstract class FileSystem { * Create a SafetyNetCloseableRegistry for a Task. This method should be called at the beginning of the task's * main thread. */ + @Internal public static void createFileSystemCloseableRegistryForTask() { SafetyNetCloseableRegistry oldRegistry = REGISTRIES.get(); if (null != oldRegistry) { @@ -111,6 +113,7 @@ public abstract class FileSystem { * Create a SafetyNetCloseableRegistry for a Task. This method should be called at the end of the task's * main thread or when the task should be canceled. */ + @Internal public static void disposeFileSystemCloseableRegistryForTask() { SafetyNetCloseableRegistry registry = REGISTRIES.get(); if (null != registry) { @@ -200,6 +203,7 @@ public abstract class FileSystem { } } + @Internal public static FileSystem getUnguardedFileSystem(URI uri) throws IOException { FileSystem fs; http://git-wip-us.apache.org/repos/asf/flink/blob/19ef82cf/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AllWindowedStream.java -- diff --git a/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AllWindowedStream.java b/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AllWindowedStream.java index ae71ce5..31ea001 100644 --- a/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AllWindowedStream.java +++ b/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AllWindowedStream.java @@ -199,7 +199,7 @@ public class AllWindowedStream { * @param function The window function. * @return The data stream that is the result of applying the window function to the window. */ - + @PublicEvolving public SingleOutputStreamOperator
flink git commit: Commit for release 1.2.0
Repository: flink Updated Branches: refs/heads/release-1.2.0-rc1 [created] 1bca12088 Commit for release 1.2.0 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/1bca1208 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/1bca1208 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/1bca1208 Branch: refs/heads/release-1.2.0-rc1 Commit: 1bca120885f1950dc5d4436ae9f19025d26aad7c Parents: 993a2e2 Author: Robert MetzgerAuthored: Tue Jan 24 13:00:32 2017 + Committer: Robert Metzger Committed: Tue Jan 24 13:00:32 2017 + -- .gitattributes | 3 -- .github/CONTRIBUTING.md | 11 .github/PULL_REQUEST_TEMPLATE.md| 17 -- .gitignore | 36 - .travis.yml | 57 deploysettings.xml | 38 - docs/_config.yml| 4 +- flink-annotations/pom.xml | 2 +- flink-clients/pom.xml | 2 +- flink-connectors/flink-avro/pom.xml | 2 +- .../flink-connector-cassandra/pom.xml | 2 +- .../flink-connector-elasticsearch/pom.xml | 2 +- .../flink-connector-elasticsearch2/pom.xml | 2 +- .../flink-connector-filesystem/pom.xml | 2 +- .../flink-connector-kafka-0.10/pom.xml | 2 +- .../flink-connector-kafka-0.8/pom.xml | 2 +- .../flink-connector-kafka-0.9/pom.xml | 2 +- .../flink-connector-kafka-base/pom.xml | 2 +- .../flink-connector-kinesis/pom.xml | 2 +- flink-connectors/flink-connector-nifi/pom.xml | 2 +- .../flink-connector-rabbitmq/pom.xml| 2 +- .../flink-connector-twitter/pom.xml | 2 +- .../flink-hadoop-compatibility/pom.xml | 2 +- flink-connectors/flink-hbase/pom.xml| 2 +- flink-connectors/flink-hcatalog/pom.xml | 2 +- flink-connectors/flink-jdbc/pom.xml | 2 +- flink-connectors/pom.xml| 2 +- flink-contrib/flink-connector-wikiedits/pom.xml | 2 +- .../flink-statebackend-rocksdb/pom.xml | 2 +- flink-contrib/flink-storm-examples/pom.xml | 2 +- flink-contrib/flink-storm/pom.xml | 2 +- flink-contrib/flink-streaming-contrib/pom.xml | 2 +- flink-contrib/flink-tweet-inputformat/pom.xml | 2 +- flink-contrib/pom.xml | 2 +- flink-core/pom.xml | 2 +- flink-dist/pom.xml | 2 +- flink-examples/flink-examples-batch/pom.xml | 2 +- flink-examples/flink-examples-streaming/pom.xml | 2 +- flink-examples/pom.xml | 2 +- flink-fs-tests/pom.xml | 2 +- flink-java/pom.xml | 2 +- flink-java8/pom.xml | 2 +- flink-libraries/flink-cep-scala/pom.xml | 2 +- flink-libraries/flink-cep/pom.xml | 2 +- flink-libraries/flink-gelly-examples/pom.xml| 2 +- flink-libraries/flink-gelly-scala/pom.xml | 2 +- flink-libraries/flink-gelly/pom.xml | 2 +- flink-libraries/flink-ml/pom.xml| 2 +- flink-libraries/flink-python/pom.xml| 2 +- flink-libraries/flink-table/pom.xml | 2 +- flink-libraries/pom.xml | 2 +- flink-mesos/pom.xml | 2 +- flink-metrics/flink-metrics-core/pom.xml| 2 +- flink-metrics/flink-metrics-dropwizard/pom.xml | 2 +- flink-metrics/flink-metrics-ganglia/pom.xml | 2 +- flink-metrics/flink-metrics-graphite/pom.xml| 2 +- flink-metrics/flink-metrics-jmx/pom.xml | 2 +- flink-metrics/flink-metrics-statsd/pom.xml | 2 +- flink-metrics/pom.xml | 2 +- flink-optimizer/pom.xml | 2 +- flink-quickstart/flink-quickstart-java/pom.xml | 2 +- .../main/resources/archetype-resources/pom.xml | 2 +- flink-quickstart/flink-quickstart-scala/pom.xml | 2 +- .../main/resources/archetype-resources/pom.xml | 2 +- flink-quickstart/pom.xml| 2 +- flink-runtime-web/pom.xml | 2 +- flink-runtime/pom.xml | 2 +- flink-scala-shell/pom.xml | 2 +- flink-scala/pom.xml | 2 +- .../flink-shaded-curator-recipes/pom.xml| 2 +- .../flink-shaded-curator-test/pom.xml | 2 +- flink-shaded-curator/pom.xml| 2 +- .../flink-shaded-hadoop2/pom.xml| 2 +- .../flink-shaded-include-yarn-tests/pom.xml | 2 +-
flink git commit: Commit for release 1.2.0
Repository: flink Updated Branches: refs/heads/release-1.2.0-rc3 [created] 1c659cf4a Commit for release 1.2.0 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/1c659cf4 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/1c659cf4 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/1c659cf4 Branch: refs/heads/release-1.2.0-rc3 Commit: 1c659cf4acca2cbc8c36494156d9ec28423e43c6 Parents: 617ff50 Author: Robert MetzgerAuthored: Sun Jan 29 21:19:15 2017 + Committer: Robert Metzger Committed: Sun Jan 29 21:19:15 2017 + -- .gitattributes | 3 -- .github/CONTRIBUTING.md | 11 .github/PULL_REQUEST_TEMPLATE.md| 17 -- .gitignore | 36 - .travis.yml | 57 deploysettings.xml | 38 - docs/_config.yml| 4 +- flink-annotations/pom.xml | 2 +- flink-clients/pom.xml | 2 +- flink-connectors/flink-avro/pom.xml | 2 +- .../flink-connector-cassandra/pom.xml | 2 +- .../flink-connector-elasticsearch/pom.xml | 2 +- .../flink-connector-elasticsearch2/pom.xml | 2 +- .../flink-connector-filesystem/pom.xml | 2 +- .../flink-connector-kafka-0.10/pom.xml | 2 +- .../flink-connector-kafka-0.8/pom.xml | 2 +- .../flink-connector-kafka-0.9/pom.xml | 2 +- .../flink-connector-kafka-base/pom.xml | 2 +- .../flink-connector-kinesis/pom.xml | 2 +- flink-connectors/flink-connector-nifi/pom.xml | 2 +- .../flink-connector-rabbitmq/pom.xml| 2 +- .../flink-connector-twitter/pom.xml | 2 +- .../flink-hadoop-compatibility/pom.xml | 2 +- flink-connectors/flink-hbase/pom.xml| 2 +- flink-connectors/flink-hcatalog/pom.xml | 2 +- flink-connectors/flink-jdbc/pom.xml | 2 +- flink-connectors/pom.xml| 2 +- flink-contrib/flink-connector-wikiedits/pom.xml | 2 +- .../flink-statebackend-rocksdb/pom.xml | 2 +- flink-contrib/flink-storm-examples/pom.xml | 2 +- flink-contrib/flink-storm/pom.xml | 2 +- flink-contrib/flink-streaming-contrib/pom.xml | 2 +- flink-contrib/flink-tweet-inputformat/pom.xml | 2 +- flink-contrib/pom.xml | 2 +- flink-core/pom.xml | 2 +- flink-dist/pom.xml | 2 +- flink-examples/flink-examples-batch/pom.xml | 2 +- flink-examples/flink-examples-streaming/pom.xml | 2 +- flink-examples/pom.xml | 2 +- flink-fs-tests/pom.xml | 2 +- flink-java/pom.xml | 2 +- flink-java8/pom.xml | 2 +- flink-libraries/flink-cep-scala/pom.xml | 2 +- flink-libraries/flink-cep/pom.xml | 2 +- flink-libraries/flink-gelly-examples/pom.xml| 2 +- flink-libraries/flink-gelly-scala/pom.xml | 2 +- flink-libraries/flink-gelly/pom.xml | 2 +- flink-libraries/flink-ml/pom.xml| 2 +- flink-libraries/flink-python/pom.xml| 2 +- flink-libraries/flink-table/pom.xml | 2 +- flink-libraries/pom.xml | 2 +- flink-mesos/pom.xml | 2 +- flink-metrics/flink-metrics-core/pom.xml| 2 +- flink-metrics/flink-metrics-dropwizard/pom.xml | 2 +- flink-metrics/flink-metrics-ganglia/pom.xml | 2 +- flink-metrics/flink-metrics-graphite/pom.xml| 2 +- flink-metrics/flink-metrics-jmx/pom.xml | 2 +- flink-metrics/flink-metrics-statsd/pom.xml | 2 +- flink-metrics/pom.xml | 2 +- flink-optimizer/pom.xml | 2 +- flink-quickstart/flink-quickstart-java/pom.xml | 2 +- .../main/resources/archetype-resources/pom.xml | 2 +- flink-quickstart/flink-quickstart-scala/pom.xml | 2 +- .../main/resources/archetype-resources/pom.xml | 2 +- flink-quickstart/pom.xml| 2 +- flink-runtime-web/pom.xml | 2 +- flink-runtime/pom.xml | 2 +- flink-scala-shell/pom.xml | 2 +- flink-scala/pom.xml | 2 +- .../flink-shaded-curator-recipes/pom.xml| 2 +- .../flink-shaded-curator-test/pom.xml | 2 +- flink-shaded-curator/pom.xml| 2 +- .../flink-shaded-hadoop2/pom.xml| 2 +- .../flink-shaded-include-yarn-tests/pom.xml | 2 +-
flink git commit: [FLINK-5492] Fix TaskManager log entry
Repository: flink Updated Branches: refs/heads/release-1.2 d4b92bcf7 -> 887c074a7 [FLINK-5492] Fix TaskManager log entry Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/887c074a Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/887c074a Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/887c074a Branch: refs/heads/release-1.2 Commit: 887c074a769d821cc210468dbe82b7b702ff6db5 Parents: d4b92bc Author: Robert MetzgerAuthored: Wed Jan 25 14:27:29 2017 +0100 Committer: Robert Metzger Committed: Wed Jan 25 14:44:02 2017 +0100 -- .../org/apache/flink/runtime/taskmanager/TaskManager.scala | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/887c074a/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala -- diff --git a/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala b/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala index d2fb717..f1e74cd 100644 --- a/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala +++ b/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala @@ -1736,11 +1736,11 @@ object TaskManager { taskManagerClass: Class[_ <: TaskManager]) : Unit = { -LOG.info(s"Starting TaskManager") +LOG.info("Starting TaskManager") // Bring up the TaskManager actor system first, bind it to the given address. - -LOG.info("Starting TaskManager actor system at {}:{}.", taskManagerHostname, actorSystemPort) + +LOG.info(s"Starting TaskManager actor system at $taskManagerHostname:$actorSystemPort.") val taskManagerSystem = try { val akkaConfig = AkkaUtils.getAkkaConfig(
flink git commit: [FLINK-5492] Fix TaskManager log entry
Repository: flink Updated Branches: refs/heads/master 34e106f63 -> cf6b3fb22 [FLINK-5492] Fix TaskManager log entry Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/cf6b3fb2 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/cf6b3fb2 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/cf6b3fb2 Branch: refs/heads/master Commit: cf6b3fb22c67a540dbfa96497d66041ba95ad358 Parents: 34e106f Author: Robert MetzgerAuthored: Wed Jan 25 14:27:29 2017 +0100 Committer: Robert Metzger Committed: Wed Jan 25 14:45:04 2017 +0100 -- .../org/apache/flink/runtime/taskmanager/TaskManager.scala | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/cf6b3fb2/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala -- diff --git a/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala b/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala index 3c142cd..7cb1902 100644 --- a/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala +++ b/flink-runtime/src/main/scala/org/apache/flink/runtime/taskmanager/TaskManager.scala @@ -1745,11 +1745,11 @@ object TaskManager { taskManagerClass: Class[_ <: TaskManager]) : Unit = { -LOG.info(s"Starting TaskManager") +LOG.info("Starting TaskManager") // Bring up the TaskManager actor system first, bind it to the given address. - -LOG.info("Starting TaskManager actor system at {}:{}.", taskManagerHostname, actorSystemPort) + +LOG.info(s"Starting TaskManager actor system at $taskManagerHostname:$actorSystemPort.") val taskManagerSystem = try { val akkaConfig = AkkaUtils.getAkkaConfig(
flink git commit: [FLINK-5637] Avoid warning while parsing YAML configuration files
Repository: flink Updated Branches: refs/heads/release-1.2 80c9f8018 -> f2240eb93 [FLINK-5637] Avoid warning while parsing YAML configuration files This closes #3216 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/f2240eb9 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/f2240eb9 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/f2240eb9 Branch: refs/heads/release-1.2 Commit: f2240eb9394b368b392ac83d37106482a005e0a2 Parents: 80c9f80 Author: Robert MetzgerAuthored: Wed Jan 25 16:31:11 2017 +0100 Committer: Robert Metzger Committed: Wed Jan 25 18:01:25 2017 +0100 -- .../org/apache/flink/configuration/GlobalConfiguration.java | 9 + .../apache/flink/configuration/GlobalConfigurationTest.java | 3 ++- flink-dist/src/main/resources/flink-conf.yaml | 2 +- 3 files changed, 8 insertions(+), 6 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/f2240eb9/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java -- diff --git a/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java b/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java index dca6307..ea9f8bf 100644 --- a/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java +++ b/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java @@ -144,11 +144,12 @@ public final class GlobalConfiguration { try (BufferedReader reader = new BufferedReader(new InputStreamReader(new FileInputStream(file{ String line; + int lineNo = 0; while ((line = reader.readLine()) != null) { - + lineNo++; // 1. check for comments String[] comments = line.split("#", 2); - String conf = comments[0]; + String conf = comments[0].trim(); // 2. get key and value if (conf.length() > 0) { @@ -156,7 +157,7 @@ public final class GlobalConfiguration { // skip line with no valid key-value pair if (kv.length == 1) { - LOG.warn("Error while trying to split key and value in configuration file " + file + ": " + line); + LOG.warn("Error while trying to split key and value in configuration file " + file + ":" + lineNo + ": \"" + line + "\""); continue; } @@ -165,7 +166,7 @@ public final class GlobalConfiguration { // sanity check if (key.length() == 0 || value.length() == 0) { - LOG.warn("Error after splitting key and value in configuration file " + file + ": " + line); + LOG.warn("Error after splitting key and value in configuration file " + file + ":" + lineNo + ": \"" + line + "\""); continue; } http://git-wip-us.apache.org/repos/asf/flink/blob/f2240eb9/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java -- diff --git a/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java b/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java index 6336a73..322e64d 100644 --- a/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java +++ b/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java @@ -56,7 +56,8 @@ public class GlobalConfigurationTest extends TestLogger { pw.println("mykey3:myvalue3"); // SKIP, missing white space after colon pw.println(" some nonsense without colon and whitespace separator"); // SKIP pw.println(" : "); // SKIP - pw.println(" "); // SKIP + pw.println(" "); // SKIP (silently) + pw.println(" "); // SKIP (silently) pw.println("mykey4: myvalue4# some comments"); // OK, skip comments only
flink git commit: [FLINK-5637] Avoid warning while parsing YAML configuration files
Repository: flink Updated Branches: refs/heads/master b9ea4cfdf -> 54d3e2668 [FLINK-5637] Avoid warning while parsing YAML configuration files Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/54d3e266 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/54d3e266 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/54d3e266 Branch: refs/heads/master Commit: 54d3e26686998a0e5e472651cd9d709d491e7093 Parents: b9ea4cf Author: Robert MetzgerAuthored: Wed Jan 25 16:31:11 2017 +0100 Committer: Robert Metzger Committed: Wed Jan 25 18:02:00 2017 +0100 -- .../org/apache/flink/configuration/GlobalConfiguration.java | 9 + .../apache/flink/configuration/GlobalConfigurationTest.java | 3 ++- flink-dist/src/main/resources/flink-conf.yaml | 2 +- 3 files changed, 8 insertions(+), 6 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/54d3e266/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java -- diff --git a/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java b/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java index dca6307..ea9f8bf 100644 --- a/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java +++ b/flink-core/src/main/java/org/apache/flink/configuration/GlobalConfiguration.java @@ -144,11 +144,12 @@ public final class GlobalConfiguration { try (BufferedReader reader = new BufferedReader(new InputStreamReader(new FileInputStream(file{ String line; + int lineNo = 0; while ((line = reader.readLine()) != null) { - + lineNo++; // 1. check for comments String[] comments = line.split("#", 2); - String conf = comments[0]; + String conf = comments[0].trim(); // 2. get key and value if (conf.length() > 0) { @@ -156,7 +157,7 @@ public final class GlobalConfiguration { // skip line with no valid key-value pair if (kv.length == 1) { - LOG.warn("Error while trying to split key and value in configuration file " + file + ": " + line); + LOG.warn("Error while trying to split key and value in configuration file " + file + ":" + lineNo + ": \"" + line + "\""); continue; } @@ -165,7 +166,7 @@ public final class GlobalConfiguration { // sanity check if (key.length() == 0 || value.length() == 0) { - LOG.warn("Error after splitting key and value in configuration file " + file + ": " + line); + LOG.warn("Error after splitting key and value in configuration file " + file + ":" + lineNo + ": \"" + line + "\""); continue; } http://git-wip-us.apache.org/repos/asf/flink/blob/54d3e266/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java -- diff --git a/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java b/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java index 6336a73..322e64d 100644 --- a/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java +++ b/flink-core/src/test/java/org/apache/flink/configuration/GlobalConfigurationTest.java @@ -56,7 +56,8 @@ public class GlobalConfigurationTest extends TestLogger { pw.println("mykey3:myvalue3"); // SKIP, missing white space after colon pw.println(" some nonsense without colon and whitespace separator"); // SKIP pw.println(" : "); // SKIP - pw.println(" "); // SKIP + pw.println(" "); // SKIP (silently) + pw.println(" "); // SKIP (silently) pw.println("mykey4: myvalue4# some comments"); // OK, skip comments only pw.println(" mykey5
flink git commit: Commit for release 1.2.0
Repository: flink Updated Branches: refs/heads/release-1.2.0-rc2 [created] 8b5b6a8b2 Commit for release 1.2.0 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/8b5b6a8b Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/8b5b6a8b Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/8b5b6a8b Branch: refs/heads/release-1.2.0-rc2 Commit: 8b5b6a8b27516578ddf05c5e77c840f5666bb28f Parents: f2240eb Author: Robert MetzgerAuthored: Wed Jan 25 17:17:29 2017 + Committer: Robert Metzger Committed: Wed Jan 25 17:17:29 2017 + -- .gitattributes | 3 -- .github/CONTRIBUTING.md | 11 .github/PULL_REQUEST_TEMPLATE.md| 17 -- .gitignore | 36 - .travis.yml | 57 deploysettings.xml | 38 - docs/_config.yml| 4 +- flink-annotations/pom.xml | 2 +- flink-clients/pom.xml | 2 +- flink-connectors/flink-avro/pom.xml | 2 +- .../flink-connector-cassandra/pom.xml | 2 +- .../flink-connector-elasticsearch/pom.xml | 2 +- .../flink-connector-elasticsearch2/pom.xml | 2 +- .../flink-connector-filesystem/pom.xml | 2 +- .../flink-connector-kafka-0.10/pom.xml | 2 +- .../flink-connector-kafka-0.8/pom.xml | 2 +- .../flink-connector-kafka-0.9/pom.xml | 2 +- .../flink-connector-kafka-base/pom.xml | 2 +- .../flink-connector-kinesis/pom.xml | 2 +- flink-connectors/flink-connector-nifi/pom.xml | 2 +- .../flink-connector-rabbitmq/pom.xml| 2 +- .../flink-connector-twitter/pom.xml | 2 +- .../flink-hadoop-compatibility/pom.xml | 2 +- flink-connectors/flink-hbase/pom.xml| 2 +- flink-connectors/flink-hcatalog/pom.xml | 2 +- flink-connectors/flink-jdbc/pom.xml | 2 +- flink-connectors/pom.xml| 2 +- flink-contrib/flink-connector-wikiedits/pom.xml | 2 +- .../flink-statebackend-rocksdb/pom.xml | 2 +- flink-contrib/flink-storm-examples/pom.xml | 2 +- flink-contrib/flink-storm/pom.xml | 2 +- flink-contrib/flink-streaming-contrib/pom.xml | 2 +- flink-contrib/flink-tweet-inputformat/pom.xml | 2 +- flink-contrib/pom.xml | 2 +- flink-core/pom.xml | 2 +- flink-dist/pom.xml | 2 +- flink-examples/flink-examples-batch/pom.xml | 2 +- flink-examples/flink-examples-streaming/pom.xml | 2 +- flink-examples/pom.xml | 2 +- flink-fs-tests/pom.xml | 2 +- flink-java/pom.xml | 2 +- flink-java8/pom.xml | 2 +- flink-libraries/flink-cep-scala/pom.xml | 2 +- flink-libraries/flink-cep/pom.xml | 2 +- flink-libraries/flink-gelly-examples/pom.xml| 2 +- flink-libraries/flink-gelly-scala/pom.xml | 2 +- flink-libraries/flink-gelly/pom.xml | 2 +- flink-libraries/flink-ml/pom.xml| 2 +- flink-libraries/flink-python/pom.xml| 2 +- flink-libraries/flink-table/pom.xml | 2 +- flink-libraries/pom.xml | 2 +- flink-mesos/pom.xml | 2 +- flink-metrics/flink-metrics-core/pom.xml| 2 +- flink-metrics/flink-metrics-dropwizard/pom.xml | 2 +- flink-metrics/flink-metrics-ganglia/pom.xml | 2 +- flink-metrics/flink-metrics-graphite/pom.xml| 2 +- flink-metrics/flink-metrics-jmx/pom.xml | 2 +- flink-metrics/flink-metrics-statsd/pom.xml | 2 +- flink-metrics/pom.xml | 2 +- flink-optimizer/pom.xml | 2 +- flink-quickstart/flink-quickstart-java/pom.xml | 2 +- .../main/resources/archetype-resources/pom.xml | 2 +- flink-quickstart/flink-quickstart-scala/pom.xml | 2 +- .../main/resources/archetype-resources/pom.xml | 2 +- flink-quickstart/pom.xml| 2 +- flink-runtime-web/pom.xml | 2 +- flink-runtime/pom.xml | 2 +- flink-scala-shell/pom.xml | 2 +- flink-scala/pom.xml | 2 +- .../flink-shaded-curator-recipes/pom.xml| 2 +- .../flink-shaded-curator-test/pom.xml | 2 +- flink-shaded-curator/pom.xml| 2 +- .../flink-shaded-hadoop2/pom.xml| 2 +- .../flink-shaded-include-yarn-tests/pom.xml | 2 +-
flink git commit: FLINK-5731 Spilt up tests into three disjoint groups
Repository: flink Updated Branches: refs/heads/master b0f0f3722 -> e7a914d4e FLINK-5731 Spilt up tests into three disjoint groups This closes #3344 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/e7a914d4 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/e7a914d4 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/e7a914d4 Branch: refs/heads/master Commit: e7a914d4e926216e1382e622556be8c8358b27c9 Parents: b0f0f37 Author: Robert MetzgerAuthored: Mon Feb 13 21:43:25 2017 +0100 Committer: Robert Metzger Committed: Wed Feb 22 20:29:49 2017 +0100 -- .travis.yml | 9 + pom.xml | 13 +++-- 2 files changed, 20 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/e7a914d4/.travis.yml -- diff --git a/.travis.yml b/.travis.yml index 1445f8d..74bec6b 100644 --- a/.travis.yml +++ b/.travis.yml @@ -19,21 +19,30 @@ matrix: env: PROFILE="-Dhadoop.version=2.7.2 -Dscala-2.11 -Pinclude-yarn-tests,flink-fast-tests-a,include-kinesis -Dmaven.javadoc.skip=true" - jdk: "oraclejdk8" env: PROFILE="-Dhadoop.version=2.7.2 -Dscala-2.11 -Pinclude-yarn-tests,flink-fast-tests-b,include-kinesis -Dmaven.javadoc.skip=true" +- jdk: "oraclejdk8" + env: PROFILE="-Dhadoop.version=2.7.2 -Dscala-2.11 -Pinclude-yarn-tests,flink-fast-tests-c,include-kinesis -Dmaven.javadoc.skip=true" - jdk: "oraclejdk8" env: PROFILE="-Dhadoop.version=2.6.3 -Pinclude-yarn-tests,flink-fast-tests-a,include-kinesis -Dmaven.javadoc.skip=true" - jdk: "oraclejdk8" env: PROFILE="-Dhadoop.version=2.6.3 -Pinclude-yarn-tests,flink-fast-tests-b,include-kinesis -Dmaven.javadoc.skip=true" +- jdk: "oraclejdk8" + env: PROFILE="-Dhadoop.version=2.6.3 -Pinclude-yarn-tests,flink-fast-tests-c,include-kinesis -Dmaven.javadoc.skip=true" - jdk: "openjdk7" env: PROFILE="-Dhadoop.version=2.4.1 -Dscala-2.11 -Pinclude-yarn-tests,flink-fast-tests-a,include-kinesis" - jdk: "openjdk7" env: PROFILE="-Dhadoop.version=2.4.1 -Dscala-2.11 -Pinclude-yarn-tests,flink-fast-tests-b,include-kinesis" +- jdk: "openjdk7" + env: PROFILE="-Dhadoop.version=2.4.1 -Dscala-2.11 -Pinclude-yarn-tests,flink-fast-tests-c,include-kinesis" - jdk: "oraclejdk7" env: PROFILE="-Dhadoop.version=2.3.0 -Pflink-fast-tests-a,include-kinesis" - jdk: "oraclejdk7" env: PROFILE="-Dhadoop.version=2.3.0 -Pflink-fast-tests-b,include-kinesis" +- jdk: "oraclejdk7" + env: PROFILE="-Dhadoop.version=2.3.0 -Pflink-fast-tests-c,include-kinesis" + git: http://git-wip-us.apache.org/repos/asf/flink/blob/e7a914d4/pom.xml -- diff --git a/pom.xml b/pom.xml index d81296e..0070835 100644 --- a/pom.xml +++ b/pom.xml @@ -767,13 +767,22 @@ under the License. flink-fast-tests-a - %regex[.*/[A-M].*] + + %regex[.*/[J-Z].*] flink-fast-tests-b - %regex[.*/[N-Z].*] + + %regex[.*/[A-IS-Z].*] + + + + flink-fast-tests-c + + + %regex[.*/[A-R].*]
flink git commit: [FLINK-5690][docs] Add note on shading to best practices guide
Repository: flink Updated Branches: refs/heads/master 8efacf588 -> d32281444 [FLINK-5690][docs] Add note on shading to best practices guide This closes #3300 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/d3228144 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/d3228144 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/d3228144 Branch: refs/heads/master Commit: d322814447e9ce437eb1adffef24699f26803038 Parents: 8efacf5 Author: Robert MetzgerAuthored: Mon Feb 13 20:50:23 2017 +0100 Committer: Robert Metzger Committed: Tue Feb 14 15:05:16 2017 +0100 -- docs/dev/best_practices.md| 314 + docs/monitoring/best_practices.md | 312 docs/monitoring/debugging_classloading.md | 14 ++ 3 files changed, 328 insertions(+), 312 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/d3228144/docs/dev/best_practices.md -- diff --git a/docs/dev/best_practices.md b/docs/dev/best_practices.md new file mode 100644 index 000..b2111c4 --- /dev/null +++ b/docs/dev/best_practices.md @@ -0,0 +1,314 @@ +--- +title: "Best Practices" +nav-parent_id: dev +nav-pos: 90 +--- + + +This page contains a collection of best practices for Flink programmers on how to solve frequently encountered problems. + + +* This will be replaced by the TOC +{:toc} + +## Parsing command line arguments and passing them around in your Flink application + + +Almost all Flink applications, both batch and streaming rely on external configuration parameters. +For example for specifying input and output sources (like paths or addresses), also system parameters (parallelism, runtime configuration) and application specific parameters (often used within the user functions). + +Since version 0.9 we are providing a simple utility called `ParameterTool` to provide at least some basic tooling for solving these problems. + +Please note that you don't have to use the `ParameterTool` explained here. Other frameworks such as [Commons CLI](https://commons.apache.org/proper/commons-cli/), +[argparse4j](http://argparse4j.sourceforge.net/) and others work well with Flink as well. + + +### Getting your configuration values into the `ParameterTool` + +The `ParameterTool` provides a set of predefined static methods for reading the configuration. The tool is internally expecting a `Map `, so its very easy to integrate it with your own configuration style. + + + From `.properties` files + +The following method will read a [Properties](https://docs.oracle.com/javase/tutorial/essential/environment/properties.html) file and provide the key/value pairs: +{% highlight java %} +String propertiesFile = "/home/sam/flink/myjob.properties"; +ParameterTool parameter = ParameterTool.fromPropertiesFile(propertiesFile); +{% endhighlight %} + + + From the command line arguments + +This allows getting arguments like `--input hdfs:///mydata --elements 42` from the command line. +{% highlight java %} +public static void main(String[] args) { + ParameterTool parameter = ParameterTool.fromArgs(args); + // .. regular code .. +{% endhighlight %} + + + From system properties + +When starting a JVM, you can pass system properties to it: `-Dinput=hdfs:///mydata`. You can also initialize the `ParameterTool` from these system properties: + +{% highlight java %} +ParameterTool parameter = ParameterTool.fromSystemProperties(); +{% endhighlight %} + + +### Using the parameters in your Flink program + +Now that we've got the parameters from somewhere (see above) we can use them in various ways. + +**Directly from the `ParameterTool`** + +The `ParameterTool` itself has methods for accessing the values. +{% highlight java %} +ParameterTool parameters = // ... +parameter.getRequired("input"); +parameter.get("output", "myDefaultValue"); +parameter.getLong("expectedCount", -1L); +parameter.getNumberOfParameters() +// .. there are more methods available. +{% endhighlight %} + +You can use the return values of these methods directly in the main() method (=the client submitting the application). +For example you could set the parallelism of a operator like this: + +{% highlight java %} +ParameterTool parameters = ParameterTool.fromArgs(args); +int parallelism = parameters.get("mapParallelism", 2); +DataSet > counts = text.flatMap(new Tokenizer()).setParallelism(parallelism); +{% endhighlight %} + +Since the `ParameterTool` is serializable, you can pass it to the functions itself: + +{% highlight java %} +ParameterTool parameters = ParameterTool.fromArgs(args); +DataSet > counts =
flink git commit: [FLINK-5613][query] querying a non-existing key is inconsistent among state backends
Repository: flink Updated Branches: refs/heads/master 24fe08148 -> da26bdc2e [FLINK-5613][query] querying a non-existing key is inconsistent among state backends Querying for a non-existing key for a state that has a default value set currently results in an UnknownKeyOrNamespace exception when the MemoryStateBackend or FsStateBackend is used but results in the default value if RocksDBStateBackend is set. This removes the special handling from the RocksDBStateBackend and makes it consistent with the other two back-ends, i.e. returning null which results in the mentioned UnknownKeyOrNamespace exception. This closes #3193 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/da26bdc2 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/da26bdc2 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/da26bdc2 Branch: refs/heads/master Commit: da26bdc2eda1c8d70ada8b1969db1d48d2095ed1 Parents: 24fe081 Author: Nico KruberAuthored: Tue Jan 17 14:26:16 2017 +0100 Committer: Robert Metzger Committed: Mon Jan 23 18:47:44 2017 +0100 -- .../streaming/state/RocksDBValueState.java | 12 -- .../flink/test/query/QueryableStateITCase.java | 160 ++- 2 files changed, 153 insertions(+), 19 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/da26bdc2/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java -- diff --git a/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java b/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java index 7724f02..b2a4fba 100644 --- a/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java +++ b/flink-contrib/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBValueState.java @@ -23,7 +23,6 @@ import org.apache.flink.api.common.state.ValueStateDescriptor; import org.apache.flink.api.common.typeutils.TypeSerializer; import org.apache.flink.core.memory.DataInputViewStreamWrapper; import org.apache.flink.core.memory.DataOutputViewStreamWrapper; -import org.apache.flink.runtime.query.netty.message.KvStateRequestSerializer; import org.apache.flink.runtime.state.internal.InternalValueState; import org.rocksdb.ColumnFamilyHandle; import org.rocksdb.RocksDBException; @@ -103,15 +102,4 @@ public class RocksDBValueState throw new RuntimeException("Error while adding data to RocksDB", e); } } - - @Override - public byte[] getSerializedValue(byte[] serializedKeyAndNamespace) throws Exception { - byte[] value = super.getSerializedValue(serializedKeyAndNamespace); - - if (value != null) { - return value; - } else { - return KvStateRequestSerializer.serializeValue(stateDesc.getDefaultValue(), stateDesc.getSerializer()); - } - } } http://git-wip-us.apache.org/repos/asf/flink/blob/da26bdc2/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java -- diff --git a/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java b/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java index 327a715..c2df6ae 100644 --- a/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java +++ b/flink-tests/src/test/java/org/apache/flink/test/query/QueryableStateITCase.java @@ -39,6 +39,7 @@ import org.apache.flink.api.java.tuple.Tuple2; import org.apache.flink.configuration.ConfigConstants; import org.apache.flink.configuration.Configuration; import org.apache.flink.configuration.QueryableStateOptions; +import org.apache.flink.contrib.streaming.state.RocksDBStateBackend; import org.apache.flink.runtime.akka.AkkaUtils; import org.apache.flink.runtime.executiongraph.ExecutionAttemptID; import org.apache.flink.runtime.executiongraph.ExecutionGraph; @@ -50,10 +51,14 @@ import org.apache.flink.runtime.messages.JobManagerMessages.JobFound; import org.apache.flink.runtime.query.KvStateLocation; import org.apache.flink.runtime.query.KvStateMessage; import org.apache.flink.runtime.query.QueryableStateClient; +import org.apache.flink.runtime.query.netty.UnknownKeyOrNamespace; import org.apache.flink.runtime.query.netty.message.KvStateRequestSerializer; +import org.apache.flink.runtime.state.AbstractStateBackend; import
flink git commit: [FLINK-5395][release script] Fix version commit and maven upload
Repository: flink Updated Branches: refs/heads/master b351aa2dd -> 5f0d8c9dd [FLINK-5395][release script] Fix version commit and maven upload Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/5f0d8c9d Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/5f0d8c9d Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/5f0d8c9d Branch: refs/heads/master Commit: 5f0d8c9dd56935f4a62ee3f57aa2d411f102bead Parents: b351aa2 Author: Robert MetzgerAuthored: Tue Jan 24 15:14:55 2017 +0100 Committer: Robert Metzger Committed: Tue Jan 24 15:14:55 2017 +0100 -- tools/create_release_files.sh | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/5f0d8c9d/tools/create_release_files.sh -- diff --git a/tools/create_release_files.sh b/tools/create_release_files.sh index ed01d97..f4c4673 100755 --- a/tools/create_release_files.sh +++ b/tools/create_release_files.sh @@ -185,7 +185,7 @@ make_source_release() { cd .. # local dist have no need to commit to remote - if [ ! IS_LOCAL_DIST ]; then + if [ "$IS_LOCAL_DIST" == "false" ]; then git commit --author="$GIT_AUTHOR" -am "Commit for release $RELEASE_VERSION" git remote add asf_push https://$USER_NAME@$GIT_REPO RELEASE_HASH=`git rev-parse HEAD` @@ -299,7 +299,7 @@ else make_binary_release "hadoop2x" "-Dhadoop.version=$HADOOP_VERSION" "$SCALA_VERSION" fi -if [ ! IS_LOCAL_DIST ] ; then +if [ "$IS_LOCAL_DIST" == "false" ] ; then copy_data deploy_to_maven fi
flink git commit: [FLINK-3150] Make YARN container invocation configurable
Repository: flink Updated Branches: refs/heads/master 6fb6967b9 -> 8f4139a42 [FLINK-3150] Make YARN container invocation configurable This closes #3056 Project: http://git-wip-us.apache.org/repos/asf/flink/repo Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/8f4139a4 Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/8f4139a4 Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/8f4139a4 Branch: refs/heads/master Commit: 8f4139a42a619738a4efa45ea0b5d20009718aba Parents: 6fb6967 Author: Nico KruberAuthored: Mon Jan 2 16:17:09 2017 +0100 Committer: Robert Metzger Committed: Wed Jan 18 11:17:02 2017 +0100 -- docs/setup/config.md| 5 + .../flink/configuration/ConfigConstants.java| 12 ++ .../clusterframework/BootstrapTools.java| 95 +--- .../clusterframework/BootstrapToolsTest.java| 148 ++ .../yarn/AbstractYarnClusterDescriptor.java | 46 -- .../flink/yarn/YarnClusterDescriptorTest.java | 149 +++ 6 files changed, 416 insertions(+), 39 deletions(-) -- http://git-wip-us.apache.org/repos/asf/flink/blob/8f4139a4/docs/setup/config.md -- diff --git a/docs/setup/config.md b/docs/setup/config.md index 1b2be8a..1a72e27 100644 --- a/docs/setup/config.md +++ b/docs/setup/config.md @@ -425,6 +425,11 @@ of the JobManager, because the same ActorSystem is used. Its not possible to use - `yarn.taskmanager.env.` Similar to the configuration prefix about, this prefix allows setting custom environment variables for the TaskManager processes. +- `yarn.container-start-command-template`: Flink uses the following template when starting on YARN: +`%java% %jvmmem% %jvmopts% %logging% %class% %args% %redirects%`. This configuration parameter allows users +to pass custom settings (such as JVM paths, arguments etc.). Note that in most cases, it is sufficient to +use the `env.java.opts` setting, which is the `%jvmopts%` variable in the String. + - `yarn.application-master.port` (Default: 0, which lets the OS choose an ephemeral port) With this configuration option, users can specify a port, a range of ports or a list of ports for the Application Master (and JobManager) RPC port. By default we recommend using the default value (0) to let the operating system choose an appropriate port. In particular when multiple AMs are running on the same physical host, fixed port assignments prevent the AM from starting. For example when running Flink on YARN on an environment with a restrictive firewall, this option allows specifying a range of allowed ports. http://git-wip-us.apache.org/repos/asf/flink/blob/8f4139a4/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java -- diff --git a/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java b/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java index fc389e0..14ba9dd 100644 --- a/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java +++ b/flink-core/src/main/java/org/apache/flink/configuration/ConfigConstants.java @@ -423,6 +423,12 @@ public final class ConfigConstants { */ @Deprecated public static final String YARN_TASK_MANAGER_ENV_PREFIX = "yarn.taskmanager.env."; + + /** +* Template for the YARN container start incovation. +*/ + public static final String YARN_CONTAINER_START_COMMAND_TEMPLATE = + "yarn.container-start-command-template"; /** * The config parameter defining the Akka actor system port for the ApplicationMaster and @@ -1146,6 +1152,12 @@ public final class ConfigConstants { public static final float DEFAULT_YARN_HEAP_CUTOFF_RATIO = 0.25f; /** +* Start command template for Flink on YARN containers +*/ + public static final String DEFAULT_YARN_CONTAINER_START_COMMAND_TEMPLATE = + "%java% %jvmmem% %jvmopts% %logging% %class% %args% %redirects%"; + + /** * Default port for the application master is 0, which means * the operating system assigns an ephemeral port */ http://git-wip-us.apache.org/repos/asf/flink/blob/8f4139a4/flink-runtime/src/main/java/org/apache/flink/runtime/clusterframework/BootstrapTools.java -- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/clusterframework/BootstrapTools.java b/flink-runtime/src/main/java/org/apache/flink/runtime/clusterframework/BootstrapTools.java index 2a33c44..1ef2cdc 100644 ---
flink-web git commit: Added a page that redirects to the latest stable docs. We can link to this page from older versions of the docs.
Repository: flink-web Updated Branches: refs/heads/asf-site 9ec0a879f -> ccdf85ae5 Added a page that redirects to the latest stable docs. We can link to this page from older versions of the docs. Project: http://git-wip-us.apache.org/repos/asf/flink-web/repo Commit: http://git-wip-us.apache.org/repos/asf/flink-web/commit/ccdf85ae Tree: http://git-wip-us.apache.org/repos/asf/flink-web/tree/ccdf85ae Diff: http://git-wip-us.apache.org/repos/asf/flink-web/diff/ccdf85ae Branch: refs/heads/asf-site Commit: ccdf85ae593228ca0a728470b73b910122262217 Parents: 9ec0a87 Author: David AndersonAuthored: Thu Jan 19 15:39:21 2017 +0100 Committer: David Anderson Committed: Thu Jan 19 15:39:21 2017 +0100 -- content/q/stable-docs.html | 30 ++ q/stable-docs.html | 30 ++ 2 files changed, 60 insertions(+) -- http://git-wip-us.apache.org/repos/asf/flink-web/blob/ccdf85ae/content/q/stable-docs.html -- diff --git a/content/q/stable-docs.html b/content/q/stable-docs.html new file mode 100644 index 000..70907a8 --- /dev/null +++ b/content/q/stable-docs.html @@ -0,0 +1,30 @@ + + + + + + +https://ci.apache.org/projects/flink/flink-docs-release-1.1/; /> + + http://git-wip-us.apache.org/repos/asf/flink-web/blob/ccdf85ae/q/stable-docs.html -- diff --git a/q/stable-docs.html b/q/stable-docs.html new file mode 100644 index 000..70907a8 --- /dev/null +++ b/q/stable-docs.html @@ -0,0 +1,30 @@ + + + + + + +https://ci.apache.org/projects/flink/flink-docs-release-1.1/; /> + +
[03/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/js/jquery-2.1.0.js -- diff --git a/content/visualizer/js/jquery-2.1.0.js b/content/visualizer/js/jquery-2.1.0.js deleted file mode 100755 index f7f4227..000 --- a/content/visualizer/js/jquery-2.1.0.js +++ /dev/null @@ -1,9111 +0,0 @@ -/*! - * jQuery JavaScript Library v2.1.0 - * http://jquery.com/ - * - * Includes Sizzle.js - * http://sizzlejs.com/ - * - * Copyright 2005, 2014 jQuery Foundation, Inc. and other contributors - * Released under the MIT license - * http://jquery.org/license - * - * Date: 2014-01-23T21:10Z - */ - -(function( global, factory ) { - - if ( typeof module === "object" && typeof module.exports === "object" ) { - // For CommonJS and CommonJS-like environments where a proper window is present, - // execute the factory and get jQuery - // For environments that do not inherently posses a window with a document - // (such as Node.js), expose a jQuery-making factory as module.exports - // This accentuates the need for the creation of a real window - // e.g. var jQuery = require("jquery")(window); - // See ticket #14549 for more info - module.exports = global.document ? - factory( global, true ) : - function( w ) { - if ( !w.document ) { - throw new Error( "jQuery requires a window with a document" ); - } - return factory( w ); - }; - } else { - factory( global ); - } - -// Pass this if window is not defined yet -}(typeof window !== "undefined" ? window : this, function( window, noGlobal ) { - -// Can't do this because several apps including ASP.NET trace -// the stack via arguments.caller.callee and Firefox dies if -// you try to trace through "use strict" call chains. (#13335) -// Support: Firefox 18+ -// - -var arr = []; - -var slice = arr.slice; - -var concat = arr.concat; - -var push = arr.push; - -var indexOf = arr.indexOf; - -var class2type = {}; - -var toString = class2type.toString; - -var hasOwn = class2type.hasOwnProperty; - -var trim = "".trim; - -var support = {}; - - - -var - // Use the correct document accordingly with window argument (sandbox) - document = window.document, - - version = "2.1.0", - - // Define a local copy of jQuery - jQuery = function( selector, context ) { - // The jQuery object is actually just the init constructor 'enhanced' - // Need init if jQuery is called (just allow error to be thrown if not included) - return new jQuery.fn.init( selector, context ); - }, - - // Matches dashed string for camelizing - rmsPrefix = /^-ms-/, - rdashAlpha = /-([\da-z])/gi, - - // Used by jQuery.camelCase as callback to replace() - fcamelCase = function( all, letter ) { - return letter.toUpperCase(); - }; - -jQuery.fn = jQuery.prototype = { - // The current version of jQuery being used - jquery: version, - - constructor: jQuery, - - // Start with an empty selector - selector: "", - - // The default length of a jQuery object is 0 - length: 0, - - toArray: function() { - return slice.call( this ); - }, - - // Get the Nth element in the matched element set OR - // Get the whole matched element set as a clean array - get: function( num ) { - return num != null ? - - // Return a 'clean' array - ( num < 0 ? this[ num + this.length ] : this[ num ] ) : - - // Return just the object - slice.call( this ); - }, - - // Take an array of elements and push it onto the stack - // (returning the new matched element set) - pushStack: function( elems ) { - - // Build a new jQuery matched element set - var ret = jQuery.merge( this.constructor(), elems ); - - // Add the old object onto the stack (as a reference) - ret.prevObject = this; - ret.context = this.context; - - // Return the newly-formed element set - return ret; - }, - - // Execute a callback for every element in the matched set. - // (You can seed the arguments with an array of args, but this is - // only used internally.) - each: function( callback, args ) { - return jQuery.each( this, callback, args ); - }, - - map: function( callback ) { - return this.pushStack( jQuery.map(this, function( elem, i ) { - return
[05/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/js/dagre-d3.js -- diff --git a/content/visualizer/js/dagre-d3.js b/content/visualizer/js/dagre-d3.js deleted file mode 100755 index 482ce82..000 --- a/content/visualizer/js/dagre-d3.js +++ /dev/null @@ -1,4560 +0,0 @@ -;(function e(t,n,r){function s(o,u){if(!n[o]){if(!t[o]){var a=typeof require=="function"&if(!u&)return a(o,!0);if(i)return i(o,!0);throw new Error("Cannot find module '"+o+"'")}var f=n[o]={exports:{}};t[o][0].call(f.exports,function(e){var n=t[o][1][e];return s(n?n:e)},f,f.exports,e,t,n,r)}return n[o].exports}var i=typeof require=="function"&for(var o=0;o
[07/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/css/graph.css -- diff --git a/content/visualizer/css/graph.css b/content/visualizer/css/graph.css deleted file mode 100755 index f4365cc..000 --- a/content/visualizer/css/graph.css +++ /dev/null @@ -1,48 +0,0 @@ -/*! - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -g.type-TK > rect { - fill: #00ffd0; -} - -svg { - border: 0px solid #999; - overflow: hidden; -} - -text { - font-weight: 300; - font-family: "Helvetica Neue", Helvetica, Arial, sans-serf; - font-size: 14px; -} - -.node rect { - stroke: #999; - stroke-width: 0px; - fill: #fff; -} - -.edgeLabel rect { - fill: #fff; -} - -.edgePath path { - stroke: #333; - stroke-width: 3px; - fill: none; -} http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/css/nephelefrontend.css -- diff --git a/content/visualizer/css/nephelefrontend.css b/content/visualizer/css/nephelefrontend.css deleted file mode 100755 index 3ca1837..000 --- a/content/visualizer/css/nephelefrontend.css +++ /dev/null @@ -1,198 +0,0 @@ -/*! - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -* -{ - padding: 0px; - margin: 0px; -} - -html, body -{ - height: 100%; - position: relative; - overflow: auto; -} - -h1 -{ - font-family: verdana; - font-size: 36px; - font-weight: normal; - font-decoration: none; - font-style: normal; - font-variant: small-caps; -} - -h1 img { - vertical-align: middle; - margin-right: 50px; -} - -h3 -{ - font-family: verdana; - font-size: 20px; - font-weight: normal; - font-decoration: none; - font-style: italic; - font-variant: normal; - text-align: center; - color: #66; - margin: 10px 10px 20px 10px; -} - -h4 -{ - font-family: Verdana; - font-size: 16px; - font-weight: normal; - font-decoration: none; - font-style: italic; - font-variant: normal; - text-align: left; - color: #00; - margin: 5px 5px 10px 5px; - padding: 2px 2px 5px 2px; - border-bottom: 1px dashed #44; -} - -input -{ - font-family: verdana, Sans-Serif; - font-size: 13px; - - border: solid 1.5px #33; - background-color: #ee; - color: #33; - - padding: 5px; - margin: 5px; -} - -.fadedPropertiesText -{ - font-family: verdana; - font-size: 18px; - color: #AA; - font-style: italic; -} - - -.mainHeading -{ - border: 1px solid #262A37; - margin: 5px; - text-align: left; - background-color: white; - background-repeat: no-repeat; - background-position: right; - height: 100px; - overflow: hidden; -} - - -.boxed -{ - border: 1px solid #262A37; - margin: 5px; -} - -.spaced -{ - margin: 5px; -} - -.footer -{ - position: absolute; - left: 0px; - right: 0px; - bottom: 0px; -} - -.error_text -{ - font-family: verdana; - font-size: 14px; - font-weight: normal; - font-decoration: none; - font-style: italic; - font-variant: normal; - text-align: center; - color: #DF0101; -} - -.translucent -{ - -moz-opacity:0 ; -
[01/78] [abbrv] flink-web git commit: Revert "Rebuild website"
Repository: flink-web Updated Branches: refs/heads/asf-site 4bb81d3f1 -> 9ec0a879f http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/js/program.js -- diff --git a/content/visualizer/js/program.js b/content/visualizer/js/program.js deleted file mode 100755 index 683185d..000 --- a/content/visualizer/js/program.js +++ /dev/null @@ -1,233 +0,0 @@ -/*! - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -var maxColumnWidth = 200; -var minColumnWidth = 100; - -// global variable for the currently requested plan -var pactPlanRequested = 0; - -/* - * This function toggels the child checkbox on and of, depending on the parent's state - */ -function toggleShowPlanBox(box) -{ - var child = $('#suspendJobDuringPlanCheck'); - - if (box.is(':checked')) { -child.attr('disabled', false); - } - else { -child.attr('disabled', true); - } -} - -/* - * Shows an error message below the upload field. - */ -function showUploadError(message) -{ - $('#upload_error_text').fadeOut("fast", function () { $('#upload_error_text')[0].innerHTML = "" + message; - $('#upload_error_text').fadeIn("slow"); } ); -} - -/* - * Checks the selected file and triggers an upload, if all is correct. - */ -function processUpload() -{ - - var filename = $('#upload_file_input').val(); - var len = filename.length; - if (len == 0) { -showUploadError("Please select a file."); - } - else if (len > 4 && filename.substr(len - 4, len) == ".jar") { -$('#upload_form')[0].submit(); - } - else { -showUploadError("Please select a .jar file."); - } -} - -/* - * This function makes sure only one checkbox is selected. - * Upon selection it initializes the drawing of the pact plan. - * Upon deselection, it clears the pact plan. - */ -function toggleCheckboxes(box) -{ - - if (box.is(':checked')) { -$('.jobItemCheckbox').attr('checked', false); -box.attr('checked', true); -var id = box.parentsUntil('.JobListItems').parent().attr('id').substr(4); -var assemblerClass = box.attr('id'); - -$('#mainCanvas').html(''); -pactPlanRequested = id; - -$.ajax({ -type: "GET", -url: "pactPlan", -data: { job: id, assemblerClass: assemblerClass}, -success: function(response) { showPreviewPlan(response); } -}); - } - else { -$('#mainCanvas').html(''); - } -} - -/* - * Function that takes the returned plan and draws it. - */ -function showPreviewPlan(data) -{ - $("#mainCanvas").empty(); -var svgElement = ""; -$("#mainCanvas").append(svgElement); -drawGraph(data.plan, "#svg-main"); -pactPlanRequested = 0; - -//activate zoom buttons -activateZoomButtons(); -// } -} - -/* - * Asynchronously loads the jobs and creates a list of them. - */ -function loadJobList() -{ - $.get("jobs", { action: "list" }, createJobList); -} - -/* - * Triggers an AJAX request to delete a job. - */ -function deleteJob(id) -{ - var name = id.substr(4); - $.get("jobs", { action: "delete", filename: name }, loadJobList); -} - -/* - * Creates and lists the returned jobs. - */ -function createJobList(data) -{ - var markup = ""; - - var lines = data.split("\n"); - for (var i = 0; i < lines.length; i++) - { -if (lines[i] == null || lines[i].length == 0) { - continue; -} - -var date = "unknown date"; -var assemblerClass = "no entry class specified"; - -var tokens = lines[i].split("\t"); -var name = tokens[0]; -if (tokens.length > 1) { - date = tokens[1]; - if (tokens.length > 2) { -assemblerClass = tokens[2]; - } -} - -var entries = assemblerClass.split("#"); -var classes = entries[0].split(","); - -markup += ''; -markup += '' + name + ''; -markup += '' + date + ''; -markup += ''; - -var j = 0; -for (var idx in classes) { - markup += ''; - markup += '' + classes[idx] + ''; -} -markup += ''; - } - - // add the contents - $('#jobsContents').html(markup); - - // register the event handler that triggers the delete when the
[26/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/colors/flink_colors.pptx -- diff --git a/content/img/logo/colors/flink_colors.pptx b/content/img/logo/colors/flink_colors.pptx deleted file mode 100644 index 8ecd0be..000 Binary files a/content/img/logo/colors/flink_colors.pptx and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/100/flink_squirrel_100_black.png -- diff --git a/content/img/logo/png/100/flink_squirrel_100_black.png b/content/img/logo/png/100/flink_squirrel_100_black.png deleted file mode 100755 index 6869f58..000 Binary files a/content/img/logo/png/100/flink_squirrel_100_black.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/100/flink_squirrel_100_color.png -- diff --git a/content/img/logo/png/100/flink_squirrel_100_color.png b/content/img/logo/png/100/flink_squirrel_100_color.png deleted file mode 100755 index c508e1e..000 Binary files a/content/img/logo/png/100/flink_squirrel_100_color.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/100/flink_squirrel_100_white.png -- diff --git a/content/img/logo/png/100/flink_squirrel_100_white.png b/content/img/logo/png/100/flink_squirrel_100_white.png deleted file mode 100755 index 088fb27..000 Binary files a/content/img/logo/png/100/flink_squirrel_100_white.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink1000_black.png -- diff --git a/content/img/logo/png/1000/flink1000_black.png b/content/img/logo/png/1000/flink1000_black.png deleted file mode 100755 index 31af663..000 Binary files a/content/img/logo/png/1000/flink1000_black.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink1000_color_black.png -- diff --git a/content/img/logo/png/1000/flink1000_color_black.png b/content/img/logo/png/1000/flink1000_color_black.png deleted file mode 100755 index 2f4b991..000 Binary files a/content/img/logo/png/1000/flink1000_color_black.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink1000_color_white.png -- diff --git a/content/img/logo/png/1000/flink1000_color_white.png b/content/img/logo/png/1000/flink1000_color_white.png deleted file mode 100755 index cd0b56e..000 Binary files a/content/img/logo/png/1000/flink1000_color_white.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink1000_white.png -- diff --git a/content/img/logo/png/1000/flink1000_white.png b/content/img/logo/png/1000/flink1000_white.png deleted file mode 100755 index 3b7b253..000 Binary files a/content/img/logo/png/1000/flink1000_white.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink_squirrel_1000.png -- diff --git a/content/img/logo/png/1000/flink_squirrel_1000.png b/content/img/logo/png/1000/flink_squirrel_1000.png deleted file mode 100755 index 37c980e..000 Binary files a/content/img/logo/png/1000/flink_squirrel_1000.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink_squirrel_black_1000.png -- diff --git a/content/img/logo/png/1000/flink_squirrel_black_1000.png b/content/img/logo/png/1000/flink_squirrel_black_1000.png deleted file mode 100755 index b68621d..000 Binary files a/content/img/logo/png/1000/flink_squirrel_black_1000.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/1000/flink_squirrel_white_1000.png -- diff --git a/content/img/logo/png/1000/flink_squirrel_white_1000.png b/content/img/logo/png/1000/flink_squirrel_white_1000.png deleted file mode 100755 index 14c7350..000 Binary files a/content/img/logo/png/1000/flink_squirrel_white_1000.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/png/200/flink2_200_black.png
[04/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/js/graphCreator.js -- diff --git a/content/visualizer/js/graphCreator.js b/content/visualizer/js/graphCreator.js deleted file mode 100755 index 86c8c81..000 --- a/content/visualizer/js/graphCreator.js +++ /dev/null @@ -1,445 +0,0 @@ -/*! - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -/* - * Document initialization. - */ -$(document).ready(function () -{ - zoom = d3.behavior.zoom("#svg-main").on("zoom", function() { - var ev = d3.event; - d3.select("#svg-main g") - .attr("transform", "translate(" + ev.translate + ") scale(" + ev.scale + ")"); - }); - zoom.scaleExtent([0.3, 3]) -}); - -//The current JSON file -var JSONData; -//The informations for the iterations -var iterationIds = new Array(); -var iterationGraphs = new Array(); -var iterationWidths = new Array(); -var iterationHeights = new Array(); - -//The zoom element -var zoom; -//informations about the enviroment -var svgWidth; -var svgHeight; - -//Renders and draws the graph -function drawGraph(data, svgID){ - JSONData = data; - - //First step: precompute all iteration graphs - - //find all iterations - iterationNodes = searchForIterationNodes(); - - //add the graphs of iterations and their sizes + Ids to arrays - if (iterationNodes != null) { - for (var i in iterationNodes) { - var itNode = iterationNodes[i]; - iterationIds.push(itNode.id); - var g0 = loadJsonToDagre(itNode); - iterationGraphs.push(g0); - var r = new dagreD3.Renderer(); - var l = dagreD3.layout() -.nodeSep(20) -.rankDir("LR"); - l = r.layout(l).run(g0, d3.select("#svg-main")); - - iterationWidths.push(l._value.width); - iterationHeights.push(l._value.height); - - //Clean svg - $("#svg-main g").empty(); - } - } - - //Continue normal - var g = loadJsonToDagre(data); - var selector = svgID + " g"; - var renderer = new dagreD3.Renderer(); - var layout = dagreD3.layout() - .nodeSep(20) - .rankDir("LR"); - var svgElement = d3.select(selector); - layout = renderer.layout(layout).run(g, svgElement); - - svgHeight = layout._value.height; - svgWidth = layout._value.width; - -var svg = d3.select("#svg-main") - .attr("width", $(document).width() - 15) - .attr("height", $(document).height() - 15 - 110) -// .attr("viewBox", "0 0 "+ ($(document).width() - 150) +" "+($(document).height() - 15 - 110)) - .call(zoom); - - // This should now draw the precomputed graphs in the svgs... . - - for (var i in iterationIds) { - var workset = searchForNode(iterationIds[i]); - renderer = new dagreD3.Renderer(); - layout = dagreD3.layout() -.nodeSep(20) -.rankDir("LR"); - selector = "#svg-"+iterationIds[i]+" g"; - svgElement = d3.select(selector); - layout = renderer.layout(layout).run(iterationGraphs[i], svgElement); - } - - //enable Overlays and register function for overlay-infos - $("a[rel]").overlay({ - onBeforeLoad: function(){ - var id = this.getTrigger().attr("nodeID") - showProperties(id); - } - }); - - } - -//Creates the dagreD3 graph object -//Responsible for adding nodes and edges -function loadJsonToDagre(data){ - - //stores all nodes that are in current graph -> no edges to nodes which are outside of current
[77/78] [abbrv] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/blog/feed.xml -- diff --git a/content/blog/feed.xml b/content/blog/feed.xml new file mode 100644 index 000..066ae2f --- /dev/null +++ b/content/blog/feed.xml @@ -0,0 +1,6041 @@ + +http://www.w3.org/2005/Atom;> + +Flink Blog Feed +Flink Blog +http://flink.apache.org/blog +http://flink.apache.org/blog/feed.xml; rel="self" type="application/rss+xml" /> + + +Apache Flink 1.1.4 Released +pThe Apache Flink community released the next bugfix version of the Apache Flink 1.1 series./p + +pThis release includes major robustness improvements for checkpoint cleanup on failures and consumption of intermediate streams. We highly recommend all users to upgrade to Flink 1.1.4./p + +div class=highlightprecode class=language-xmlspan class=ntlt;dependencygt;/span + span class=ntlt;groupIdgt;/spanorg.apache.flinkspan class=ntlt;/groupIdgt;/span + span class=ntlt;artifactIdgt;/spanflink-javaspan class=ntlt;/artifactIdgt;/span + span class=ntlt;versiongt;/span1.1.4span class=ntlt;/versiongt;/span +span class=ntlt;/dependencygt;/span +span class=ntlt;dependencygt;/span + span class=ntlt;groupIdgt;/spanorg.apache.flinkspan class=ntlt;/groupIdgt;/span + span class=ntlt;artifactIdgt;/spanflink-streaming-java_2.10span class=ntlt;/artifactIdgt;/span + span class=ntlt;versiongt;/span1.1.4span class=ntlt;/versiongt;/span +span class=ntlt;/dependencygt;/span +span class=ntlt;dependencygt;/span + span class=ntlt;groupIdgt;/spanorg.apache.flinkspan class=ntlt;/groupIdgt;/span + span class=ntlt;artifactIdgt;/spanflink-clients_2.10span class=ntlt;/artifactIdgt;/span + span class=ntlt;versiongt;/span1.1.4span class=ntlt;/versiongt;/span +span class=ntlt;/dependencygt;/span/code/pre/div + +pYou can find the binaries on the updated a href=http://flink.apache.org/downloads.htmlDownloads page/a./p + +h2 id=note-for-rocksdb-backend-usersNote for RocksDB Backend Users/h2 + +pWe updated Flinkâs RocksDB dependency version from code4.5.1/code to code4.11.2/code. Between these versions some of RocksDBâs internal configuration defaults changed that would affect the memory footprint of running Flink with RocksDB. Therefore, we manually reset them to the previous defaults. If you want to run with the new Rocks 4.11.2 defaults, you can do this via:/p + +div class=highlightprecode class=language-javaspan class=nRocksDBStateBackend/span span class=nbackend/span span class=o=/span span class=knew/span span class=nfRocksDBStateBackend/spanspan class=o(/spanspan class=squot;...quot;/spanspan class=o);/span +span class=c1// Use the new default options. Otherwise, the default for RocksDB 4.5.1/span +span class=c1// `PredefinedOptions.DEFAULT_ROCKS_4_5_1` will be used./span +span class=nbackend/spanspan class=o./spanspan class=nasetPredefinedOptions/spanspan class=o(/spanspan class=nPredefinedOptions/spanspan class=o./spanspan class=naDEFAULT/spanspan class=o);/span/code/pre/div + +h2 id=release-notes---flink---version-114Release Notes - Flink - Version 1.1.4/h2 + +h3 id=sub-taskSub-task/h3 +ul +li[a href=https://issues.apache.org/jira/browse/FLINK-4510FLINK-4510/a;] - Always create CheckpointCoordinator +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4984FLINK-4984/a;] - Add Cancellation Barriers to BarrierTracker and BarrierBuffer +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4985FLINK-4985/a;] - Report Declined/Canceled Checkpoints to Checkpoint Coordinator +/li +/ul + +h3 id=bugBug/h3 +ul +li[a href=https://issues.apache.org/jira/browse/FLINK-2662FLINK-2662/a;] - CompilerException: quot;Bug: Plan generation for Unions picked a ship strategy between binary plan operators.quot; +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-3680FLINK-3680/a;] - Remove or improve (not set) text in the Job Plan UI +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-3813FLINK-3813/a;] - YARNSessionFIFOITCase.testDetachedMode failed on Travis +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4108FLINK-4108/a;] - NPE in Row.productArity +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4506FLINK-4506/a;] - CsvOutputFormat defaults allowNullValues to false, even though doc and declaration says true +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4581FLINK-4581/a;] - Table API throws quot;No suitable driver found for jdbc:calcitequot; +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4586FLINK-4586/a;] - NumberSequenceIterator and Accumulator threading issue +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4619FLINK-4619/a;] - JobManager does not answer to client when restore from savepoint fails +/li +li[a
[44/78] [abbrv] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/js/dagre-d3.js -- diff --git a/content/visualizer/js/dagre-d3.js b/content/visualizer/js/dagre-d3.js new file mode 100755 index 000..482ce82 --- /dev/null +++ b/content/visualizer/js/dagre-d3.js @@ -0,0 +1,4560 @@ +;(function e(t,n,r){function s(o,u){if(!n[o]){if(!t[o]){var a=typeof require=="function"&if(!u&)return a(o,!0);if(i)return i(o,!0);throw new Error("Cannot find module '"+o+"'")}var f=n[o]={exports:{}};t[o][0].call(f.exports,function(e){var n=t[o][1][e];return s(n?n:e)},f,f.exports,e,t,n,r)}return n[o].exports}var i=typeof require=="function"&for(var o=0;o
[70/78] [abbrv] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/how-to-contribute.html -- diff --git a/content/how-to-contribute.html b/content/how-to-contribute.html new file mode 100644 index 000..87ad4a0 --- /dev/null +++ b/content/how-to-contribute.html @@ -0,0 +1,345 @@ + + + + + + + +Apache Flink: How To Contribute + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + +How To Contribute + + Apache Flink is developed by an open and friendly community. Everybody is cordially welcome to join the community and contribute to Apache Flink. There are several ways to interact with the community and to contribute to Flink including asking questions, filing bug reports, proposing new features, joining discussions on the mailing lists, contributing code or documentation, improving the website, or testing release candidates. + + + + Ask questions! + File a bug report + Propose an improvement or a new feature + Help others and join the discussions + Test a release candidate + Contribute code + Contribute documentation + Improve the website + More ways to contribute⦠+ Submit a Contributor License Agreement + How to become a committer + + + + +Ask questions! + +The Apache Flink community is eager to help and to answer your questions. We have a user mailing list, hang out in an IRC channel, and watch Stack Overflow on the http://stackoverflow.com/questions/tagged/apache-flink;>[apache-flink] tag. + + + +File a bug report + +Please let us know if you experienced a problem with Flink and file a bug report. Open http://issues.apache.org/jira/browse/FLINK;>Flinkâs JIRA and click on the blue Create button at the top. Please give detailed information about the problem you encountered and, if possible, add a description that helps to reproduce the problem. Thank you very much. + + + +Propose an improvement or a new feature + +Our community is constantly looking for feedback to improve Apache Flink. If you have an idea how to improve Flink or have a new feature in mind that would be beneficial for Flink users, please open an issue in http://issues.apache.org/jira/browse/FLINK;>Flinkâs JIRA. The improvement or new feature should be described in appropriate detail and include the scope and its requirements if possible. Detailed information is important for a few reasons: + + + It ensures your requirements are met when the improvement or feature is implemented. + It helps to estimate the effort and to design a solution that addresses your needs. + It allow for constructive discussions that might arise around this issue. + + +Detailed information is also required, if you plan to contribute the improvement or feature you proposed yourself. Please read the Contribute code guide in this case as well. + +We recommend to first reach consensus with the community on whether a new feature is required and how to implement a new feature, before starting with the implementation. Some features might be out of scope of the project, and itâs best to
[17/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/news/2015/03/02/february-2015-in-flink.html -- diff --git a/content/news/2015/03/02/february-2015-in-flink.html b/content/news/2015/03/02/february-2015-in-flink.html deleted file mode 100644 index a32ba35..000 --- a/content/news/2015/03/02/february-2015-in-flink.html +++ /dev/null @@ -1,308 +0,0 @@ - - - - - - - -Apache Flink: February 2015 in the Flink community - - - - -https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> - - - - - - - - -https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -Download Flink - - -Home - - -Introduction to Flink - - -Flink Use Cases - - -Powered by Flink - - -Ecosystem - - -Community Project Info - - -How to Contribute - - -Flink Blog - - - - - - - - - Documentation - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) - http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) - - - - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart - - - - - https://github.com/apache/flink; target="_blank">Flink on GitHub - - - - - - - - - - - - - - - -Project FAQ - - -https://twitter.com/apacheflink; target="_blank">@ApacheFlink - - -Plan Visualizer - - - - - - - - - - - February 2015 in the Flink community - - -02 Mar 2015 - -February might be the shortest month of the year, but this does not -mean that the Flink community has not been busy adding features to the -system and fixing bugs. Hereâs a rundown of the activity in the Flink -community last month. - -0.8.1 release - -Flink 0.8.1 was released. This bugfixing release resolves a total of 22 issues. - -New committer - -https://github.com/mxm;>Max Michels has been voted a committer by the Flink PMC. - -Flink adapter for Apache SAMOA - -http://samoa.incubator.apache.org;>Apache SAMOA (incubating) is a -distributed streaming machine learning (ML) framework with a -programming abstraction for distributed streaming ML algorithms. SAMOA -runs on a variety of backend engines, currently Apache Storm and -Apache S4. A https://github.com/apache/incubator-samoa/pull/11;>pull -request is -available at the SAMOA repository that adds a Flink adapter for SAMOA. - -Easy Flink deployment on Google Compute Cloud - -Flink is now integrated in bdutil, Googleâs open source tool for -creating and configuring (Hadoop) clusters in Google Compute -Engine. Deployment of Flink clusters in now supported starting with -https://groups.google.com/forum/#!topic/gcp-hadoop-announce/uVJ_6y9cGKM;>bdutil -1.2.0. - -Flink on the Web - -A new blog post on http://flink.apache.org/news/2015/02/09/streaming-example.html;>Flink -Streaming -was published at the blog. Flink was mentioned in several articles on -the web. Here are some examples: - - - -http://dataconomy.com/how-flink-became-an-apache-top-level-project/;>How Flink became an Apache Top-Level Project - - -https://www.linkedin.com/pulse/stale-synchronous-parallelism-new-frontier-apache-flink-nam-luc-tran?utm_content=buffer461afutm_medium=socialutm_source=linkedin.comutm_campaign=buffer;>Stale Synchronous Parallelism: The new frontier for Apache Flink? - - -http://www.hadoopsphere.com/2015/02/distributed-data-processing-with-apache.html;>Distributed data processing with Apache Flink - - -http://www.hadoopsphere.com/2015/02/ciao-latency-hallo-speed.html;>Ciao latency, hello speed - - - -In the Flink master - -The following features have been now merged in Flinkâs master repository. - -Gelly - -Gelly, Flinkâs Graph API allows users to manipulate graph-shaped data -directly. Hereâs for example a calculation of shortest paths in a -graph: - -GraphLong, Double, Double graph = Graph.fromDataSet(vertices, edges, env); -
[06/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/js/d3.js -- diff --git a/content/visualizer/js/d3.js b/content/visualizer/js/d3.js deleted file mode 100755 index 2d09bc9..000 --- a/content/visualizer/js/d3.js +++ /dev/null @@ -1,9270 +0,0 @@ -/** - * D3 renderer from http://d3js.org/ - * Copyright (c) 2010-2014, Michael Bostock - * - * Under BSD License (as stated under http://d3js.org/) - */ - -!function() { - var d3 = { -version: "3.4.6" - }; - if (!Date.now) Date.now = function() { -return +new Date(); - }; - var d3_arraySlice = [].slice, d3_array = function(list) { -return d3_arraySlice.call(list); - }; - var d3_document = document, d3_documentElement = d3_document.documentElement, d3_window = window; - try { -d3_array(d3_documentElement.childNodes)[0].nodeType; - } catch (e) { -d3_array = function(list) { - var i = list.length, array = new Array(i); - while (i--) array[i] = list[i]; - return array; -}; - } - try { -d3_document.createElement("div").style.setProperty("opacity", 0, ""); - } catch (error) { -var d3_element_prototype = d3_window.Element.prototype, d3_element_setAttribute = d3_element_prototype.setAttribute, d3_element_setAttributeNS = d3_element_prototype.setAttributeNS, d3_style_prototype = d3_window.CSSStyleDeclaration.prototype, d3_style_setProperty = d3_style_prototype.setProperty; -d3_element_prototype.setAttribute = function(name, value) { - d3_element_setAttribute.call(this, name, value + ""); -}; -d3_element_prototype.setAttributeNS = function(space, local, value) { - d3_element_setAttributeNS.call(this, space, local, value + ""); -}; -d3_style_prototype.setProperty = function(name, value, priority) { - d3_style_setProperty.call(this, name, value + "", priority); -}; - } - d3.ascending = d3_ascending; - function d3_ascending(a, b) { -return a < b ? -1 : a > b ? 1 : a >= b ? 0 : NaN; - } - d3.descending = function(a, b) { -return b < a ? -1 : b > a ? 1 : b >= a ? 0 : NaN; - }; - d3.min = function(array, f) { -var i = -1, n = array.length, a, b; -if (arguments.length === 1) { - while (++i < n && !((a = array[i]) != null && a <= a)) a = undefined; - while (++i < n) if ((b = array[i]) != null && a > b) a = b; -} else { - while (++i < n && !((a = f.call(array, array[i], i)) != null && a <= a)) a = undefined; - while (++i < n) if ((b = f.call(array, array[i], i)) != null && a > b) a = b; -} -return a; - }; - d3.max = function(array, f) { -var i = -1, n = array.length, a, b; -if (arguments.length === 1) { - while (++i < n && !((a = array[i]) != null && a <= a)) a = undefined; - while (++i < n) if ((b = array[i]) != null && b > a) a = b; -} else { - while (++i < n && !((a = f.call(array, array[i], i)) != null && a <= a)) a = undefined; - while (++i < n) if ((b = f.call(array, array[i], i)) != null && b > a) a = b; -} -return a; - }; - d3.extent = function(array, f) { -var i = -1, n = array.length, a, b, c; -if (arguments.length === 1) { - while (++i < n && !((a = c = array[i]) != null && a <= a)) a = c = undefined; - while (++i < n) if ((b = array[i]) != null) { -if (a > b) a = b; -if (c < b) c = b; - } -} else { - while (++i < n && !((a = c = f.call(array, array[i], i)) != null && a <= a)) a = undefined; - while (++i < n) if ((b = f.call(array, array[i], i)) != null) { -if (a > b) a = b; -if (c < b) c = b; - } -} -return [ a, c ]; - }; - d3.sum = function(array, f) { -var s = 0, n = array.length, a, i = -1; -if (arguments.length === 1) { - while (++i < n) if (!isNaN(a = +array[i])) s += a; -} else { - while (++i < n) if (!isNaN(a = +f.call(array, array[i], i))) s += a; -} -return s; - }; - function d3_number(x) { -return x != null && !isNaN(x); - } - d3.mean = function(array, f) { -var s = 0, n = array.length, a, i = -1, j = n; -if (arguments.length === 1) { - while (++i < n) if (d3_number(a = array[i])) s += a; else --j; -} else { - while (++i < n) if (d3_number(a = f.call(array, array[i], i))) s += a; else --j; -} -return j ? s / j : undefined; - }; - d3.quantile = function(values, p) { -var H = (values.length - 1) * p + 1, h = Math.floor(H), v = +values[h - 1], e = H - h; -return e ? v + e * (values[h] - v) : v; - }; - d3.median = function(array, f) { -if (arguments.length > 1) array = array.map(f); -array = array.filter(d3_number); -return array.length ? d3.quantile(array.sort(d3_ascending), .5) : undefined; - }; - function d3_bisector(compare) { -return { - left: function(a, x, lo, hi) { -if (arguments.length < 3) lo = 0; -if (arguments.length < 4) hi = a.length; -
[32/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/faq.html -- diff --git a/content/faq.html b/content/faq.html deleted file mode 100644 index 7653ff8..000 --- a/content/faq.html +++ /dev/null @@ -1,683 +0,0 @@ - - - - - - - -Apache Flink: Frequently Asked Questions (FAQ) - - - - -https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> - - - - - - - - -https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -Download Flink - - -Home - - -Introduction to Flink - - -Flink Use Cases - - -Powered by Flink - - -Ecosystem - - -Community Project Info - - -How to Contribute - - -Flink Blog - - - - - - - - - Documentation - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) - http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) - - - - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart - - - - - https://github.com/apache/flink; target="_blank">Flink on GitHub - - - - - - - - - - - - - - - -Project FAQ - - -https://twitter.com/apacheflink; target="_blank">@ApacheFlink - - -Plan Visualizer - - - - - - - - - -Frequently Asked Questions (FAQ) - - - -The following questions are frequently asked with regard to the Flink project in general. If you have further questions, make sure to consult the documentation or ask the community. - - - - General - Is Flink a Hadoop Project? - Do I have to install Apache Hadoop to use Flink? - - - Usage - How do I assess the progress of a Flink program? - How can I figure out why a program failed? - How do I debug Flink programs? - What is the parallelism? How do I set it? - - - Errors - Why am I getting a âNonSerializableExceptionâ ? - In Scala API, I get an error about implicit values and evidence parameters - I get an error message saying that not enough buffers are available. How do I fix this? - My job fails early with a java.io.EOFException. What could be the cause? - My job fails with various exceptions from the HDFS/Hadoop code. What can I do? - In Eclipse, I get compilation errors in the Scala projects - My program does not compute the correct result. Why are my custom key types - I get a java.lang.InstantiationException for my data type, what is wrong? - I canât stop Flink with the provided stop-scripts. What can I do? - I got an OutOfMemoryException. What can I do? - Why do the TaskManager log files become so huge? - The slot allocated for my task manager has been released. What should I do? - - - YARN Deployment - The YARN session runs only for a few seconds - My YARN containers are killed because they use too much memory - The YARN session crashes with a HDFS permission exception during startup - My job is not reacting to a job cancellation? - - - Features - What kind of fault-tolerance does Flink provide? - Are Hadoop-like utilities, such as Counters and the DistributedCache supported? - - - - - - -General - -Is Flink a Hadoop Project? - -Flink is a data processing system and an alternative to Hadoopâs -MapReduce component. It comes with its own runtime rather than building on top -of MapReduce. As such, it can work completely independently of the Hadoop -ecosystem. However, Flink can also access Hadoopâs distributed file -system (HDFS) to read and write data, and Hadoopâs next-generation resource -manager (YARN) to provision cluster resources. Since most Flink users are -using Hadoop HDFS to store their data, Flink already ships the required libraries to -access HDFS. - -Do I have to install Apache Hadoop to use Flink? - -No. Flink can run without a Hadoop installation. However, a
[14/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/news/2015/09/16/off-heap-memory.html -- diff --git a/content/news/2015/09/16/off-heap-memory.html b/content/news/2015/09/16/off-heap-memory.html deleted file mode 100644 index 60d7bda..000 --- a/content/news/2015/09/16/off-heap-memory.html +++ /dev/null @@ -1,1083 +0,0 @@ - - - - - - - -Apache Flink: Off-heap Memory in Apache Flink and the curious JIT compiler - - - - -https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> - - - - - - - - -https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -Download Flink - - -Home - - -Introduction to Flink - - -Flink Use Cases - - -Powered by Flink - - -Ecosystem - - -Community Project Info - - -How to Contribute - - -Flink Blog - - - - - - - - - Documentation - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) - http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) - - - - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart - - - - - https://github.com/apache/flink; target="_blank">Flink on GitHub - - - - - - - - - - - - - - - -Project FAQ - - -https://twitter.com/apacheflink; target="_blank">@ApacheFlink - - -Plan Visualizer - - - - - - - - - - - Off-heap Memory in Apache Flink and the curious JIT compiler - - -16 Sep 2015 by Stephan Ewen (https://twitter.com/stephanewen;>@stephanewen) - -Running data-intensive code in the JVM and making it well-behaved is tricky. Systems that put billions of data objects naively onto the JVM heap face unpredictable OutOfMemoryErrors and Garbage Collection stalls. Of course, you still want to to keep your data in memory as much as possible, for speed and responsiveness of the processing applications. In that context, âoff-heapâ has become almost something like a magic word to solve these problems. - -In this blog post, we will look at how Flink exploits off-heap memory. The feature is part of the upcoming release, but you can try it out with the latest nightly builds. We will also give a few interesting insights into the behavior for Javaâs JIT compiler for highly optimized methods and loops. - -Recap: Memory Management in Flink - -To understand Flinkâs approach to off-heap memory, we need to recap Flinkâs approach to custom managed memory. We have written an earlier blog post about how Flink manages JVM memory itself - -As a summary, the core part is that Flink implements its algorithms not against Java objects, arrays, or lists, but actually against a data structure similar to java.nio.ByteBuffer. Flink uses its own specialized version, called https://github.com/apache/flink/blob/release-0.9.1-rc1/flink-core/src/main/java/org/apache/flink/core/memory/MemorySegment.java;>MemorySegment on which algorithms put and get at specific positions ints, longs, byte arrays, etc, and compare and copy memory. The memory segments are held and distributed by a central component (called MemoryManager) from which algorithms request segments according to their calculated memory budgets. - -Donât believe that this can be fast? Have a look at the benchmarks in the earlier blogpost, which show that it is actually often much faster than working on objects, due to better control over data layout (cache efficiency, data size), and reducing the pressure on Javaâs Garbage Collector. - -This form of memory management has been in Flink for a long time. Anecdotally, the first public demo of Flinkâs predecessor project Stratosphere, at the VLDB conference in 2010, was running its programs with custom managed memory (although I believe few attendees were aware of that). - -Why actually bother with off-heap memory? - -Given that Flink has a sophisticated level of managing on-heap
[08/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/css/bootstrap.css -- diff --git a/content/visualizer/css/bootstrap.css b/content/visualizer/css/bootstrap.css deleted file mode 100755 index fc41022..000 --- a/content/visualizer/css/bootstrap.css +++ /dev/null @@ -1,5785 +0,0 @@ -/*! - * Bootstrap v3.1.1 (http://getbootstrap.com) - * Copyright 2011-2014 Twitter, Inc. - * Licensed under MIT (https://github.com/twbs/bootstrap/blob/master/LICENSE) - */ - -/*! normalize.css v3.0.0 | MIT License | git.io/normalize */ -html { - font-family: sans-serif; - -webkit-text-size-adjust: 100%; - -ms-text-size-adjust: 100%; -} -body { - margin: 0; -} -article, -aside, -details, -figcaption, -figure, -footer, -header, -hgroup, -main, -nav, -section, -summary { - display: block; -} -audio, -canvas, -progress, -video { - display: inline-block; - vertical-align: baseline; -} -audio:not([controls]) { - display: none; - height: 0; -} -[hidden], -template { - display: none; -} -a { - background: transparent; -} -a:active, -a:hover { - outline: 0; -} -abbr[title] { - border-bottom: 1px dotted; -} -b, -strong { - font-weight: bold; -} -dfn { - font-style: italic; -} -h1 { - margin: .67em 0; - font-size: 2em; -} -mark { - color: #000; - background: #ff0; -} -small { - font-size: 80%; -} -sub, -sup { - position: relative; - font-size: 75%; - line-height: 0; - vertical-align: baseline; -} -sup { - top: -.5em; -} -sub { - bottom: -.25em; -} -img { - border: 0; -} -svg:not(:root) { - overflow: hidden; -} -figure { - margin: 1em 40px; -} -hr { - height: 0; - -moz-box-sizing: content-box; - box-sizing: content-box; -} -pre { - overflow: auto; -} -code, -kbd, -pre, -samp { - font-family: monospace, monospace; - font-size: 1em; -} -button, -input, -optgroup, -select, -textarea { - margin: 0; - font: inherit; - color: inherit; -} -button { - overflow: visible; -} -button, -select { - text-transform: none; -} -button, -html input[type="button"], -input[type="reset"], -input[type="submit"] { - -webkit-appearance: button; - cursor: pointer; -} -button[disabled], -html input[disabled] { - cursor: default; -} -button::-moz-focus-inner, -input::-moz-focus-inner { - padding: 0; - border: 0; -} -input { - line-height: normal; -} -input[type="checkbox"], -input[type="radio"] { - box-sizing: border-box; - padding: 0; -} -input[type="number"]::-webkit-inner-spin-button, -input[type="number"]::-webkit-outer-spin-button { - height: auto; -} -input[type="search"] { - -webkit-box-sizing: content-box; - -moz-box-sizing: content-box; - box-sizing: content-box; - -webkit-appearance: textfield; -} -input[type="search"]::-webkit-search-cancel-button, -input[type="search"]::-webkit-search-decoration { - -webkit-appearance: none; -} -fieldset { - padding: .35em .625em .75em; - margin: 0 2px; - border: 1px solid #c0c0c0; -} -legend { - padding: 0; - border: 0; -} -textarea { - overflow: auto; -} -optgroup { - font-weight: bold; -} -table { - border-spacing: 0; - border-collapse: collapse; -} -td, -th { - padding: 0; -} -@media print { - * { -color: #000 !important; -text-shadow: none !important; -background: transparent !important; -box-shadow: none !important; - } - a, - a:visited { -text-decoration: underline; - } - a[href]:after { -content: " (" attr(href) ")"; - } - abbr[title]:after { -content: " (" attr(title) ")"; - } - a[href^="javascript:"]:after, - a[href^="#"]:after { -content: ""; - } - pre, - blockquote { -border: 1px solid #999; - -page-break-inside: avoid; - } - thead { -display: table-header-group; - } - tr, - img { -page-break-inside: avoid; - } - img { -max-width: 100% !important; - } - p, - h2, - h3 { -orphans: 3; -widows: 3; - } - h2, - h3 { -page-break-after: avoid; - } - select { -background: #fff !important; - } - .navbar { -display: none; - } - .table td, - .table th { -background-color: #fff !important; - } - .btn > .caret, - .dropup > .btn > .caret { -border-top-color: #000 !important; - } - .label { -border: 1px solid #000; - } - .table { -border-collapse: collapse !important; - } - .table-bordered th, - .table-bordered td { -border: 1px solid #ddd !important; - } -} -* { - -webkit-box-sizing: border-box; - -moz-box-sizing: border-box; - box-sizing: border-box; -} -*:before, -*:after { - -webkit-box-sizing: border-box; - -moz-box-sizing: border-box; - box-sizing: border-box; -} -html { - font-size: 62.5%; - - -webkit-tap-highlight-color: rgba(0, 0, 0, 0); -} -body { - font-family: "Helvetica Neue", Helvetica, Arial, sans-serif; - font-size: 14px; - line-height: 1.42857143; - color: #333; - background-color: #fff; -} -input, -button, -select, -textarea { - font-family: inherit; -
[47/78] [abbrv] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/css/bootstrap.css -- diff --git a/content/visualizer/css/bootstrap.css b/content/visualizer/css/bootstrap.css new file mode 100755 index 000..fc41022 --- /dev/null +++ b/content/visualizer/css/bootstrap.css @@ -0,0 +1,5785 @@ +/*! + * Bootstrap v3.1.1 (http://getbootstrap.com) + * Copyright 2011-2014 Twitter, Inc. + * Licensed under MIT (https://github.com/twbs/bootstrap/blob/master/LICENSE) + */ + +/*! normalize.css v3.0.0 | MIT License | git.io/normalize */ +html { + font-family: sans-serif; + -webkit-text-size-adjust: 100%; + -ms-text-size-adjust: 100%; +} +body { + margin: 0; +} +article, +aside, +details, +figcaption, +figure, +footer, +header, +hgroup, +main, +nav, +section, +summary { + display: block; +} +audio, +canvas, +progress, +video { + display: inline-block; + vertical-align: baseline; +} +audio:not([controls]) { + display: none; + height: 0; +} +[hidden], +template { + display: none; +} +a { + background: transparent; +} +a:active, +a:hover { + outline: 0; +} +abbr[title] { + border-bottom: 1px dotted; +} +b, +strong { + font-weight: bold; +} +dfn { + font-style: italic; +} +h1 { + margin: .67em 0; + font-size: 2em; +} +mark { + color: #000; + background: #ff0; +} +small { + font-size: 80%; +} +sub, +sup { + position: relative; + font-size: 75%; + line-height: 0; + vertical-align: baseline; +} +sup { + top: -.5em; +} +sub { + bottom: -.25em; +} +img { + border: 0; +} +svg:not(:root) { + overflow: hidden; +} +figure { + margin: 1em 40px; +} +hr { + height: 0; + -moz-box-sizing: content-box; + box-sizing: content-box; +} +pre { + overflow: auto; +} +code, +kbd, +pre, +samp { + font-family: monospace, monospace; + font-size: 1em; +} +button, +input, +optgroup, +select, +textarea { + margin: 0; + font: inherit; + color: inherit; +} +button { + overflow: visible; +} +button, +select { + text-transform: none; +} +button, +html input[type="button"], +input[type="reset"], +input[type="submit"] { + -webkit-appearance: button; + cursor: pointer; +} +button[disabled], +html input[disabled] { + cursor: default; +} +button::-moz-focus-inner, +input::-moz-focus-inner { + padding: 0; + border: 0; +} +input { + line-height: normal; +} +input[type="checkbox"], +input[type="radio"] { + box-sizing: border-box; + padding: 0; +} +input[type="number"]::-webkit-inner-spin-button, +input[type="number"]::-webkit-outer-spin-button { + height: auto; +} +input[type="search"] { + -webkit-box-sizing: content-box; + -moz-box-sizing: content-box; + box-sizing: content-box; + -webkit-appearance: textfield; +} +input[type="search"]::-webkit-search-cancel-button, +input[type="search"]::-webkit-search-decoration { + -webkit-appearance: none; +} +fieldset { + padding: .35em .625em .75em; + margin: 0 2px; + border: 1px solid #c0c0c0; +} +legend { + padding: 0; + border: 0; +} +textarea { + overflow: auto; +} +optgroup { + font-weight: bold; +} +table { + border-spacing: 0; + border-collapse: collapse; +} +td, +th { + padding: 0; +} +@media print { + * { +color: #000 !important; +text-shadow: none !important; +background: transparent !important; +box-shadow: none !important; + } + a, + a:visited { +text-decoration: underline; + } + a[href]:after { +content: " (" attr(href) ")"; + } + abbr[title]:after { +content: " (" attr(title) ")"; + } + a[href^="javascript:"]:after, + a[href^="#"]:after { +content: ""; + } + pre, + blockquote { +border: 1px solid #999; + +page-break-inside: avoid; + } + thead { +display: table-header-group; + } + tr, + img { +page-break-inside: avoid; + } + img { +max-width: 100% !important; + } + p, + h2, + h3 { +orphans: 3; +widows: 3; + } + h2, + h3 { +page-break-after: avoid; + } + select { +background: #fff !important; + } + .navbar { +display: none; + } + .table td, + .table th { +background-color: #fff !important; + } + .btn > .caret, + .dropup > .btn > .caret { +border-top-color: #000 !important; + } + .label { +border: 1px solid #000; + } + .table { +border-collapse: collapse !important; + } + .table-bordered th, + .table-bordered td { +border: 1px solid #ddd !important; + } +} +* { + -webkit-box-sizing: border-box; + -moz-box-sizing: border-box; + box-sizing: border-box; +} +*:before, +*:after { + -webkit-box-sizing: border-box; + -moz-box-sizing: border-box; + box-sizing: border-box; +} +html { + font-size: 62.5%; + + -webkit-tap-highlight-color: rgba(0, 0, 0, 0); +} +body { + font-family: "Helvetica Neue", Helvetica, Arial, sans-serif; + font-size: 14px; + line-height: 1.42857143; + color: #333; + background-color: #fff; +} +input, +button, +select, +textarea { + font-family: inherit; +
[31/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/how-to-contribute.html -- diff --git a/content/how-to-contribute.html b/content/how-to-contribute.html deleted file mode 100644 index 87ad4a0..000 --- a/content/how-to-contribute.html +++ /dev/null @@ -1,345 +0,0 @@ - - - - - - - -Apache Flink: How To Contribute - - - - -https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> - - - - - - - - -https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -Download Flink - - -Home - - -Introduction to Flink - - -Flink Use Cases - - -Powered by Flink - - -Ecosystem - - -Community Project Info - - -How to Contribute - - -Flink Blog - - - - - - - - - Documentation - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) - http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) - - - - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart - - - - - https://github.com/apache/flink; target="_blank">Flink on GitHub - - - - - - - - - - - - - - - -Project FAQ - - -https://twitter.com/apacheflink; target="_blank">@ApacheFlink - - -Plan Visualizer - - - - - - - - - -How To Contribute - - Apache Flink is developed by an open and friendly community. Everybody is cordially welcome to join the community and contribute to Apache Flink. There are several ways to interact with the community and to contribute to Flink including asking questions, filing bug reports, proposing new features, joining discussions on the mailing lists, contributing code or documentation, improving the website, or testing release candidates. - - - - Ask questions! - File a bug report - Propose an improvement or a new feature - Help others and join the discussions - Test a release candidate - Contribute code - Contribute documentation - Improve the website - More ways to contribute⦠- Submit a Contributor License Agreement - How to become a committer - - - - -Ask questions! - -The Apache Flink community is eager to help and to answer your questions. We have a user mailing list, hang out in an IRC channel, and watch Stack Overflow on the http://stackoverflow.com/questions/tagged/apache-flink;>[apache-flink] tag. - - - -File a bug report - -Please let us know if you experienced a problem with Flink and file a bug report. Open http://issues.apache.org/jira/browse/FLINK;>Flinkâs JIRA and click on the blue Create button at the top. Please give detailed information about the problem you encountered and, if possible, add a description that helps to reproduce the problem. Thank you very much. - - - -Propose an improvement or a new feature - -Our community is constantly looking for feedback to improve Apache Flink. If you have an idea how to improve Flink or have a new feature in mind that would be beneficial for Flink users, please open an issue in http://issues.apache.org/jira/browse/FLINK;>Flinkâs JIRA. The improvement or new feature should be described in appropriate detail and include the scope and its requirements if possible. Detailed information is important for a few reasons: - - - It ensures your requirements are met when the improvement or feature is implemented. - It helps to estimate the effort and to design a solution that addresses your needs. - It allow for constructive discussions that might arise around this issue. - - -Detailed information is also required, if you plan to contribute the improvement or feature you proposed yourself. Please read the Contribute code guide in this case as well. - -We recommend to first reach consensus with the community on whether a new feature is required and how to implement a new feature, before starting with the implementation. Some features might be out of scope of the project, and itâs best to
[34/78] [abbrv] flink-web git commit: Revert "Rebuild website"
an Gates -PMC, Committer -gates - - -https://avatars0.githubusercontent.com/u/2388347?s=50; /> -Fabian Hueske -PMC, Committer -fhueske - - -https://avatars3.githubusercontent.com/u/498957?v=3s=50; /> -Vasia Kalavri -PMC, Committer -vasia - - /tr - -https://avatars0.githubusercontent.com/u/68551?s=50; /> -Aljoscha Krettek -PMC, Committer -aljoscha - - -https://avatars2.githubusercontent.com/u/2550549?s=50; /> -Andra Lungu -Committer -andra - - -https://avatars0.githubusercontent.com/u/89049?s=50; /> -Robert Metzger -PMC, Committer -rmetzger - - -https://avatars2.githubusercontent.com/u/837221?s=50; /> -Maximilian Michels -PMC, Committer -mxm - - -https://avatars2.githubusercontent.com/u/1941681?s=50; /> -Chiwan Park -Committer -chiwanpark - - -https://avatars1.githubusercontent.com/u/5756858?s=50; /> -Till Rohrmann -PMC, Committer -trohrmann - - -https://avatars0.githubusercontent.com/u/105434?s=50; /> -Henry Saputra -PMC, Committer -hsaputra - - -https://avatars0.githubusercontent.com/u/8959638?s=50; /> -Matthias J. Sax -Committer -mjsax - - -https://avatars1.githubusercontent.com/u/409707?s=50; /> -Sebastian Schelter -PMC, Committer -ssc - - -https://avatars2.githubusercontent.com/u/1925554?s=50; /> -Kostas Tzoumas -PMC, Committer -ktzoumas - - - -Timo Walther -PMC, Committer -twalthr - - -https://avatars1.githubusercontent.com/u/1826769?s=50; /> -Daniel Warneke -PMC, Committer -warneke - - -https://avatars1.githubusercontent.com/u/4425616?s=50; /> -ChengXiang Li -Committer -chengxiang - - -https://avatars1.githubusercontent.com/u/569655?s=50; /> -Greg Hogan -Committer -greg - - -https://avatars1.githubusercontent.com/u/5284370?s=50; /> -Tzu-Li (Gordon) Tai -Committer -tzulitai - - - -You can reach committers directly at apache-id@apache.org. A list of all contributors can be found https://cwiki.apache.org/confluence/display/FLINK/List+of+contributors;>here. - -Former mentors - -The following people were very kind to mentor the project while in incubation. - - - - -Name -Role -Apache ID - - - -Ashutosh Chauhan -Former PPMC, Mentor -hashutosh - - - -Ted Dunning -Former PPMC, Mentor -tdunning - - - -Alan Gates -Former PPMC, Mentor -gates - - /tr - - -Owen O'Malley -Former PPMC, Mentor -omalley - - - -Sean Owen -Former PPMC, Mentor -srowen - - - -Henry Saputra -Former PPMC, Mentor -hsaputra - - - -Slides - -Note: Keep in mind that code examples on slides have a chance of being incomplete or outdated. Always refer to the http://ci.apache.org/projects/flink/flink-docs-release-1.2;>latest documentation for an up to date reference. - -2016 - - - Stefan Richter: A look at Apache Flink 1.2 and beyond Apache Flink Meetup Berlin, November 2016: http://www.slideshare.net/StefanRichter10/a-look-at-flink-12;>SlideShare - Robert Metzger: Apache Flink Community Updates November 2016 Apache Flink Meetup Berlin, November 2016: http://www.slideshare.net/robertmetzger1/apache-flink-community-updates-november-2016-berlin-meetup;>SlideShare - Aljoscha Krettek: Apache Flink for IoT: How Event-Time Processing Enables Easy and Accurate Analytics Big Data Spain, Madrid November 2016: http://www.slideshare.net/dataArtisans/aljoscha-krettek-apache-flink-for-iot-how-eventtime-processing-enables-easy-and-accurate-analytics;>SlideShare - Stephan Ewen: Keynote -The maturing data streaming ecosystem and Apache Flinkâs accelerated growth Apache Big Data Europe 2016, Seville November 2016: http://www.slideshare.net/dataArtisans/keynote-stephan-ewen-stream-processing-as-a-foundational-paradigm-and-apache-flinks-approach-to-it;>SlideShare - Kostas Tzoumas: Stream Processing with Apache Flink® Apache Flink London Meetup, November 2016: http://www.slideshare.net/dataArtisans/kostas-tzoumas-stream-processing-with-apache-flink;>SlideShare - Kostas Tzoumas: Apache Flink®: State of the Union and Whatâs Next Strata + Hadoop World New York, September 2016: http://www.slideshare.net/dataArtisans/kostas-tzoumas-apache-flink-state-of-the-union-and-whats-next;>SlideShare - Kostas Tzoumas Stephan Ewen: Keynote -The maturing data streaming ecosystem and Apache Flinkâs accelerated growth Flink Forward, Berlin September 2016: http://www.slideshare.net/FlinkForward/kostas-tzoumasstpehan-ewen-keynote-the-maturing-data-streaming-ecosystem-and-apache-flinks-accelerated-growth;>SlideShare - Robert Metzger: Connecting Apache Flink to the World - Reviewing the streaming
[12/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/news/2016/03/08/release-1.0.0.html -- diff --git a/content/news/2016/03/08/release-1.0.0.html b/content/news/2016/03/08/release-1.0.0.html deleted file mode 100644 index 13313d8..000 --- a/content/news/2016/03/08/release-1.0.0.html +++ /dev/null @@ -1,319 +0,0 @@ - - - - - - - -Apache Flink: Announcing Apache Flink 1.0.0 - - - - -https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> - - - - - - - - -https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -Download Flink - - -Home - - -Introduction to Flink - - -Flink Use Cases - - -Powered by Flink - - -Ecosystem - - -Community Project Info - - -How to Contribute - - -Flink Blog - - - - - - - - - Documentation - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) - http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) - - - - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart - - - - - https://github.com/apache/flink; target="_blank">Flink on GitHub - - - - - - - - - - - - - - - -Project FAQ - - -https://twitter.com/apacheflink; target="_blank">@ApacheFlink - - -Plan Visualizer - - - - - - - - - - - Announcing Apache Flink 1.0.0 - - -08 Mar 2016 - -The Apache Flink community is pleased to announce the availability of the 1.0.0 release. The community put significant effort into improving and extending Apache Flink since the last release, focusing on improving the experience of writing and executing data stream processing pipelines in production. - - - - - -Flink version 1.0.0 marks the beginning of the 1.X.X series of releases, which will maintain backwards compatibility with 1.0.0. This means that applications written against stable APIs of Flink 1.0.0 will compile and run with all Flink versions in the 1. series. This is the first time we are formally guaranteeing compatibility in Flinkâs history, and we therefore see this release as a major milestone of the project, perhaps the most important since graduation as a top-level project. - -Apart from backwards compatibility, Flink 1.0.0 brings a variety of new user-facing features, as well as tons of bug fixes. About 64 contributors provided bug fixes, improvements, and new features such that in total more than 450 JIRA issues could be resolved. - -We encourage everyone to http://flink.apache.org/downloads.html;>download the release and https://ci.apache.org/projects/flink/flink-docs-release-1.0/;>check out the documentation. Feedback through the Flink http://flink.apache.org/community.html#mailing-lists;>mailing lists is, as always, very welcome! - -Interface stability annotations - -Flink 1.0.0 introduces interface stability annotations for API classes and methods. Interfaces defined as @Public are guaranteed to remain stable across all releases of the 1.x series. The @PublicEvolving annotation marks API features that may be subject to change in future versions. - -Flinkâs stability annotations will help users to implement applications that compile and execute unchanged against future versions of Flink 1.x. This greatly reduces the complexity for users when upgrading to a newer Flink release. - -Out-of-core state support - -Flink 1.0.0 adds a new state backend that uses RocksDB to store state (both windows and user-defined key-value state). http://rocksdb.org/;>RocksDB is an embedded key/value store database, originally developed by Facebook. -When using this backend, active state in streaming programs can grow well beyond memory. The RocksDB files are stored in a distributed file system such as HDFS or S3 for backups. - -Savepoints and version upgrades - -Savepoints are checkpoints of the state of a running streaming job that can be manually
[56/78] [abbrv] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/03/02/february-2015-in-flink.html -- diff --git a/content/news/2015/03/02/february-2015-in-flink.html b/content/news/2015/03/02/february-2015-in-flink.html new file mode 100644 index 000..1e3a09b --- /dev/null +++ b/content/news/2015/03/02/february-2015-in-flink.html @@ -0,0 +1,308 @@ + + + + + + + +Apache Flink: February 2015 in the Flink community + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + + + February 2015 in the Flink community + + +02 Mar 2015 + +February might be the shortest month of the year, but this does not +mean that the Flink community has not been busy adding features to the +system and fixing bugs. Hereâs a rundown of the activity in the Flink +community last month. + +0.8.1 release + +Flink 0.8.1 was released. This bugfixing release resolves a total of 22 issues. + +New committer + +https://github.com/mxm;>Max Michels has been voted a committer by the Flink PMC. + +Flink adapter for Apache SAMOA + +http://samoa.incubator.apache.org;>Apache SAMOA (incubating) is a +distributed streaming machine learning (ML) framework with a +programming abstraction for distributed streaming ML algorithms. SAMOA +runs on a variety of backend engines, currently Apache Storm and +Apache S4. A https://github.com/apache/incubator-samoa/pull/11;>pull +request is +available at the SAMOA repository that adds a Flink adapter for SAMOA. + +Easy Flink deployment on Google Compute Cloud + +Flink is now integrated in bdutil, Googleâs open source tool for +creating and configuring (Hadoop) clusters in Google Compute +Engine. Deployment of Flink clusters in now supported starting with +https://groups.google.com/forum/#!topic/gcp-hadoop-announce/uVJ_6y9cGKM;>bdutil +1.2.0. + +Flink on the Web + +A new blog post on http://flink.apache.org/news/2015/02/09/streaming-example.html;>Flink +Streaming +was published at the blog. Flink was mentioned in several articles on +the web. Here are some examples: + + + +http://dataconomy.com/how-flink-became-an-apache-top-level-project/;>How Flink became an Apache Top-Level Project + + +https://www.linkedin.com/pulse/stale-synchronous-parallelism-new-frontier-apache-flink-nam-luc-tran?utm_content=buffer461afutm_medium=socialutm_source=linkedin.comutm_campaign=buffer;>Stale Synchronous Parallelism: The new frontier for Apache Flink? + + +http://www.hadoopsphere.com/2015/02/distributed-data-processing-with-apache.html;>Distributed data processing with Apache Flink + + +http://www.hadoopsphere.com/2015/02/ciao-latency-hallo-speed.html;>Ciao latency, hello speed + + + +In the Flink master + +The following features have been now merged in Flinkâs master repository. + +Gelly + +Gelly, Flinkâs Graph API allows users to manipulate graph-shaped data +directly. Hereâs for example a calculation of shortest paths in a +graph: + +GraphLong, Double, Double graph = Graph.fromDataSet(vertices, edges, env); + +DataSetVertexLong,
[54/78] [abbrv] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/08/24/introducing-flink-gelly.html -- diff --git a/content/news/2015/08/24/introducing-flink-gelly.html b/content/news/2015/08/24/introducing-flink-gelly.html new file mode 100644 index 000..714e75a --- /dev/null +++ b/content/news/2015/08/24/introducing-flink-gelly.html @@ -0,0 +1,649 @@ + + + + + + + +Apache Flink: Introducing Gelly: Graph Processing with Apache Flink + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + + + Introducing Gelly: Graph Processing with Apache Flink + + +24 Aug 2015 + +This blog post introduces Gelly, Apache Flinkâs graph-processing API and library. Flinkâs native support +for iterations makes it a suitable platform for large-scale graph analytics. +By leveraging delta iterations, Gelly is able to map various graph processing models such as +vertex-centric or gather-sum-apply to Flink dataflows. + +Gelly allows Flink users to perform end-to-end data analysis in a single system. +Gelly can be seamlessly used with Flinkâs DataSet API, +which means that pre-processing, graph creation, analysis, and post-processing can be done +in the same application. At the end of this post, we will go through a step-by-step example +in order to demonstrate that loading, transformation, filtering, graph creation, and analysis +can be performed in a single Flink program. + +Overview + + + What is Gelly? + Graph Representation and Creation + Transformations and Utilities + Iterative Graph Processing + Library of Graph Algorithms + Use-Case: Music Profiles + Ongoing and Future Work + + + + +What is Gelly? + +Gelly is a Graph API for Flink. It is currently supported in both Java and Scala. +The Scala methods are implemented as wrappers on top of the basic Java operations. +The API contains a set of utility functions for graph analysis, supports iterative graph +processing and introduces a library of graph algorithms. + + + + + +Back to top + +Graph Representation and Creation + +In Gelly, a graph is represented by a DataSet of vertices and a DataSet of edges. +A vertex is defined by its unique ID and a value, whereas an edge is defined by its source ID, +target ID, and value. A vertex or edge for which a value is not specified will simply have the +value type set to NullValue. + +A graph can be created from: + + + DataSet of edges and an optional DataSet of vertices using Graph.fromDataSet() + DataSet of Tuple3 and an optional DataSet of Tuple2 using Graph.fromTupleDataSet() + Collection of edges and an optional Collection of vertices using Graph.fromCollection() + + +In all three cases, if the vertices are not provided, +Gelly will automatically produce the vertex IDs from the edge source and target IDs. + +Back to top + +Transformations and Utilities + +These are methods of the Graph class and include common graph metrics, transformations +and mutations as well as neighborhood aggregations. + +Common Graph Metrics +These methods can be used to
[21/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/logo/svg/white_filled.svg -- diff --git a/content/img/logo/svg/white_filled.svg b/content/img/logo/svg/white_filled.svg deleted file mode 100755 index 28bd9f9..000 --- a/content/img/logo/svg/white_filled.svg +++ /dev/null @@ -1,231 +0,0 @@ - - -http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd;> -http://www.w3.org/2000/svg; xmlns:xlink="http://www.w3.org/1999/xlink; x="0px" y="0px" -width="2000px" height="1280px" viewBox="0 0 2000 1280" enable-background="new 0 0 2000 1280" xml:space="preserve"> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -Filled white - http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/managed-state.png -- diff --git a/content/img/managed-state.png b/content/img/managed-state.png deleted file mode 100755 index 3dbb2fd..000 Binary files a/content/img/managed-state.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/memory_heap_division.png -- diff --git a/content/img/memory_heap_division.png b/content/img/memory_heap_division.png deleted file mode 100644 index 2b4c2e2..000 Binary files a/content/img/memory_heap_division.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/navbar-brand-logo.jpg -- diff --git a/content/img/navbar-brand-logo.jpg b/content/img/navbar-brand-logo.jpg deleted file mode 100755 index 5993ee8..000 Binary files a/content/img/navbar-brand-logo.jpg and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/navbar-brand-logo.png -- diff --git a/content/img/navbar-brand-logo.png b/content/img/navbar-brand-logo.png deleted file mode 100644 index 152f74e..000 Binary files a/content/img/navbar-brand-logo.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/one_runtime.png -- diff --git a/content/img/one_runtime.png b/content/img/one_runtime.png deleted file mode 100644 index 9cb4363..000 Binary files a/content/img/one_runtime.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/optimizer_choice.png -- diff --git a/content/img/optimizer_choice.png b/content/img/optimizer_choice.png deleted file mode 100644 index 1f8004b..000 Binary files a/content/img/optimizer_choice.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/otto-group-logo.jpg -- diff --git a/content/img/otto-group-logo.jpg b/content/img/otto-group-logo.jpg deleted file mode 100644 index f578af6..000 Binary files a/content/img/otto-group-logo.jpg and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/out_of_order_stream.png -- diff --git a/content/img/out_of_order_stream.png b/content/img/out_of_order_stream.png deleted file mode 100644 index 20ad09d..000 Binary files a/content/img/out_of_order_stream.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/parallel_dataflows.png -- diff --git a/content/img/parallel_dataflows.png b/content/img/parallel_dataflows.png deleted file mode 100644 index dbd83b2..000 Binary files a/content/img/parallel_dataflows.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/img/poweredby/alibaba-logo.png
[04/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/js/graphCreator.js -- diff --git a/content/visualizer/js/graphCreator.js b/content/visualizer/js/graphCreator.js new file mode 100755 index 000..86c8c81 --- /dev/null +++ b/content/visualizer/js/graphCreator.js @@ -0,0 +1,445 @@ +/*! + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/* + * Document initialization. + */ +$(document).ready(function () +{ + zoom = d3.behavior.zoom("#svg-main").on("zoom", function() { + var ev = d3.event; + d3.select("#svg-main g") + .attr("transform", "translate(" + ev.translate + ") scale(" + ev.scale + ")"); + }); + zoom.scaleExtent([0.3, 3]) +}); + +//The current JSON file +var JSONData; +//The informations for the iterations +var iterationIds = new Array(); +var iterationGraphs = new Array(); +var iterationWidths = new Array(); +var iterationHeights = new Array(); + +//The zoom element +var zoom; +//informations about the enviroment +var svgWidth; +var svgHeight; + +//Renders and draws the graph +function drawGraph(data, svgID){ + JSONData = data; + + //First step: precompute all iteration graphs + + //find all iterations + iterationNodes = searchForIterationNodes(); + + //add the graphs of iterations and their sizes + Ids to arrays + if (iterationNodes != null) { + for (var i in iterationNodes) { + var itNode = iterationNodes[i]; + iterationIds.push(itNode.id); + var g0 = loadJsonToDagre(itNode); + iterationGraphs.push(g0); + var r = new dagreD3.Renderer(); + var l = dagreD3.layout() +.nodeSep(20) +.rankDir("LR"); + l = r.layout(l).run(g0, d3.select("#svg-main")); + + iterationWidths.push(l._value.width); + iterationHeights.push(l._value.height); + + //Clean svg + $("#svg-main g").empty(); + } + } + + //Continue normal + var g = loadJsonToDagre(data); + var selector = svgID + " g"; + var renderer = new dagreD3.Renderer(); + var layout = dagreD3.layout() + .nodeSep(20) + .rankDir("LR"); + var svgElement = d3.select(selector); + layout = renderer.layout(layout).run(g, svgElement); + + svgHeight = layout._value.height; + svgWidth = layout._value.width; + +var svg = d3.select("#svg-main") + .attr("width", $(document).width() - 15) + .attr("height", $(document).height() - 15 - 110) +// .attr("viewBox", "0 0 "+ ($(document).width() - 150) +" "+($(document).height() - 15 - 110)) + .call(zoom); + + // This should now draw the precomputed graphs in the svgs... . + + for (var i in iterationIds) { + var workset = searchForNode(iterationIds[i]); + renderer = new dagreD3.Renderer(); + layout = dagreD3.layout() +.nodeSep(20) +.rankDir("LR"); + selector = "#svg-"+iterationIds[i]+" g"; + svgElement = d3.select(selector); + layout = renderer.layout(layout).run(iterationGraphs[i], svgElement); + } + + //enable Overlays and register function for overlay-infos + $("a[rel]").overlay({ + onBeforeLoad: function(){ + var id = this.getTrigger().attr("nodeID") + showProperties(id); + } + }); + + } + +//Creates the dagreD3 graph object +//Responsible for adding nodes and edges +function loadJsonToDagre(data){ + + //stores all nodes that are in current graph -> no edges to nodes which are outside of current
[17/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/03/02/february-2015-in-flink.html -- diff --git a/content/news/2015/03/02/february-2015-in-flink.html b/content/news/2015/03/02/february-2015-in-flink.html new file mode 100644 index 000..1e3a09b --- /dev/null +++ b/content/news/2015/03/02/february-2015-in-flink.html @@ -0,0 +1,308 @@ + + + + + + + +Apache Flink: February 2015 in the Flink community + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + + + February 2015 in the Flink community + + +02 Mar 2015 + +February might be the shortest month of the year, but this does not +mean that the Flink community has not been busy adding features to the +system and fixing bugs. Hereâs a rundown of the activity in the Flink +community last month. + +0.8.1 release + +Flink 0.8.1 was released. This bugfixing release resolves a total of 22 issues. + +New committer + +https://github.com/mxm;>Max Michels has been voted a committer by the Flink PMC. + +Flink adapter for Apache SAMOA + +http://samoa.incubator.apache.org;>Apache SAMOA (incubating) is a +distributed streaming machine learning (ML) framework with a +programming abstraction for distributed streaming ML algorithms. SAMOA +runs on a variety of backend engines, currently Apache Storm and +Apache S4. A https://github.com/apache/incubator-samoa/pull/11;>pull +request is +available at the SAMOA repository that adds a Flink adapter for SAMOA. + +Easy Flink deployment on Google Compute Cloud + +Flink is now integrated in bdutil, Googleâs open source tool for +creating and configuring (Hadoop) clusters in Google Compute +Engine. Deployment of Flink clusters in now supported starting with +https://groups.google.com/forum/#!topic/gcp-hadoop-announce/uVJ_6y9cGKM;>bdutil +1.2.0. + +Flink on the Web + +A new blog post on http://flink.apache.org/news/2015/02/09/streaming-example.html;>Flink +Streaming +was published at the blog. Flink was mentioned in several articles on +the web. Here are some examples: + + + +http://dataconomy.com/how-flink-became-an-apache-top-level-project/;>How Flink became an Apache Top-Level Project + + +https://www.linkedin.com/pulse/stale-synchronous-parallelism-new-frontier-apache-flink-nam-luc-tran?utm_content=buffer461afutm_medium=socialutm_source=linkedin.comutm_campaign=buffer;>Stale Synchronous Parallelism: The new frontier for Apache Flink? + + +http://www.hadoopsphere.com/2015/02/distributed-data-processing-with-apache.html;>Distributed data processing with Apache Flink + + +http://www.hadoopsphere.com/2015/02/ciao-latency-hallo-speed.html;>Ciao latency, hello speed + + + +In the Flink master + +The following features have been now merged in Flinkâs master repository. + +Gelly + +Gelly, Flinkâs Graph API allows users to manipulate graph-shaped data +directly. Hereâs for example a calculation of shortest paths in a +graph: + +GraphLong, Double, Double graph = Graph.fromDataSet(vertices, edges, env); + +DataSetVertexLong,
[01/39] flink-web git commit: Rebuild site
Repository: flink-web Updated Branches: refs/heads/asf-site 16a92b0c1 -> 9ec0a879f http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/js/program.js -- diff --git a/content/visualizer/js/program.js b/content/visualizer/js/program.js new file mode 100755 index 000..683185d --- /dev/null +++ b/content/visualizer/js/program.js @@ -0,0 +1,233 @@ +/*! + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +var maxColumnWidth = 200; +var minColumnWidth = 100; + +// global variable for the currently requested plan +var pactPlanRequested = 0; + +/* + * This function toggels the child checkbox on and of, depending on the parent's state + */ +function toggleShowPlanBox(box) +{ + var child = $('#suspendJobDuringPlanCheck'); + + if (box.is(':checked')) { +child.attr('disabled', false); + } + else { +child.attr('disabled', true); + } +} + +/* + * Shows an error message below the upload field. + */ +function showUploadError(message) +{ + $('#upload_error_text').fadeOut("fast", function () { $('#upload_error_text')[0].innerHTML = "" + message; + $('#upload_error_text').fadeIn("slow"); } ); +} + +/* + * Checks the selected file and triggers an upload, if all is correct. + */ +function processUpload() +{ + + var filename = $('#upload_file_input').val(); + var len = filename.length; + if (len == 0) { +showUploadError("Please select a file."); + } + else if (len > 4 && filename.substr(len - 4, len) == ".jar") { +$('#upload_form')[0].submit(); + } + else { +showUploadError("Please select a .jar file."); + } +} + +/* + * This function makes sure only one checkbox is selected. + * Upon selection it initializes the drawing of the pact plan. + * Upon deselection, it clears the pact plan. + */ +function toggleCheckboxes(box) +{ + + if (box.is(':checked')) { +$('.jobItemCheckbox').attr('checked', false); +box.attr('checked', true); +var id = box.parentsUntil('.JobListItems').parent().attr('id').substr(4); +var assemblerClass = box.attr('id'); + +$('#mainCanvas').html(''); +pactPlanRequested = id; + +$.ajax({ +type: "GET", +url: "pactPlan", +data: { job: id, assemblerClass: assemblerClass}, +success: function(response) { showPreviewPlan(response); } +}); + } + else { +$('#mainCanvas').html(''); + } +} + +/* + * Function that takes the returned plan and draws it. + */ +function showPreviewPlan(data) +{ + $("#mainCanvas").empty(); +var svgElement = ""; +$("#mainCanvas").append(svgElement); +drawGraph(data.plan, "#svg-main"); +pactPlanRequested = 0; + +//activate zoom buttons +activateZoomButtons(); +// } +} + +/* + * Asynchronously loads the jobs and creates a list of them. + */ +function loadJobList() +{ + $.get("jobs", { action: "list" }, createJobList); +} + +/* + * Triggers an AJAX request to delete a job. + */ +function deleteJob(id) +{ + var name = id.substr(4); + $.get("jobs", { action: "delete", filename: name }, loadJobList); +} + +/* + * Creates and lists the returned jobs. + */ +function createJobList(data) +{ + var markup = ""; + + var lines = data.split("\n"); + for (var i = 0; i < lines.length; i++) + { +if (lines[i] == null || lines[i].length == 0) { + continue; +} + +var date = "unknown date"; +var assemblerClass = "no entry class specified"; + +var tokens = lines[i].split("\t"); +var name = tokens[0]; +if (tokens.length > 1) { + date = tokens[1]; + if (tokens.length > 2) { +assemblerClass = tokens[2]; + } +} + +var entries = assemblerClass.split("#"); +var classes = entries[0].split(","); + +markup += ''; +markup += '' + name + ''; +markup += '' + date + ''; +markup += ''; + +var j = 0; +for (var idx in classes) { + markup += ''; + markup += '' + classes[idx] + ''; +} +markup += ''; + } + + // add the contents + $('#jobsContents').html(markup); + + // register the event handler that triggers the delete when the
[25/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/svg/color_black.svg -- diff --git a/content/img/logo/svg/color_black.svg b/content/img/logo/svg/color_black.svg new file mode 100755 index 000..d54707c --- /dev/null +++ b/content/img/logo/svg/color_black.svg @@ -0,0 +1,1561 @@ + + +http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd;> +http://www.w3.org/2000/svg; xmlns:xlink="http://www.w3.org/1999/xlink; x="0px" y="0px" +width="2000px" height="1280px" viewBox="0 0 2000 1280" enable-background="new 0 0 2000 1280" xml:space="preserve"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
[16/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/05/11/Juggling-with-Bits-and-Bytes.html -- diff --git a/content/news/2015/05/11/Juggling-with-Bits-and-Bytes.html b/content/news/2015/05/11/Juggling-with-Bits-and-Bytes.html new file mode 100644 index 000..a7a5436 --- /dev/null +++ b/content/news/2015/05/11/Juggling-with-Bits-and-Bytes.html @@ -0,0 +1,383 @@ + + + + + + + +Apache Flink: Juggling with Bits and Bytes + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + + + Juggling with Bits and Bytes + + +11 May 2015 by Fabian Hüske (https://twitter.com/fhueske;>@fhueske) + +How Apache Flink operates on binary data + +Nowadays, a lot of open-source systems for analyzing large data sets are implemented in Java or other JVM-based programming languages. The most well-known example is Apache Hadoop, but also newer frameworks such as Apache Spark, Apache Drill, and also Apache Flink run on JVMs. A common challenge that JVM-based data analysis engines face is to store large amounts of data in memory - both for caching and for efficient processing such as sorting and joining of data. Managing the JVM memory well makes the difference between a system that is hard to configure and has unpredictable reliability and performance and a system that behaves robustly with few configuration knobs. + +In this blog post we discuss how Apache Flink manages memory, talk about its custom data de/serialization stack, and show how it operates on binary data. + +Data Objects? Letâs put them on the heap! + +The most straight-forward approach to process lots of data in a JVM is to put it as objects on the heap and operate on these objects. Caching a data set as objects would be as simple as maintaining a list containing an object for each record. An in-memory sort would simply sort the list of objects. +However, this approach has a few notable drawbacks. First of all it is not trivial to watch and control heap memory usage when a lot of objects are created and invalidated constantly. Memory overallocation instantly kills the JVM with an OutOfMemoryError. Another aspect is garbage collection on multi-GB JVMs which are flooded with new objects. The overhead of garbage collection in such environments can easily reach 50% and more. Finally, Java objects come with a certain space overhead depending on the JVM and platform. For data sets with many small objects this can significantly reduce the effectively usable amount of memory. Given proficient system design and careful, use-case specific system parameter tuning, heap memory usage can be more or less controlled and OutOfMemoryErrors avoided. However, such setups are rather fragile especially if data characteristics or the execution environment change. + +What is Flink doing about that? + +Apache Flink has its roots at a research project which aimed to combine the best technologies of MapReduce-based systems and parallel database systems. Coming from this background, Flink has always had
[06/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/js/d3.js -- diff --git a/content/visualizer/js/d3.js b/content/visualizer/js/d3.js new file mode 100755 index 000..2d09bc9 --- /dev/null +++ b/content/visualizer/js/d3.js @@ -0,0 +1,9270 @@ +/** + * D3 renderer from http://d3js.org/ + * Copyright (c) 2010-2014, Michael Bostock + * + * Under BSD License (as stated under http://d3js.org/) + */ + +!function() { + var d3 = { +version: "3.4.6" + }; + if (!Date.now) Date.now = function() { +return +new Date(); + }; + var d3_arraySlice = [].slice, d3_array = function(list) { +return d3_arraySlice.call(list); + }; + var d3_document = document, d3_documentElement = d3_document.documentElement, d3_window = window; + try { +d3_array(d3_documentElement.childNodes)[0].nodeType; + } catch (e) { +d3_array = function(list) { + var i = list.length, array = new Array(i); + while (i--) array[i] = list[i]; + return array; +}; + } + try { +d3_document.createElement("div").style.setProperty("opacity", 0, ""); + } catch (error) { +var d3_element_prototype = d3_window.Element.prototype, d3_element_setAttribute = d3_element_prototype.setAttribute, d3_element_setAttributeNS = d3_element_prototype.setAttributeNS, d3_style_prototype = d3_window.CSSStyleDeclaration.prototype, d3_style_setProperty = d3_style_prototype.setProperty; +d3_element_prototype.setAttribute = function(name, value) { + d3_element_setAttribute.call(this, name, value + ""); +}; +d3_element_prototype.setAttributeNS = function(space, local, value) { + d3_element_setAttributeNS.call(this, space, local, value + ""); +}; +d3_style_prototype.setProperty = function(name, value, priority) { + d3_style_setProperty.call(this, name, value + "", priority); +}; + } + d3.ascending = d3_ascending; + function d3_ascending(a, b) { +return a < b ? -1 : a > b ? 1 : a >= b ? 0 : NaN; + } + d3.descending = function(a, b) { +return b < a ? -1 : b > a ? 1 : b >= a ? 0 : NaN; + }; + d3.min = function(array, f) { +var i = -1, n = array.length, a, b; +if (arguments.length === 1) { + while (++i < n && !((a = array[i]) != null && a <= a)) a = undefined; + while (++i < n) if ((b = array[i]) != null && a > b) a = b; +} else { + while (++i < n && !((a = f.call(array, array[i], i)) != null && a <= a)) a = undefined; + while (++i < n) if ((b = f.call(array, array[i], i)) != null && a > b) a = b; +} +return a; + }; + d3.max = function(array, f) { +var i = -1, n = array.length, a, b; +if (arguments.length === 1) { + while (++i < n && !((a = array[i]) != null && a <= a)) a = undefined; + while (++i < n) if ((b = array[i]) != null && b > a) a = b; +} else { + while (++i < n && !((a = f.call(array, array[i], i)) != null && a <= a)) a = undefined; + while (++i < n) if ((b = f.call(array, array[i], i)) != null && b > a) a = b; +} +return a; + }; + d3.extent = function(array, f) { +var i = -1, n = array.length, a, b, c; +if (arguments.length === 1) { + while (++i < n && !((a = c = array[i]) != null && a <= a)) a = c = undefined; + while (++i < n) if ((b = array[i]) != null) { +if (a > b) a = b; +if (c < b) c = b; + } +} else { + while (++i < n && !((a = c = f.call(array, array[i], i)) != null && a <= a)) a = undefined; + while (++i < n) if ((b = f.call(array, array[i], i)) != null) { +if (a > b) a = b; +if (c < b) c = b; + } +} +return [ a, c ]; + }; + d3.sum = function(array, f) { +var s = 0, n = array.length, a, i = -1; +if (arguments.length === 1) { + while (++i < n) if (!isNaN(a = +array[i])) s += a; +} else { + while (++i < n) if (!isNaN(a = +f.call(array, array[i], i))) s += a; +} +return s; + }; + function d3_number(x) { +return x != null && !isNaN(x); + } + d3.mean = function(array, f) { +var s = 0, n = array.length, a, i = -1, j = n; +if (arguments.length === 1) { + while (++i < n) if (d3_number(a = array[i])) s += a; else --j; +} else { + while (++i < n) if (d3_number(a = f.call(array, array[i], i))) s += a; else --j; +} +return j ? s / j : undefined; + }; + d3.quantile = function(values, p) { +var H = (values.length - 1) * p + 1, h = Math.floor(H), v = +values[h - 1], e = H - h; +return e ? v + e * (values[h] - v) : v; + }; + d3.median = function(array, f) { +if (arguments.length > 1) array = array.map(f); +array = array.filter(d3_number); +return array.length ? d3.quantile(array.sort(d3_ascending), .5) : undefined; + }; + function d3_bisector(compare) { +return { + left: function(a, x, lo, hi) { +if (arguments.length < 3) lo = 0; +if (arguments.length < 4) hi = a.length; +
[38/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/blog/feed.xml -- diff --git a/content/blog/feed.xml b/content/blog/feed.xml new file mode 100644 index 000..066ae2f --- /dev/null +++ b/content/blog/feed.xml @@ -0,0 +1,6041 @@ + +http://www.w3.org/2005/Atom;> + +Flink Blog Feed +Flink Blog +http://flink.apache.org/blog +http://flink.apache.org/blog/feed.xml; rel="self" type="application/rss+xml" /> + + +Apache Flink 1.1.4 Released +pThe Apache Flink community released the next bugfix version of the Apache Flink 1.1 series./p + +pThis release includes major robustness improvements for checkpoint cleanup on failures and consumption of intermediate streams. We highly recommend all users to upgrade to Flink 1.1.4./p + +div class=highlightprecode class=language-xmlspan class=ntlt;dependencygt;/span + span class=ntlt;groupIdgt;/spanorg.apache.flinkspan class=ntlt;/groupIdgt;/span + span class=ntlt;artifactIdgt;/spanflink-javaspan class=ntlt;/artifactIdgt;/span + span class=ntlt;versiongt;/span1.1.4span class=ntlt;/versiongt;/span +span class=ntlt;/dependencygt;/span +span class=ntlt;dependencygt;/span + span class=ntlt;groupIdgt;/spanorg.apache.flinkspan class=ntlt;/groupIdgt;/span + span class=ntlt;artifactIdgt;/spanflink-streaming-java_2.10span class=ntlt;/artifactIdgt;/span + span class=ntlt;versiongt;/span1.1.4span class=ntlt;/versiongt;/span +span class=ntlt;/dependencygt;/span +span class=ntlt;dependencygt;/span + span class=ntlt;groupIdgt;/spanorg.apache.flinkspan class=ntlt;/groupIdgt;/span + span class=ntlt;artifactIdgt;/spanflink-clients_2.10span class=ntlt;/artifactIdgt;/span + span class=ntlt;versiongt;/span1.1.4span class=ntlt;/versiongt;/span +span class=ntlt;/dependencygt;/span/code/pre/div + +pYou can find the binaries on the updated a href=http://flink.apache.org/downloads.htmlDownloads page/a./p + +h2 id=note-for-rocksdb-backend-usersNote for RocksDB Backend Users/h2 + +pWe updated Flinkâs RocksDB dependency version from code4.5.1/code to code4.11.2/code. Between these versions some of RocksDBâs internal configuration defaults changed that would affect the memory footprint of running Flink with RocksDB. Therefore, we manually reset them to the previous defaults. If you want to run with the new Rocks 4.11.2 defaults, you can do this via:/p + +div class=highlightprecode class=language-javaspan class=nRocksDBStateBackend/span span class=nbackend/span span class=o=/span span class=knew/span span class=nfRocksDBStateBackend/spanspan class=o(/spanspan class=squot;...quot;/spanspan class=o);/span +span class=c1// Use the new default options. Otherwise, the default for RocksDB 4.5.1/span +span class=c1// `PredefinedOptions.DEFAULT_ROCKS_4_5_1` will be used./span +span class=nbackend/spanspan class=o./spanspan class=nasetPredefinedOptions/spanspan class=o(/spanspan class=nPredefinedOptions/spanspan class=o./spanspan class=naDEFAULT/spanspan class=o);/span/code/pre/div + +h2 id=release-notes---flink---version-114Release Notes - Flink - Version 1.1.4/h2 + +h3 id=sub-taskSub-task/h3 +ul +li[a href=https://issues.apache.org/jira/browse/FLINK-4510FLINK-4510/a;] - Always create CheckpointCoordinator +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4984FLINK-4984/a;] - Add Cancellation Barriers to BarrierTracker and BarrierBuffer +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4985FLINK-4985/a;] - Report Declined/Canceled Checkpoints to Checkpoint Coordinator +/li +/ul + +h3 id=bugBug/h3 +ul +li[a href=https://issues.apache.org/jira/browse/FLINK-2662FLINK-2662/a;] - CompilerException: quot;Bug: Plan generation for Unions picked a ship strategy between binary plan operators.quot; +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-3680FLINK-3680/a;] - Remove or improve (not set) text in the Job Plan UI +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-3813FLINK-3813/a;] - YARNSessionFIFOITCase.testDetachedMode failed on Travis +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4108FLINK-4108/a;] - NPE in Row.productArity +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4506FLINK-4506/a;] - CsvOutputFormat defaults allowNullValues to false, even though doc and declaration says true +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4581FLINK-4581/a;] - Table API throws quot;No suitable driver found for jdbc:calcitequot; +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4586FLINK-4586/a;] - NumberSequenceIterator and Accumulator threading issue +/li +li[a href=https://issues.apache.org/jira/browse/FLINK-4619FLINK-4619/a;] - JobManager does not answer to client when restore from savepoint fails +/li +li[a
[39/39] flink-web git commit: Rebuild site
Rebuild site Project: http://git-wip-us.apache.org/repos/asf/flink-web/repo Commit: http://git-wip-us.apache.org/repos/asf/flink-web/commit/9ec0a879 Tree: http://git-wip-us.apache.org/repos/asf/flink-web/tree/9ec0a879 Diff: http://git-wip-us.apache.org/repos/asf/flink-web/diff/9ec0a879 Branch: refs/heads/asf-site Commit: 9ec0a879f23ebb9c0533bf7e001ebf830a8bdd6f Parents: 16a92b0 Author: Robert MetzgerAuthored: Wed Jan 18 15:04:07 2017 +0100 Committer: Robert Metzger Committed: Wed Jan 18 15:04:07 2017 +0100 -- content/blog/feed.xml | 6041 content/blog/index.html | 747 ++ content/blog/page2/index.html | 743 ++ content/blog/page3/index.html | 756 ++ content/blog/page4/index.html | 760 ++ .../release_1.0.0-changelog_known_issues.html | 1139 +++ content/blog/release_1.1.0-changelog.html | 689 ++ content/community.html | 743 ++ content/contribute-code.html| 519 + content/contribute-documentation.html | 248 + content/css/flink.css | 286 + content/css/syntax.css | 60 + content/doap_flink.rdf | 51 + content/documentation.html | 196 + content/downloads.html | 320 + content/ecosystem.html | 278 + content/faq.html| 683 ++ content/favicon.ico | Bin 0 -> 1150 bytes content/features.html | 511 + content/how-to-contribute.html | 345 + content/img/alibaba-logo.png| Bin 0 -> 22897 bytes content/img/assets/WhatIsFlink.png | Bin 0 -> 248892 bytes content/img/assets/grep.png | Bin 0 -> 121239 bytes content/img/assets/hadoop-img.png | Bin 0 -> 67841 bytes content/img/assets/optimizer-visual.png | Bin 0 -> 204452 bytes content/img/assets/pagerank.pdf | Bin 0 -> 21052 bytes content/img/assets/pagerank.png | Bin 0 -> 135851 bytes content/img/blog/GSA-plan.png | Bin 0 -> 54646 bytes content/img/blog/appeared-in.png| Bin 0 -> 317098 bytes content/img/blog/blog_basic_window.png | Bin 0 -> 144158 bytes content/img/blog/blog_data_driven.png | Bin 0 -> 96886 bytes content/img/blog/blog_multi_input.png | Bin 0 -> 112104 bytes content/img/blog/blog_social_media.png | Bin 0 -> 98666 bytes content/img/blog/blog_stream_join.png | Bin 0 -> 112305 bytes content/img/blog/cep-monitoring.svg | 2838 ++ content/img/blog/code-growth.png| Bin 0 -> 87120 bytes content/img/blog/commit-stats.png | Bin 0 -> 5 bytes content/img/blog/community-growth.png | Bin 0 -> 102186 bytes content/img/blog/data-serialization.png | Bin 0 -> 395455 bytes content/img/blog/emr-firefoxsettings.png| Bin 0 -> 176706 bytes content/img/blog/emr-hadoopversion.png | Bin 0 -> 97799 bytes content/img/blog/emr-jobmanager.png | Bin 0 -> 84576 bytes content/img/blog/emr-running.png| Bin 0 -> 78210 bytes content/img/blog/emr-security.png | Bin 0 -> 76951 bytes content/img/blog/emr-yarnappmaster.png | Bin 0 -> 85393 bytes content/img/blog/feature-timeline.png | Bin 0 -> 182292 bytes content/img/blog/ff-speakers.png| Bin 0 -> 367370 bytes content/img/blog/flink-1.0.png | Bin 0 -> 114360 bytes content/img/blog/flink-dow-2016.png | Bin 0 -> 3251 bytes content/img/blog/flink-forward-banner.png | Bin 0 -> 37458 bytes content/img/blog/flink-hod-2016.png | Bin 0 -> 3090 bytes content/img/blog/flink-lines-of-code-2016.png | Bin 0 -> 3935 bytes content/img/blog/flink-meetups-dec2016.png | Bin 0 -> 560426 bytes content/img/blog/flink-releases-2016.png| Bin 0 -> 109796 bytes content/img/blog/flink-stack.png| Bin 0 -> 32934 bytes content/img/blog/flink-storm.png| Bin 0 -> 127217 bytes content/img/blog/flinkSer-int-gc.png| Bin 0 -> 12405 bytes content/img/blog/flinkSer-int-mem.png | Bin 0 -> 15689 bytes content/img/blog/github-stats-2016.png | Bin 0 -> 93704 bytes content/img/blog/hadoop-summit.png | Bin 0 -> 330780 bytes content/img/blog/hcompat-flow.png | Bin 0 -> 102575 bytes content/img/blog/hcompat-logos.png | Bin 0 -> 320253 bytes content/img/blog/iteration.png | Bin 0 -> 24531 bytes
[02/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/js/jquery.tools.min.js -- diff --git a/content/visualizer/js/jquery.tools.min.js b/content/visualizer/js/jquery.tools.min.js new file mode 100755 index 000..dc914d8 --- /dev/null +++ b/content/visualizer/js/jquery.tools.min.js @@ -0,0 +1,20 @@ +/*! + * jQuery Tools v1.2.7 - The missing UI library for the Web + * + * overlay/overlay.js + * scrollable/scrollable.js + * tabs/tabs.js + * tooltip/tooltip.js + * + * NO COPYRIGHTS OR LICENSES. DO WHAT YOU LIKE. + * + * http://flowplayer.org/tools/ + * + */ +/*! jQuery v1.7.2 jquery.com | jquery.org/license */ +(function(a,b){function cy(a){return f.isWindow(a)?a:a.nodeType===9?a.defaultView||a.parentWindow:!1}function cu(a){if(!cj[a]){var b=c.body,d=f("<"+a+">").appendTo(b),e=d.css("display");d.remove();if(e==="none"||e===""){ck||(ck=c.createElement("iframe"),ck.frameBorder=ck.width=ck.height=0),b.appendChild(ck);if(!cl||!ck.createElement)cl=(ck.contentWindow||ck.contentDocument).document,cl.write((f.support.boxModel?"":"")+""),cl.close();d=cl.createElement(a),cl.body.appendChild(d),e=f.css(d,"display"),b.removeChild(ck)}cj[a]=e}return cj[a]}function ct(a,b){var c={};f.each(cp.concat.apply([],cp.slice(0,b)),function(){c[this]=a});return c}function cs(){cq=b}function cr(){setTimeout(cs,0);return cq=f.now()}function ci(){try{return new a.ActiveXObject("Microsoft.XMLHTTP")}catch(b){}}function ch(){try{return new a.XMLHttpRequest}catch(b){}}function cb(a,c){a.dataFilter&&(c=a.dataFilter(c,a.dataType));var d=a.dataTypes,e={},g,h,i=d.length,j,k=d[0],l,m,n,o,p;for(g=1; g0){if(c!=="border")for(;e
[33/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/contribute-documentation.html -- diff --git a/content/contribute-documentation.html b/content/contribute-documentation.html new file mode 100644 index 000..11f9000 --- /dev/null +++ b/content/contribute-documentation.html @@ -0,0 +1,248 @@ + + + + + + + +Apache Flink: Contribute Documentation + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + +Contribute Documentation + + Good documentation is crucial for any kind of software. This is especially true for sophisticated software systems such as distributed data processing engines like Apache Flink. The Apache Flink community aims to provide concise, precise, and complete documentation and welcomes any contribution to improve Apache Flinkâs documentation. + + + + Obtain the documentation sources + Before you start working on the documentation ⦠+ Update or extend the documentation + Submit your contribution + + + + +Obtain the documentation sources + +Apache Flinkâs documentation is maintained in the same http://git-scm.com/;>git repository as the code base. This is done to ensure that code and documentation can be easily kept in sync. + +The easiest way to contribute documentation is to fork https://github.com/apache/flink;>Flinkâs mirrored repository on Github into your own Github account by clicking on the fork button at the top right. If you have no Github account, you can create one for free. + +Next, clone your fork to your local machine. + +git clone https://github.com/your-user-name/flink.git + + +The documentation is located in the docs/ subdirectory of the Flink code base. + +Before you start working on the documentation ⦠+ +⦠please make sure there exists a https://issues.apache.org/jira/browse/FLINK;>JIRA issue that corresponds to your contribution. We require all documentation changes to refer to a JIRA issue, except for trivial fixes such as typos. + +Update or extend the documentation + +The Flink documentation is written in http://daringfireball.net/projects/markdown/;>Markdown. Markdown is a lightweight markup language which can be translated to HTML. + +In order to update or extend the documentation you have to modify the Markdown (.md) files. Please verify your changes by starting the build script in preview mode. + +cd docs +./build_docs.sh -p + + +The script compiles the Markdown files into static HTML pages and starts a local webserver. Open your browser at http://localhost:4000 to view the compiled documentation including your changes. The served documentation is automatically re-compiled and updated when you modify and save Markdown files and refresh your browser. + +Please feel free to ask any questions you have on the developer mailing list. + +Submit your contribution + +The Flink project accepts documentation contributions through the https://github.com/apache/flink;>GitHub Mirror as https://help.github.com/articles/using-pull-requests;>Pull Requests. Pull requests are a simple way of offering a
[36/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/blog/release_1.0.0-changelog_known_issues.html -- diff --git a/content/blog/release_1.0.0-changelog_known_issues.html b/content/blog/release_1.0.0-changelog_known_issues.html new file mode 100644 index 000..9b7ddc7 --- /dev/null +++ b/content/blog/release_1.0.0-changelog_known_issues.html @@ -0,0 +1,1139 @@ + + + + + + + +Apache Flink: Release 1.0.0 â Changelog and Known Issues + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + +Release 1.0.0 â Changelog and Known Issues + + + Changelog + Sub-task + Bug + Improvement + Task + Test + Wish + + + Known Issues + + +Changelog + +Sub-task + + +[https://issues.apache.org/jira/browse/FLINK-7;>FLINK-7] - [GitHub] Enable Range Partitioner + +[https://issues.apache.org/jira/browse/FLINK-146;>FLINK-146] - Sorted output not working + +[https://issues.apache.org/jira/browse/FLINK-1982;>FLINK-1982] - Remove dependencies on Record for Flink runtime and core + +[https://issues.apache.org/jira/browse/FLINK-2676;>FLINK-2676] - Add abstraction for keyed window state + +[https://issues.apache.org/jira/browse/FLINK-2680;>FLINK-2680] - Create a dedicated aligned-event time window operator + +[https://issues.apache.org/jira/browse/FLINK-2730;>FLINK-2730] - Add CPU/Network utilization graphs to new web dashboard + +[https://issues.apache.org/jira/browse/FLINK-2732;>FLINK-2732] - Add access to the TaskManagers log file and out file in the web dashboard. + +[https://issues.apache.org/jira/browse/FLINK-2850;>FLINK-2850] - Limit the types of jobs which can run in detached mode + +[https://issues.apache.org/jira/browse/FLINK-2853;>FLINK-2853] - Apply JMH on MutableHashTablePerformanceBenchmark class. + +[https://issues.apache.org/jira/browse/FLINK-2869;>FLINK-2869] - Apply JMH on IOManagerPerformanceBenchmark class. + +[https://issues.apache.org/jira/browse/FLINK-2889;>FLINK-2889] - Apply JMH on LongSerializationSpeedBenchmark class + +[https://issues.apache.org/jira/browse/FLINK-2890;>FLINK-2890] - Apply JMH on StringSerializationSpeedBenchmark class. + +[https://issues.apache.org/jira/browse/FLINK-2900;>FLINK-2900] - Remove Record-API dependencies from Hadoop Compat module + +[https://issues.apache.org/jira/browse/FLINK-2901;>FLINK-2901] - Several flink-test ITCases depend on Record API features + +[https://issues.apache.org/jira/browse/FLINK-2906;>FLINK-2906] - Remove Record-API + +[https://issues.apache.org/jira/browse/FLINK-2919;>FLINK-2919] - Apply JMH on FieldAccessMinibenchmark class. + +[https://issues.apache.org/jira/browse/FLINK-2920;>FLINK-2920] - Apply JMH on KryoVersusAvroMinibenchmark class. + +[https://issues.apache.org/jira/browse/FLINK-2933;>FLINK-2933] - Flink scala libraries exposed with maven should carry scala version +
[05/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/js/dagre-d3.js -- diff --git a/content/visualizer/js/dagre-d3.js b/content/visualizer/js/dagre-d3.js new file mode 100755 index 000..482ce82 --- /dev/null +++ b/content/visualizer/js/dagre-d3.js @@ -0,0 +1,4560 @@ +;(function e(t,n,r){function s(o,u){if(!n[o]){if(!t[o]){var a=typeof require=="function"&if(!u&)return a(o,!0);if(i)return i(o,!0);throw new Error("Cannot find module '"+o+"'")}var f=n[o]={exports:{}};t[o][0].call(f.exports,function(e){var n=t[o][1][e];return s(n?n:e)},f,f.exports,e,t,n,r)}return n[o].exports}var i=typeof require=="function"&for(var o=0;o
[21/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/svg/white_filled.svg -- diff --git a/content/img/logo/svg/white_filled.svg b/content/img/logo/svg/white_filled.svg new file mode 100755 index 000..28bd9f9 --- /dev/null +++ b/content/img/logo/svg/white_filled.svg @@ -0,0 +1,231 @@ + + +http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd;> +http://www.w3.org/2000/svg; xmlns:xlink="http://www.w3.org/1999/xlink; x="0px" y="0px" +width="2000px" height="1280px" viewBox="0 0 2000 1280" enable-background="new 0 0 2000 1280" xml:space="preserve"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Filled white + http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/managed-state.png -- diff --git a/content/img/managed-state.png b/content/img/managed-state.png new file mode 100755 index 000..3dbb2fd Binary files /dev/null and b/content/img/managed-state.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/memory_heap_division.png -- diff --git a/content/img/memory_heap_division.png b/content/img/memory_heap_division.png new file mode 100644 index 000..2b4c2e2 Binary files /dev/null and b/content/img/memory_heap_division.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/navbar-brand-logo.jpg -- diff --git a/content/img/navbar-brand-logo.jpg b/content/img/navbar-brand-logo.jpg new file mode 100755 index 000..5993ee8 Binary files /dev/null and b/content/img/navbar-brand-logo.jpg differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/navbar-brand-logo.png -- diff --git a/content/img/navbar-brand-logo.png b/content/img/navbar-brand-logo.png new file mode 100644 index 000..152f74e Binary files /dev/null and b/content/img/navbar-brand-logo.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/one_runtime.png -- diff --git a/content/img/one_runtime.png b/content/img/one_runtime.png new file mode 100644 index 000..9cb4363 Binary files /dev/null and b/content/img/one_runtime.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/optimizer_choice.png -- diff --git a/content/img/optimizer_choice.png b/content/img/optimizer_choice.png new file mode 100644 index 000..1f8004b Binary files /dev/null and b/content/img/optimizer_choice.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/otto-group-logo.jpg -- diff --git a/content/img/otto-group-logo.jpg b/content/img/otto-group-logo.jpg new file mode 100644 index 000..f578af6 Binary files /dev/null and b/content/img/otto-group-logo.jpg differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/out_of_order_stream.png -- diff --git a/content/img/out_of_order_stream.png b/content/img/out_of_order_stream.png new file mode 100644 index 000..20ad09d Binary files /dev/null and b/content/img/out_of_order_stream.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/parallel_dataflows.png -- diff --git a/content/img/parallel_dataflows.png b/content/img/parallel_dataflows.png new file mode 100644 index 000..dbd83b2 Binary files /dev/null and b/content/img/parallel_dataflows.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/poweredby/alibaba-logo.png -- diff --git
[12/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2016/03/08/release-1.0.0.html -- diff --git a/content/news/2016/03/08/release-1.0.0.html b/content/news/2016/03/08/release-1.0.0.html new file mode 100644 index 000..13313d8 --- /dev/null +++ b/content/news/2016/03/08/release-1.0.0.html @@ -0,0 +1,319 @@ + + + + + + + +Apache Flink: Announcing Apache Flink 1.0.0 + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + + + Announcing Apache Flink 1.0.0 + + +08 Mar 2016 + +The Apache Flink community is pleased to announce the availability of the 1.0.0 release. The community put significant effort into improving and extending Apache Flink since the last release, focusing on improving the experience of writing and executing data stream processing pipelines in production. + + + + + +Flink version 1.0.0 marks the beginning of the 1.X.X series of releases, which will maintain backwards compatibility with 1.0.0. This means that applications written against stable APIs of Flink 1.0.0 will compile and run with all Flink versions in the 1. series. This is the first time we are formally guaranteeing compatibility in Flinkâs history, and we therefore see this release as a major milestone of the project, perhaps the most important since graduation as a top-level project. + +Apart from backwards compatibility, Flink 1.0.0 brings a variety of new user-facing features, as well as tons of bug fixes. About 64 contributors provided bug fixes, improvements, and new features such that in total more than 450 JIRA issues could be resolved. + +We encourage everyone to http://flink.apache.org/downloads.html;>download the release and https://ci.apache.org/projects/flink/flink-docs-release-1.0/;>check out the documentation. Feedback through the Flink http://flink.apache.org/community.html#mailing-lists;>mailing lists is, as always, very welcome! + +Interface stability annotations + +Flink 1.0.0 introduces interface stability annotations for API classes and methods. Interfaces defined as @Public are guaranteed to remain stable across all releases of the 1.x series. The @PublicEvolving annotation marks API features that may be subject to change in future versions. + +Flinkâs stability annotations will help users to implement applications that compile and execute unchanged against future versions of Flink 1.x. This greatly reduces the complexity for users when upgrading to a newer Flink release. + +Out-of-core state support + +Flink 1.0.0 adds a new state backend that uses RocksDB to store state (both windows and user-defined key-value state). http://rocksdb.org/;>RocksDB is an embedded key/value store database, originally developed by Facebook. +When using this backend, active state in streaming programs can grow well beyond memory. The RocksDB files are stored in a distributed file system such as HDFS or S3 for backups. + +Savepoints and version upgrades + +Savepoints are checkpoints of the state of a running streaming job that can be manually
[14/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/09/16/off-heap-memory.html -- diff --git a/content/news/2015/09/16/off-heap-memory.html b/content/news/2015/09/16/off-heap-memory.html new file mode 100644 index 000..60d7bda --- /dev/null +++ b/content/news/2015/09/16/off-heap-memory.html @@ -0,0 +1,1083 @@ + + + + + + + +Apache Flink: Off-heap Memory in Apache Flink and the curious JIT compiler + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + + + Off-heap Memory in Apache Flink and the curious JIT compiler + + +16 Sep 2015 by Stephan Ewen (https://twitter.com/stephanewen;>@stephanewen) + +Running data-intensive code in the JVM and making it well-behaved is tricky. Systems that put billions of data objects naively onto the JVM heap face unpredictable OutOfMemoryErrors and Garbage Collection stalls. Of course, you still want to to keep your data in memory as much as possible, for speed and responsiveness of the processing applications. In that context, âoff-heapâ has become almost something like a magic word to solve these problems. + +In this blog post, we will look at how Flink exploits off-heap memory. The feature is part of the upcoming release, but you can try it out with the latest nightly builds. We will also give a few interesting insights into the behavior for Javaâs JIT compiler for highly optimized methods and loops. + +Recap: Memory Management in Flink + +To understand Flinkâs approach to off-heap memory, we need to recap Flinkâs approach to custom managed memory. We have written an earlier blog post about how Flink manages JVM memory itself + +As a summary, the core part is that Flink implements its algorithms not against Java objects, arrays, or lists, but actually against a data structure similar to java.nio.ByteBuffer. Flink uses its own specialized version, called https://github.com/apache/flink/blob/release-0.9.1-rc1/flink-core/src/main/java/org/apache/flink/core/memory/MemorySegment.java;>MemorySegment on which algorithms put and get at specific positions ints, longs, byte arrays, etc, and compare and copy memory. The memory segments are held and distributed by a central component (called MemoryManager) from which algorithms request segments according to their calculated memory budgets. + +Donât believe that this can be fast? Have a look at the benchmarks in the earlier blogpost, which show that it is actually often much faster than working on objects, due to better control over data layout (cache efficiency, data size), and reducing the pressure on Javaâs Garbage Collector. + +This form of memory management has been in Flink for a long time. Anecdotally, the first public demo of Flinkâs predecessor project Stratosphere, at the VLDB conference in 2010, was running its programs with custom managed memory (although I believe few attendees were aware of that). + +Why actually bother with off-heap memory? + +Given that Flink has a sophisticated level of managing on-heap memory,
[29/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/commit-stats.png -- diff --git a/content/img/blog/commit-stats.png b/content/img/blog/commit-stats.png new file mode 100644 index 000..dfeb7b3 Binary files /dev/null and b/content/img/blog/commit-stats.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/community-growth.png -- diff --git a/content/img/blog/community-growth.png b/content/img/blog/community-growth.png new file mode 100644 index 000..e37df51 Binary files /dev/null and b/content/img/blog/community-growth.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/data-serialization.png -- diff --git a/content/img/blog/data-serialization.png b/content/img/blog/data-serialization.png new file mode 100755 index 000..80667f6 Binary files /dev/null and b/content/img/blog/data-serialization.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/emr-firefoxsettings.png -- diff --git a/content/img/blog/emr-firefoxsettings.png b/content/img/blog/emr-firefoxsettings.png new file mode 100755 index 000..cc515a4 Binary files /dev/null and b/content/img/blog/emr-firefoxsettings.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/emr-hadoopversion.png -- diff --git a/content/img/blog/emr-hadoopversion.png b/content/img/blog/emr-hadoopversion.png new file mode 100755 index 000..7646b75 Binary files /dev/null and b/content/img/blog/emr-hadoopversion.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/emr-jobmanager.png -- diff --git a/content/img/blog/emr-jobmanager.png b/content/img/blog/emr-jobmanager.png new file mode 100755 index 000..4cbe5cc Binary files /dev/null and b/content/img/blog/emr-jobmanager.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/emr-running.png -- diff --git a/content/img/blog/emr-running.png b/content/img/blog/emr-running.png new file mode 100755 index 000..44d21d2 Binary files /dev/null and b/content/img/blog/emr-running.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/emr-security.png -- diff --git a/content/img/blog/emr-security.png b/content/img/blog/emr-security.png new file mode 100755 index 000..fee4fed Binary files /dev/null and b/content/img/blog/emr-security.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/emr-yarnappmaster.png -- diff --git a/content/img/blog/emr-yarnappmaster.png b/content/img/blog/emr-yarnappmaster.png new file mode 100755 index 000..62be9ed Binary files /dev/null and b/content/img/blog/emr-yarnappmaster.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/feature-timeline.png -- diff --git a/content/img/blog/feature-timeline.png b/content/img/blog/feature-timeline.png new file mode 100644 index 000..1f56747 Binary files /dev/null and b/content/img/blog/feature-timeline.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/ff-speakers.png -- diff --git a/content/img/blog/ff-speakers.png b/content/img/blog/ff-speakers.png new file mode 100644 index 000..429a2c8 Binary files /dev/null and b/content/img/blog/ff-speakers.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/flink-1.0.png -- diff --git a/content/img/blog/flink-1.0.png b/content/img/blog/flink-1.0.png new file mode 100644 index 000..88a006b Binary files /dev/null and b/content/img/blog/flink-1.0.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/flink-dow-2016.png -- diff --git a/content/img/blog/flink-dow-2016.png b/content/img/blog/flink-dow-2016.png new file mode 100644 index 000..c90d768 Binary files /dev/null and b/content/img/blog/flink-dow-2016.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/flink-forward-banner.png
[26/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/colors/flink_colors.pptx -- diff --git a/content/img/logo/colors/flink_colors.pptx b/content/img/logo/colors/flink_colors.pptx new file mode 100644 index 000..8ecd0be Binary files /dev/null and b/content/img/logo/colors/flink_colors.pptx differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/100/flink_squirrel_100_black.png -- diff --git a/content/img/logo/png/100/flink_squirrel_100_black.png b/content/img/logo/png/100/flink_squirrel_100_black.png new file mode 100755 index 000..6869f58 Binary files /dev/null and b/content/img/logo/png/100/flink_squirrel_100_black.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/100/flink_squirrel_100_color.png -- diff --git a/content/img/logo/png/100/flink_squirrel_100_color.png b/content/img/logo/png/100/flink_squirrel_100_color.png new file mode 100755 index 000..c508e1e Binary files /dev/null and b/content/img/logo/png/100/flink_squirrel_100_color.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/100/flink_squirrel_100_white.png -- diff --git a/content/img/logo/png/100/flink_squirrel_100_white.png b/content/img/logo/png/100/flink_squirrel_100_white.png new file mode 100755 index 000..088fb27 Binary files /dev/null and b/content/img/logo/png/100/flink_squirrel_100_white.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink1000_black.png -- diff --git a/content/img/logo/png/1000/flink1000_black.png b/content/img/logo/png/1000/flink1000_black.png new file mode 100755 index 000..31af663 Binary files /dev/null and b/content/img/logo/png/1000/flink1000_black.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink1000_color_black.png -- diff --git a/content/img/logo/png/1000/flink1000_color_black.png b/content/img/logo/png/1000/flink1000_color_black.png new file mode 100755 index 000..2f4b991 Binary files /dev/null and b/content/img/logo/png/1000/flink1000_color_black.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink1000_color_white.png -- diff --git a/content/img/logo/png/1000/flink1000_color_white.png b/content/img/logo/png/1000/flink1000_color_white.png new file mode 100755 index 000..cd0b56e Binary files /dev/null and b/content/img/logo/png/1000/flink1000_color_white.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink1000_white.png -- diff --git a/content/img/logo/png/1000/flink1000_white.png b/content/img/logo/png/1000/flink1000_white.png new file mode 100755 index 000..3b7b253 Binary files /dev/null and b/content/img/logo/png/1000/flink1000_white.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink_squirrel_1000.png -- diff --git a/content/img/logo/png/1000/flink_squirrel_1000.png b/content/img/logo/png/1000/flink_squirrel_1000.png new file mode 100755 index 000..37c980e Binary files /dev/null and b/content/img/logo/png/1000/flink_squirrel_1000.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink_squirrel_black_1000.png -- diff --git a/content/img/logo/png/1000/flink_squirrel_black_1000.png b/content/img/logo/png/1000/flink_squirrel_black_1000.png new file mode 100755 index 000..b68621d Binary files /dev/null and b/content/img/logo/png/1000/flink_squirrel_black_1000.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/1000/flink_squirrel_white_1000.png -- diff --git a/content/img/logo/png/1000/flink_squirrel_white_1000.png b/content/img/logo/png/1000/flink_squirrel_white_1000.png new file mode 100755 index 000..14c7350 Binary files /dev/null and b/content/img/logo/png/1000/flink_squirrel_white_1000.png differ http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/png/200/flink2_200_black.png -- diff --git
[08/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/css/bootstrap.css -- diff --git a/content/visualizer/css/bootstrap.css b/content/visualizer/css/bootstrap.css new file mode 100755 index 000..fc41022 --- /dev/null +++ b/content/visualizer/css/bootstrap.css @@ -0,0 +1,5785 @@ +/*! + * Bootstrap v3.1.1 (http://getbootstrap.com) + * Copyright 2011-2014 Twitter, Inc. + * Licensed under MIT (https://github.com/twbs/bootstrap/blob/master/LICENSE) + */ + +/*! normalize.css v3.0.0 | MIT License | git.io/normalize */ +html { + font-family: sans-serif; + -webkit-text-size-adjust: 100%; + -ms-text-size-adjust: 100%; +} +body { + margin: 0; +} +article, +aside, +details, +figcaption, +figure, +footer, +header, +hgroup, +main, +nav, +section, +summary { + display: block; +} +audio, +canvas, +progress, +video { + display: inline-block; + vertical-align: baseline; +} +audio:not([controls]) { + display: none; + height: 0; +} +[hidden], +template { + display: none; +} +a { + background: transparent; +} +a:active, +a:hover { + outline: 0; +} +abbr[title] { + border-bottom: 1px dotted; +} +b, +strong { + font-weight: bold; +} +dfn { + font-style: italic; +} +h1 { + margin: .67em 0; + font-size: 2em; +} +mark { + color: #000; + background: #ff0; +} +small { + font-size: 80%; +} +sub, +sup { + position: relative; + font-size: 75%; + line-height: 0; + vertical-align: baseline; +} +sup { + top: -.5em; +} +sub { + bottom: -.25em; +} +img { + border: 0; +} +svg:not(:root) { + overflow: hidden; +} +figure { + margin: 1em 40px; +} +hr { + height: 0; + -moz-box-sizing: content-box; + box-sizing: content-box; +} +pre { + overflow: auto; +} +code, +kbd, +pre, +samp { + font-family: monospace, monospace; + font-size: 1em; +} +button, +input, +optgroup, +select, +textarea { + margin: 0; + font: inherit; + color: inherit; +} +button { + overflow: visible; +} +button, +select { + text-transform: none; +} +button, +html input[type="button"], +input[type="reset"], +input[type="submit"] { + -webkit-appearance: button; + cursor: pointer; +} +button[disabled], +html input[disabled] { + cursor: default; +} +button::-moz-focus-inner, +input::-moz-focus-inner { + padding: 0; + border: 0; +} +input { + line-height: normal; +} +input[type="checkbox"], +input[type="radio"] { + box-sizing: border-box; + padding: 0; +} +input[type="number"]::-webkit-inner-spin-button, +input[type="number"]::-webkit-outer-spin-button { + height: auto; +} +input[type="search"] { + -webkit-box-sizing: content-box; + -moz-box-sizing: content-box; + box-sizing: content-box; + -webkit-appearance: textfield; +} +input[type="search"]::-webkit-search-cancel-button, +input[type="search"]::-webkit-search-decoration { + -webkit-appearance: none; +} +fieldset { + padding: .35em .625em .75em; + margin: 0 2px; + border: 1px solid #c0c0c0; +} +legend { + padding: 0; + border: 0; +} +textarea { + overflow: auto; +} +optgroup { + font-weight: bold; +} +table { + border-spacing: 0; + border-collapse: collapse; +} +td, +th { + padding: 0; +} +@media print { + * { +color: #000 !important; +text-shadow: none !important; +background: transparent !important; +box-shadow: none !important; + } + a, + a:visited { +text-decoration: underline; + } + a[href]:after { +content: " (" attr(href) ")"; + } + abbr[title]:after { +content: " (" attr(title) ")"; + } + a[href^="javascript:"]:after, + a[href^="#"]:after { +content: ""; + } + pre, + blockquote { +border: 1px solid #999; + +page-break-inside: avoid; + } + thead { +display: table-header-group; + } + tr, + img { +page-break-inside: avoid; + } + img { +max-width: 100% !important; + } + p, + h2, + h3 { +orphans: 3; +widows: 3; + } + h2, + h3 { +page-break-after: avoid; + } + select { +background: #fff !important; + } + .navbar { +display: none; + } + .table td, + .table th { +background-color: #fff !important; + } + .btn > .caret, + .dropup > .btn > .caret { +border-top-color: #000 !important; + } + .label { +border: 1px solid #000; + } + .table { +border-collapse: collapse !important; + } + .table-bordered th, + .table-bordered td { +border: 1px solid #ddd !important; + } +} +* { + -webkit-box-sizing: border-box; + -moz-box-sizing: border-box; + box-sizing: border-box; +} +*:before, +*:after { + -webkit-box-sizing: border-box; + -moz-box-sizing: border-box; + box-sizing: border-box; +} +html { + font-size: 62.5%; + + -webkit-tap-highlight-color: rgba(0, 0, 0, 0); +} +body { + font-family: "Helvetica Neue", Helvetica, Arial, sans-serif; + font-size: 14px; + line-height: 1.42857143; + color: #333; + background-color: #fff; +} +input, +button, +select, +textarea { + font-family: inherit; +
[32/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/faq.html -- diff --git a/content/faq.html b/content/faq.html new file mode 100644 index 000..7653ff8 --- /dev/null +++ b/content/faq.html @@ -0,0 +1,683 @@ + + + + + + + +Apache Flink: Frequently Asked Questions (FAQ) + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + +Frequently Asked Questions (FAQ) + + + +The following questions are frequently asked with regard to the Flink project in general. If you have further questions, make sure to consult the documentation or ask the community. + + + + General + Is Flink a Hadoop Project? + Do I have to install Apache Hadoop to use Flink? + + + Usage + How do I assess the progress of a Flink program? + How can I figure out why a program failed? + How do I debug Flink programs? + What is the parallelism? How do I set it? + + + Errors + Why am I getting a âNonSerializableExceptionâ ? + In Scala API, I get an error about implicit values and evidence parameters + I get an error message saying that not enough buffers are available. How do I fix this? + My job fails early with a java.io.EOFException. What could be the cause? + My job fails with various exceptions from the HDFS/Hadoop code. What can I do? + In Eclipse, I get compilation errors in the Scala projects + My program does not compute the correct result. Why are my custom key types + I get a java.lang.InstantiationException for my data type, what is wrong? + I canât stop Flink with the provided stop-scripts. What can I do? + I got an OutOfMemoryException. What can I do? + Why do the TaskManager log files become so huge? + The slot allocated for my task manager has been released. What should I do? + + + YARN Deployment + The YARN session runs only for a few seconds + My YARN containers are killed because they use too much memory + The YARN session crashes with a HDFS permission exception during startup + My job is not reacting to a job cancellation? + + + Features + What kind of fault-tolerance does Flink provide? + Are Hadoop-like utilities, such as Counters and the DistributedCache supported? + + + + + + +General + +Is Flink a Hadoop Project? + +Flink is a data processing system and an alternative to Hadoopâs +MapReduce component. It comes with its own runtime rather than building on top +of MapReduce. As such, it can work completely independently of the Hadoop +ecosystem. However, Flink can also access Hadoopâs distributed file +system (HDFS) to read and write data, and Hadoopâs next-generation resource +manager (YARN) to provision cluster resources. Since most Flink users are +using Hadoop HDFS to store their data, Flink already ships the required libraries to +access HDFS. + +Do I have to install Apache Hadoop to use Flink? + +No. Flink can run without a Hadoop installation. However, a very
[13/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/12/04/Introducing-windows.html -- diff --git a/content/news/2015/12/04/Introducing-windows.html b/content/news/2015/12/04/Introducing-windows.html new file mode 100644 index 000..aa0834f --- /dev/null +++ b/content/news/2015/12/04/Introducing-windows.html @@ -0,0 +1,349 @@ + + + + + + + +Apache Flink: Introducing Stream Windows in Apache Flink + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + + + Introducing Stream Windows in Apache Flink + + +04 Dec 2015 by Fabian Hueske (https://twitter.com/fhueske;>@fhueske) + +The data analysis space is witnessing an evolution from batch to stream processing for many use cases. Although batch can be handled as a special case of stream processing, analyzing never-ending streaming data often requires a shift in the mindset and comes with its own terminology (for example, âwindowingâ and âat-least-onceâ/âexactly-onceâ processing). This shift and the new terminology can be quite confusing for people being new to the space of stream processing. Apache Flink is a production-ready stream processor with an easy-to-use yet very expressive API to define advanced stream analysis programs. Flinkâs API features very flexible window definitions on data streams which let it stand out among other open source stream processors. + +In this blog post, we discuss the concept of windows for stream processing, present Flinkâs built-in windows, and explain its support for custom windowing semantics. + +What are windows and what are they good for? + +Consider the example of a traffic sensor that counts every 15 seconds the number of vehicles passing a certain location. The resulting stream could look like: + + + + + +If you would like to know, how many vehicles passed that location, you would simply sum the individual counts. However, the nature of a sensor stream is that it continuously produces data. Such a stream never ends and it is not possible to compute a final sum that can be returned. Instead, it is possible to compute rolling sums, i.e., return for each input event an updated sum record. This would yield a new stream of partial sums. + + + + + +However, a stream of partial sums might not be what we are looking for, because it constantly updates the count and even more important, some information such as variation over time is lost. Hence, we might want to rephrase our question and ask for the number of cars that pass the location every minute. This requires us to group the elements of the stream into finite sets, each set corresponding to sixty seconds. This operation is called a tumbling windows operation. + + + + + +Tumbling windows discretize a stream into non-overlapping windows. For certain applications it is important that windows are not disjunct because an application might require smoothed aggregates. For example, we can compute every thirty seconds the number of cars passed in the last minute. Such windows
[24/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/svg/color_white.svg -- diff --git a/content/img/logo/svg/color_white.svg b/content/img/logo/svg/color_white.svg new file mode 100755 index 000..e24ac1c --- /dev/null +++ b/content/img/logo/svg/color_white.svg @@ -0,0 +1,1563 @@ + + +http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd;> +http://www.w3.org/2000/svg; xmlns:xlink="http://www.w3.org/1999/xlink; x="0px" y="0px" +width="2000px" height="1280px" viewBox="0 0 2000 1280" enable-background="new 0 0 2000 1280" xml:space="preserve"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
[15/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2015/08/24/introducing-flink-gelly.html -- diff --git a/content/news/2015/08/24/introducing-flink-gelly.html b/content/news/2015/08/24/introducing-flink-gelly.html new file mode 100644 index 000..714e75a --- /dev/null +++ b/content/news/2015/08/24/introducing-flink-gelly.html @@ -0,0 +1,649 @@ + + + + + + + +Apache Flink: Introducing Gelly: Graph Processing with Apache Flink + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + + + Introducing Gelly: Graph Processing with Apache Flink + + +24 Aug 2015 + +This blog post introduces Gelly, Apache Flinkâs graph-processing API and library. Flinkâs native support +for iterations makes it a suitable platform for large-scale graph analytics. +By leveraging delta iterations, Gelly is able to map various graph processing models such as +vertex-centric or gather-sum-apply to Flink dataflows. + +Gelly allows Flink users to perform end-to-end data analysis in a single system. +Gelly can be seamlessly used with Flinkâs DataSet API, +which means that pre-processing, graph creation, analysis, and post-processing can be done +in the same application. At the end of this post, we will go through a step-by-step example +in order to demonstrate that loading, transformation, filtering, graph creation, and analysis +can be performed in a single Flink program. + +Overview + + + What is Gelly? + Graph Representation and Creation + Transformations and Utilities + Iterative Graph Processing + Library of Graph Algorithms + Use-Case: Music Profiles + Ongoing and Future Work + + + + +What is Gelly? + +Gelly is a Graph API for Flink. It is currently supported in both Java and Scala. +The Scala methods are implemented as wrappers on top of the basic Java operations. +The API contains a set of utility functions for graph analysis, supports iterative graph +processing and introduces a library of graph algorithms. + + + + + +Back to top + +Graph Representation and Creation + +In Gelly, a graph is represented by a DataSet of vertices and a DataSet of edges. +A vertex is defined by its unique ID and a value, whereas an edge is defined by its source ID, +target ID, and value. A vertex or edge for which a value is not specified will simply have the +value type set to NullValue. + +A graph can be created from: + + + DataSet of edges and an optional DataSet of vertices using Graph.fromDataSet() + DataSet of Tuple3 and an optional DataSet of Tuple2 using Graph.fromTupleDataSet() + Collection of edges and an optional Collection of vertices using Graph.fromCollection() + + +In all three cases, if the vertices are not provided, +Gelly will automatically produce the vertex IDs from the edge source and target IDs. + +Back to top + +Transformations and Utilities + +These are methods of the Graph class and include common graph metrics, transformations +and mutations as well as neighborhood aggregations. + +Common Graph Metrics +These methods can be used to
[22/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/logo/svg/flink_logotypes.svg -- diff --git a/content/img/logo/svg/flink_logotypes.svg b/content/img/logo/svg/flink_logotypes.svg new file mode 100755 index 000..006243a --- /dev/null +++ b/content/img/logo/svg/flink_logotypes.svg @@ -0,0 +1,3835 @@ + + +http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd;> +http://www.w3.org/2000/svg; xmlns:xlink="http://www.w3.org/1999/xlink; x="0px" y="0px" +width="799.736px" height="803.173px" viewBox="168.959 68.626 799.736 803.173" +enable-background="new 168.959 68.626 799.736 803.173" xml:space="preserve"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
[30/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/cep-monitoring.svg -- diff --git a/content/img/blog/cep-monitoring.svg b/content/img/blog/cep-monitoring.svg new file mode 100644 index 000..02cca81 --- /dev/null +++ b/content/img/blog/cep-monitoring.svg @@ -0,0 +1,2838 @@ + + + +http://purl.org/dc/elements/1.1/; + xmlns:cc="http://creativecommons.org/ns#; + xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#; + xmlns:svg="http://www.w3.org/2000/svg; + xmlns="http://www.w3.org/2000/svg; + xmlns:xlink="http://www.w3.org/1999/xlink; + xmlns:sodipodi="http://sodipodi.sourceforge.net/DTD/sodipodi-0.dtd; + xmlns:inkscape="http://www.inkscape.org/namespaces/inkscape; + width="210mm" + height="148mm" + viewBox="0 0 744.09449 524.40944" + id="svg2" + version="1.1" + inkscape:version="0.91 r13725" + sodipodi:docname="CEP_final.svg"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +image/svg+xml +http://purl.org/dc/dcmitype/StillImage; /> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/img/blog/code-growth.png -- diff --git a/content/img/blog/code-growth.png b/content/img/blog/code-growth.png new file mode 100644 index 000..196c907 Binary files /dev/null and b/content/img/blog/code-growth.png differ
[flink-web] Git Push Summary [forced push!] [Forced Update!]
Repository: flink-web Updated Branches: refs/heads/asf-site 9ec0a879f -> 4bb81d3f1 (forced update)
[31/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/how-to-contribute.html -- diff --git a/content/how-to-contribute.html b/content/how-to-contribute.html new file mode 100644 index 000..87ad4a0 --- /dev/null +++ b/content/how-to-contribute.html @@ -0,0 +1,345 @@ + + + + + + + +Apache Flink: How To Contribute + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + +How To Contribute + + Apache Flink is developed by an open and friendly community. Everybody is cordially welcome to join the community and contribute to Apache Flink. There are several ways to interact with the community and to contribute to Flink including asking questions, filing bug reports, proposing new features, joining discussions on the mailing lists, contributing code or documentation, improving the website, or testing release candidates. + + + + Ask questions! + File a bug report + Propose an improvement or a new feature + Help others and join the discussions + Test a release candidate + Contribute code + Contribute documentation + Improve the website + More ways to contribute⦠+ Submit a Contributor License Agreement + How to become a committer + + + + +Ask questions! + +The Apache Flink community is eager to help and to answer your questions. We have a user mailing list, hang out in an IRC channel, and watch Stack Overflow on the http://stackoverflow.com/questions/tagged/apache-flink;>[apache-flink] tag. + + + +File a bug report + +Please let us know if you experienced a problem with Flink and file a bug report. Open http://issues.apache.org/jira/browse/FLINK;>Flinkâs JIRA and click on the blue Create button at the top. Please give detailed information about the problem you encountered and, if possible, add a description that helps to reproduce the problem. Thank you very much. + + + +Propose an improvement or a new feature + +Our community is constantly looking for feedback to improve Apache Flink. If you have an idea how to improve Flink or have a new feature in mind that would be beneficial for Flink users, please open an issue in http://issues.apache.org/jira/browse/FLINK;>Flinkâs JIRA. The improvement or new feature should be described in appropriate detail and include the scope and its requirements if possible. Detailed information is important for a few reasons: + + + It ensures your requirements are met when the improvement or feature is implemented. + It helps to estimate the effort and to design a solution that addresses your needs. + It allow for constructive discussions that might arise around this issue. + + +Detailed information is also required, if you plan to contribute the improvement or feature you proposed yourself. Please read the Contribute code guide in this case as well. + +We recommend to first reach consensus with the community on whether a new feature is required and how to implement a new feature, before starting with the implementation. Some features might be out of scope of the project, and itâs best to
[11/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/news/2016/05/24/stream-sql.html -- diff --git a/content/news/2016/05/24/stream-sql.html b/content/news/2016/05/24/stream-sql.html new file mode 100644 index 000..f22a063 --- /dev/null +++ b/content/news/2016/05/24/stream-sql.html @@ -0,0 +1,321 @@ + + + + + + + +Apache Flink: Stream Processing for Everyone with SQL and Apache Flink + + + + +https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> + + + + + + + + +https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Download Flink + + +Home + + +Introduction to Flink + + +Flink Use Cases + + +Powered by Flink + + +Ecosystem + + +Community Project Info + + +How to Contribute + + +Flink Blog + + + + + + + + + Documentation + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) + http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) + + + + + + http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart + + + + + https://github.com/apache/flink; target="_blank">Flink on GitHub + + + + + + + + + + + + + + + +Project FAQ + + +https://twitter.com/apacheflink; target="_blank">@ApacheFlink + + +Plan Visualizer + + + + + + + + + + + Stream Processing for Everyone with SQL and Apache Flink + + +24 May 2016 by Fabian Hueske (https://twitter.com/fhueske;>@fhueske) + +The capabilities of open source systems for distributed stream processing have evolved significantly over the last years. Initially, the first systems in the field (notably https://storm.apache.org;>Apache Storm) provided low latency processing, but were limited to at-least-once guarantees, processing-time semantics, and rather low-level APIs. Since then, several new systems emerged and pushed the state of the art of open source stream processing in several dimensions. Today, users of Apache Flink or https://beam.incubator.apache.org;>Apache Beam can use fluent Scala and Java APIs to implement stream processing jobs that operate in event-time with exactly-once semantics at high throughput and low latency. + +In the meantime, stream processing has taken off in the industry. We are witnessing a rapidly growing interest in stream processing which is reflected by prevalent deployments of streaming processing infrastructure such as https://kafka.apache.org;>Apache Kafka and Apache Flink. The increasing number of available data streams results in a demand for people that can analyze streaming data and turn it into real-time insights. However, stream data analysis requires a special skill set including knowledge of streaming concepts such as the characteristics of unbounded streams, windows, time, and state as well as the skills to implement stream analysis jobs usually against Java or Scala APIs. People with this skill set are rare and hard to find. + +About six months ago, the Apache Flink community started an effort to add a SQL interface for stream data analysis. SQL is the standard language to access and process data. Everybody who occasionally analyzes data is familiar with SQL. Consequently, a SQL interface for stream data processing will make this technology accessible to a much wider audience. Moreover, SQL support for streaming data will also enable new use cases such as interactive and ad-hoc stream analysis and significantly simplify many applications including stream ingestion and simple transformations. In this blog post, we report on the current status, architectural design, and future plans of the Apache Flink community to implement support for SQL as a language for analyzing data streams. + +Where did we come from? + +With the http://flink.apache.org/news/2015/04/13/release-0.9.0-milestone1.html;>0.9.0-milestone1 release, Apache Flink added an API
[07/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/css/graph.css -- diff --git a/content/visualizer/css/graph.css b/content/visualizer/css/graph.css new file mode 100755 index 000..f4365cc --- /dev/null +++ b/content/visualizer/css/graph.css @@ -0,0 +1,48 @@ +/*! + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +g.type-TK > rect { + fill: #00ffd0; +} + +svg { + border: 0px solid #999; + overflow: hidden; +} + +text { + font-weight: 300; + font-family: "Helvetica Neue", Helvetica, Arial, sans-serf; + font-size: 14px; +} + +.node rect { + stroke: #999; + stroke-width: 0px; + fill: #fff; +} + +.edgeLabel rect { + fill: #fff; +} + +.edgePath path { + stroke: #333; + stroke-width: 3px; + fill: none; +} http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/visualizer/css/nephelefrontend.css -- diff --git a/content/visualizer/css/nephelefrontend.css b/content/visualizer/css/nephelefrontend.css new file mode 100755 index 000..3ca1837 --- /dev/null +++ b/content/visualizer/css/nephelefrontend.css @@ -0,0 +1,198 @@ +/*! + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +* +{ + padding: 0px; + margin: 0px; +} + +html, body +{ + height: 100%; + position: relative; + overflow: auto; +} + +h1 +{ + font-family: verdana; + font-size: 36px; + font-weight: normal; + font-decoration: none; + font-style: normal; + font-variant: small-caps; +} + +h1 img { + vertical-align: middle; + margin-right: 50px; +} + +h3 +{ + font-family: verdana; + font-size: 20px; + font-weight: normal; + font-decoration: none; + font-style: italic; + font-variant: normal; + text-align: center; + color: #66; + margin: 10px 10px 20px 10px; +} + +h4 +{ + font-family: Verdana; + font-size: 16px; + font-weight: normal; + font-decoration: none; + font-style: italic; + font-variant: normal; + text-align: left; + color: #00; + margin: 5px 5px 10px 5px; + padding: 2px 2px 5px 2px; + border-bottom: 1px dashed #44; +} + +input +{ + font-family: verdana, Sans-Serif; + font-size: 13px; + + border: solid 1.5px #33; + background-color: #ee; + color: #33; + + padding: 5px; + margin: 5px; +} + +.fadedPropertiesText +{ + font-family: verdana; + font-size: 18px; + color: #AA; + font-style: italic; +} + + +.mainHeading +{ + border: 1px solid #262A37; + margin: 5px; + text-align: left; + background-color: white; + background-repeat: no-repeat; + background-position: right; + height: 100px; + overflow: hidden; +} + + +.boxed +{ + border: 1px solid #262A37; + margin: 5px; +} + +.spaced +{ + margin: 5px; +} + +.footer +{ + position: absolute; + left: 0px; + right: 0px; + bottom: 0px; +} + +.error_text +{ + font-family: verdana; + font-size: 14px; + font-weight: normal; + font-decoration: none; + font-style: italic; + font-variant: normal; + text-align: center; + color: #DF0101; +} + +.translucent +{ + -moz-opacity:0 ; +
[20/39] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/js/jquery.jcarousel.min.js -- diff --git a/content/js/jquery.jcarousel.min.js b/content/js/jquery.jcarousel.min.js new file mode 100644 index 000..7421030 --- /dev/null +++ b/content/js/jquery.jcarousel.min.js @@ -0,0 +1,4 @@ +/*! jCarousel - v0.3.4 - 2015-09-23 +* http://sorgalla.com/jcarousel/ +* Copyright (c) 2006-2015 Jan Sorgalla; Licensed MIT */ +!function(a){"use strict";var b=a.jCarousel={};b.version="0.3.4";var c=/^([+\-]=)?(.+)$/;b.parseTarget=function(a){var b=!1,d="object"!=typeof a?c.exec(a):null;return d?(a=parseInt(d[2],10)||0,d[1]&&(b=!0,"-="===d[1]&&(a*=-1))):"object"!=typeof a&&(a=parseInt(a,10)||0),{target:a,relative:b}},b.detectCarousel=function(a){for(var b;a.length>0;){if(b=a.filter("[data-jcarousel]"),b.length>0)return b;if(b=a.find("[data-jcarousel]"),b.length>0)return b;a=a.parent()}return null},b.base=function(c){return{version:b.version,_options:{},_element:null,_carousel:null,_init:a.noop,_create:a.noop,_destroy:a.noop,_reload:a.noop,create:function(){return this._element.attr("data-"+c.toLowerCase(),!0).data(c,this),!1===this._trigger("create")?this:(this._create(),this._trigger("createend"),this)},destroy:function(){return!1===this._trigger("destroy")?this:(this._destroy(),this._trigger("destroyend"),this._element.removeData(c).removeAttr("data-"+c.toLowerCase()),this)},reload:function(a){return!1===t his._trigger("reload")?this:(a&(a),this._reload(),this._trigger("reloadend"),this)},element:function(){return this._element},options:function(b,c){if(0===arguments.length)return a.extend({},this._options);if("string"==typeof b){if("undefined"==typeof c)return"undefined"==typeof this._options[b]?null:this._options[b];this._options[b]=c}else this._options=a.extend({},this._options,b);return this},carousel:function(){return this._carousel||(this._carousel=b.detectCarousel(this.options("carousel")||this._element),this._carousel||a.error('Could not detect carousel for plugin "'+c+'"')),this._carousel},_trigger:function(b,d,e){var f,g=!1;return e=[this].concat(e||[]),(d||this._element).each(function(){f=a.Event((c+":"+b).toLowerCase()),a(this).trigger(f,e),f.isDefaultPrevented()&&(g=!0)}),!g}}},b.plugin=function(c,d){var e=a[c]=function(b,c){this._element=a(b),this.options(c),this._init(),this.create()};return e.fn=e.prototype=a.extend({},b.base(c),d),a.fn[c]=function(b){var d=Array.prototype.slice.call(arguments,1),f=this;return this.each("string"==typeof b?function(){var e=a(this).data(c);if(!e)return a.error("Cannot call methods on "+c+' prior to initialization; attempted to call method "'+b+'"');if(!a.isFunction(e[b])||"_"===b.charAt(0))return a.error('No such method "'+b+'" for '+c+" instance");var g=e[b].apply(e,d);return g!==e&&"undefined"!=typeof g?(f=g,!1):void 0}:function(){var d=a(this).data(c);d instanceof e?d.reload(b):new e(this,b)}),f},e}}(jQuery),function(a,b){"use strict";var c=function(a){return parseFloat(a)||0};a.jCarousel.plugin("jcarousel",{animating:!1,tail:0,inTail:!1,resizeTimer:null,lt:null,vertical:!1,rtl:!1,circular:!1,underflow:!1,relative:!1,_options:{list:function(){return this.element().children().eq(0)},items:function(){return this.list().children()},animation:400,transitions:!1,wrap:null,vertical:null,rtl:null,center:!1},_list:null,_items:null,_target:a(),_first:a(),_last:a(),_visible:a(),_fullyvisible:a(),_init:functio n(){var a=this;return this.onWindowResize=function(){a.resizeTimer&(a.resizeTimer),a.resizeTimer=setTimeout(function(){a.reload()},100)},this},_create:function(){this._reload(),a(b).on("resize.jcarousel",this.onWindowResize)},_destroy:function(){a(b).off("resize.jcarousel",this.onWindowResize)},_reload:function(){this.vertical=this.options("vertical"),null==this.vertical&&(this.vertical=this.list().height()>this.list().width()),this.rtl=this.options("rtl"),null==this.rtl&&(this.rtl=function(b){if("rtl"===(""+b.attr("dir")).toLowerCase())return!0;var c=!1;return b.parents("[dir]").each(function(){return/rtl/i.test(a(this).attr("dir"))?(c=!0,!1):void 0}),c}(this._element)),this.lt=this.vertical?"top":"left",this.relative="relative"===this.list().css("position"),this._list=null,this._items=null;var b=this.index(this._target)>=0?this._target:this.closest();this.circular="circular"===this.options("wrap"),this.underflow=!1;var c={left:0,top:0};return b.length>0&&(this._prepar e(b),this.list().find("[data-jcarousel-clone]").remove(),this._items=null,this.underflow=this._fullyvisible.length>=this.items().length,this.circular=this.circular&&!this.underflow,c[this.lt]=this._position(b)+"px"),this.move(c),this},list:function(){if(null===this._list){var b=this.options("list");this._list=a.isFunction(b)?b.call(this):this._element.find(b)}return this._list},items:function(){if(null===this._items){var
[34/39] flink-web git commit: Rebuild site
an Gates +PMC, Committer +gates + + +https://avatars0.githubusercontent.com/u/2388347?s=50; /> +Fabian Hueske +PMC, Committer +fhueske + + +https://avatars3.githubusercontent.com/u/498957?v=3s=50; /> +Vasia Kalavri +PMC, Committer +vasia + + + +https://avatars0.githubusercontent.com/u/68551?s=50; /> +Aljoscha Krettek +PMC, Committer +aljoscha + + +https://avatars2.githubusercontent.com/u/2550549?s=50; /> +Andra Lungu +Committer +andra + + +https://avatars0.githubusercontent.com/u/89049?s=50; /> +Robert Metzger +PMC, Committer +rmetzger + + +https://avatars2.githubusercontent.com/u/837221?s=50; /> +Maximilian Michels +PMC, Committer +mxm + + +https://avatars2.githubusercontent.com/u/1941681?s=50; /> +Chiwan Park +Committer +chiwanpark + + +https://avatars1.githubusercontent.com/u/5756858?s=50; /> +Till Rohrmann +PMC, Committer +trohrmann + + +https://avatars0.githubusercontent.com/u/105434?s=50; /> +Henry Saputra +PMC, Committer +hsaputra + + +https://avatars0.githubusercontent.com/u/8959638?s=50; /> +Matthias J. Sax +Committer +mjsax + + +https://avatars1.githubusercontent.com/u/409707?s=50; /> +Sebastian Schelter +PMC, Committer +ssc + + +https://avatars2.githubusercontent.com/u/1925554?s=50; /> +Kostas Tzoumas +PMC, Committer +ktzoumas + + + +Timo Walther +PMC, Committer +twalthr + + +https://avatars1.githubusercontent.com/u/1826769?s=50; /> +Daniel Warneke +PMC, Committer +warneke + + +https://avatars1.githubusercontent.com/u/4425616?s=50; /> +ChengXiang Li +Committer +chengxiang + + +https://avatars1.githubusercontent.com/u/569655?s=50; /> +Greg Hogan +Committer +greg + + +https://avatars1.githubusercontent.com/u/5284370?s=50; /> +Tzu-Li (Gordon) Tai +Committer +tzulitai + + + +You can reach committers directly at apache-id@apache.org. A list of all contributors can be found https://cwiki.apache.org/confluence/display/FLINK/List+of+contributors;>here. + +Former mentors + +The following people were very kind to mentor the project while in incubation. + + + + +Name +Role +Apache ID + + + +Ashutosh Chauhan +Former PPMC, Mentor +hashutosh + + + +Ted Dunning +Former PPMC, Mentor +tdunning + + + +Alan Gates +Former PPMC, Mentor +gates + + + + +Owen O'Malley +Former PPMC, Mentor +omalley + + + +Sean Owen +Former PPMC, Mentor +srowen + + + +Henry Saputra +Former PPMC, Mentor +hsaputra + + + +Slides + +Note: Keep in mind that code examples on slides have a chance of being incomplete or outdated. Always refer to the http://ci.apache.org/projects/flink/flink-docs-release-1.2;>latest documentation for an up to date reference. + +2016 + + + Stefan Richter: A look at Apache Flink 1.2 and beyond Apache Flink Meetup Berlin, November 2016: http://www.slideshare.net/StefanRichter10/a-look-at-flink-12;>SlideShare + Robert Metzger: Apache Flink Community Updates November 2016 Apache Flink Meetup Berlin, November 2016: http://www.slideshare.net/robertmetzger1/apache-flink-community-updates-november-2016-berlin-meetup;>SlideShare + Aljoscha Krettek: Apache Flink for IoT: How Event-Time Processing Enables Easy and Accurate Analytics Big Data Spain, Madrid November 2016: http://www.slideshare.net/dataArtisans/aljoscha-krettek-apache-flink-for-iot-how-eventtime-processing-enables-easy-and-accurate-analytics;>SlideShare + Stephan Ewen: Keynote -The maturing data streaming ecosystem and Apache Flinkâs accelerated growth Apache Big Data Europe 2016, Seville November 2016: http://www.slideshare.net/dataArtisans/keynote-stephan-ewen-stream-processing-as-a-foundational-paradigm-and-apache-flinks-approach-to-it;>SlideShare + Kostas Tzoumas: Stream Processing with Apache Flink® Apache Flink London Meetup, November 2016: http://www.slideshare.net/dataArtisans/kostas-tzoumas-stream-processing-with-apache-flink;>SlideShare + Kostas Tzoumas: Apache Flink®: State of the Union and Whatâs Next Strata + Hadoop World New York, September 2016: http://www.slideshare.net/dataArtisans/kostas-tzoumas-apache-flink-state-of-the-union-and-whats-next;>SlideShare + Kostas Tzoumas Stephan Ewen: Keynote -The maturing data streaming ecosystem and Apache Flinkâs accelerated growth Flink Forward, Berlin September 2016: http://www.slideshare.net/FlinkForward/kostas-tzoumasstpehan-ewen-keynote-the-maturing-data-streaming-ecosystem-and-apache-flinks-accelerated-growth;>SlideShare + Robert Metzger: Connecting Apache Flink to the World - Reviewing the streaming connectors F
[35/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/blog/release_1.1.0-changelog.html -- diff --git a/content/blog/release_1.1.0-changelog.html b/content/blog/release_1.1.0-changelog.html deleted file mode 100644 index 15d6ddc..000 --- a/content/blog/release_1.1.0-changelog.html +++ /dev/null @@ -1,689 +0,0 @@ - - - - - - - -Apache Flink: Release 1.1.0 â Changelog - - - - -https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> - - - - - - - - -https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -Download Flink - - -Home - - -Introduction to Flink - - -Flink Use Cases - - -Powered by Flink - - -Ecosystem - - -Community Project Info - - -How to Contribute - - -Flink Blog - - - - - - - - - Documentation - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) - http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) - - - - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart - - - - - https://github.com/apache/flink; target="_blank">Flink on GitHub - - - - - - - - - - - - - - - -Project FAQ - - -https://twitter.com/apacheflink; target="_blank">@ApacheFlink - - -Plan Visualizer - - - - - - - - - -Release 1.1.0 â Changelog - - - Changelog - Sub-task - Bug - Improvement - New Feature - Task - Test - Wish - - - - -Changelog - -The 1.1.0 release https://issues.apache.org/jira/issues/?jql=project+%3D+FLINK+AND+fixVersion+%3D+1.1.0;>resolved 457 JIRA issues in total. - -Sub-task - - - https://issues.apache.org/jira/browse/FLINK-1502;>FLINK-1502: Expose metrics to graphite, ganglia and JMX. - https://issues.apache.org/jira/browse/FLINK-1550;>FLINK-1550: Show JVM Metrics for JobManager - https://issues.apache.org/jira/browse/FLINK-3129;>FLINK-3129: Add tooling to ensure interface stability - https://issues.apache.org/jira/browse/FLINK-3141;>FLINK-3141: Design of NULL values handling in operation - https://issues.apache.org/jira/browse/FLINK-3223;>FLINK-3223: Translate Table API query into logical relational plans for Apache Calcite - https://issues.apache.org/jira/browse/FLINK-3225;>FLINK-3225: Optimize logical Table API plans in Calcite - https://issues.apache.org/jira/browse/FLINK-3226;>FLINK-3226: Translate optimized logical Table API plans into physical plans representing DataSet programs - https://issues.apache.org/jira/browse/FLINK-3229;>FLINK-3229: Kinesis streaming consumer with integration of Flink's checkpointing mechanics - https://issues.apache.org/jira/browse/FLINK-3230;>FLINK-3230: Kinesis streaming producer - https://issues.apache.org/jira/browse/FLINK-3231;>FLINK-3231: Handle Kinesis-side resharding in Kinesis streaming consumer - https://issues.apache.org/jira/browse/FLINK-3327;>FLINK-3327: Attach the ExecutionConfig to the JobGraph and make it accessible to the AbstractInvocable. - https://issues.apache.org/jira/browse/FLINK-3489;>FLINK-3489: Refactor Table API before merging into master - https://issues.apache.org/jira/browse/FLINK-3544;>FLINK-3544: ResourceManager runtime components - https://issues.apache.org/jira/browse/FLINK-3545;>FLINK-3545: ResourceManager: YARN integration - https://issues.apache.org/jira/browse/FLINK-3547;>FLINK-3547: Add support for streaming projection, selection, and union - https://issues.apache.org/jira/browse/FLINK-3550;>FLINK-3550: Rework stream join example - https://issues.apache.org/jira/browse/FLINK-3552;>FLINK-3552: Change socket WordCount to be properly windowed - https://issues.apache.org/jira/browse/FLINK-3573;>FLINK-3573: Implement more String functions for Table API - https://issues.apache.org/jira/browse/FLINK-3574;>FLINK-3574: Implement math functions for Table API -
[48/78] [abbrv] flink-web git commit: Rebuild site
http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/q/quickstart-SNAPSHOT.sh -- diff --git a/content/q/quickstart-SNAPSHOT.sh b/content/q/quickstart-SNAPSHOT.sh new file mode 100755 index 000..1b1f2d2 --- /dev/null +++ b/content/q/quickstart-SNAPSHOT.sh @@ -0,0 +1,47 @@ +#!/usr/bin/env bash + + +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + + + +PACKAGE=quickstart + +mvn archetype:generate \ + -DarchetypeGroupId=org.apache.flink \ + -DarchetypeArtifactId=flink-quickstart-java \ + -DarchetypeVersion=1.2-SNAPSHOT \ + -DgroupId=org.myorg.quickstart \ + -DartifactId=$PACKAGE \ + -Dversion=0.1 \ + -Dpackage=org.myorg.quickstart \ + -DinteractiveMode=false \ + -DarchetypeCatalog=https://repository.apache.org/content/repositories/snapshots/ + +# +# Give some guidance +# +echo -e "\\n\\n" +echo -e "\\tA sample quickstart Flink Job has been created." +echo -e "\\tSwitch into the directory using" +echo -e "\\t\\t cd $PACKAGE" +echo -e "\\tImport the project there using your favorite IDE (Import it as a maven project)" +echo -e "\\tBuild a jar inside the directory using" +echo -e "\\t\\t mvn clean package" +echo -e "\\tYou will find the runnable jar in $PACKAGE/target" +echo -e "\\tConsult our website if you have any troubles: http://flink.apache.org/community.html#mailing-lists; +echo -e "\\n\\n" http://git-wip-us.apache.org/repos/asf/flink-web/blob/9ec0a879/content/q/quickstart-scala-SNAPSHOT.sh -- diff --git a/content/q/quickstart-scala-SNAPSHOT.sh b/content/q/quickstart-scala-SNAPSHOT.sh new file mode 100755 index 000..9cfeb2e --- /dev/null +++ b/content/q/quickstart-scala-SNAPSHOT.sh @@ -0,0 +1,47 @@ +#!/usr/bin/env bash + + +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + + + +PACKAGE=quickstart + +mvn archetype:generate \ + -DarchetypeGroupId=org.apache.flink \ + -DarchetypeArtifactId=flink-quickstart-scala \ + -DarchetypeVersion=1.2-SNAPSHOT \ + -DgroupId=org.myorg.quickstart \ + -DartifactId=$PACKAGE \ + -Dversion=0.1 \ + -Dpackage=org.myorg.quickstart \ + -DinteractiveMode=false \ + -DarchetypeCatalog=https://repository.apache.org/content/repositories/snapshots/ + +# +# Give some guidance +# +echo -e "\\n\\n" +echo -e
[15/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/news/2015/08/24/introducing-flink-gelly.html -- diff --git a/content/news/2015/08/24/introducing-flink-gelly.html b/content/news/2015/08/24/introducing-flink-gelly.html deleted file mode 100644 index 714e75a..000 --- a/content/news/2015/08/24/introducing-flink-gelly.html +++ /dev/null @@ -1,649 +0,0 @@ - - - - - - - -Apache Flink: Introducing Gelly: Graph Processing with Apache Flink - - - - -https://maxcdn.bootstrapcdn.com/bootstrap/3.3.4/css/bootstrap.min.css;> - - - - - - - - -https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -Download Flink - - -Home - - -Introduction to Flink - - -Flink Use Cases - - -Powered by Flink - - -Ecosystem - - -Community Project Info - - -How to Contribute - - -Flink Blog - - - - - - - - - Documentation - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1; target="_blank">1.1 (Latest stable release) - http://ci.apache.org/projects/flink/flink-docs-release-1.2; target="_blank">1.2 (Snapshot) - - - - - - http://ci.apache.org/projects/flink/flink-docs-release-1.1/quickstart/setup_quickstart.html; target="_blank">Quickstart - - - - - https://github.com/apache/flink; target="_blank">Flink on GitHub - - - - - - - - - - - - - - - -Project FAQ - - -https://twitter.com/apacheflink; target="_blank">@ApacheFlink - - -Plan Visualizer - - - - - - - - - - - Introducing Gelly: Graph Processing with Apache Flink - - -24 Aug 2015 - -This blog post introduces Gelly, Apache Flinkâs graph-processing API and library. Flinkâs native support -for iterations makes it a suitable platform for large-scale graph analytics. -By leveraging delta iterations, Gelly is able to map various graph processing models such as -vertex-centric or gather-sum-apply to Flink dataflows. - -Gelly allows Flink users to perform end-to-end data analysis in a single system. -Gelly can be seamlessly used with Flinkâs DataSet API, -which means that pre-processing, graph creation, analysis, and post-processing can be done -in the same application. At the end of this post, we will go through a step-by-step example -in order to demonstrate that loading, transformation, filtering, graph creation, and analysis -can be performed in a single Flink program. - -Overview - - - What is Gelly? - Graph Representation and Creation - Transformations and Utilities - Iterative Graph Processing - Library of Graph Algorithms - Use-Case: Music Profiles - Ongoing and Future Work - - - - -What is Gelly? - -Gelly is a Graph API for Flink. It is currently supported in both Java and Scala. -The Scala methods are implemented as wrappers on top of the basic Java operations. -The API contains a set of utility functions for graph analysis, supports iterative graph -processing and introduces a library of graph algorithms. - - - - - -Back to top - -Graph Representation and Creation - -In Gelly, a graph is represented by a DataSet of vertices and a DataSet of edges. -A vertex is defined by its unique ID and a value, whereas an edge is defined by its source ID, -target ID, and value. A vertex or edge for which a value is not specified will simply have the -value type set to NullValue. - -A graph can be created from: - - - DataSet of edges and an optional DataSet of vertices using Graph.fromDataSet() - DataSet of Tuple3 and an optional DataSet of Tuple2 using Graph.fromTupleDataSet() - Collection of edges and an optional Collection of vertices using Graph.fromCollection() - - -In all three cases, if the vertices are not provided, -Gelly will automatically produce the vertex IDs from the edge source and target IDs. - -Back to top - -Transformations and Utilities - -These are methods of the Graph class and include common graph metrics, transformations -and mutations as well as neighborhood aggregations. - -Common Graph Metrics -These methods can be used
[02/78] [abbrv] flink-web git commit: Revert "Rebuild website"
http://git-wip-us.apache.org/repos/asf/flink-web/blob/16a92b0c/content/visualizer/js/jquery.tools.min.js -- diff --git a/content/visualizer/js/jquery.tools.min.js b/content/visualizer/js/jquery.tools.min.js deleted file mode 100755 index dc914d8..000 --- a/content/visualizer/js/jquery.tools.min.js +++ /dev/null @@ -1,20 +0,0 @@ -/*! - * jQuery Tools v1.2.7 - The missing UI library for the Web - * - * overlay/overlay.js - * scrollable/scrollable.js - * tabs/tabs.js - * tooltip/tooltip.js - * - * NO COPYRIGHTS OR LICENSES. DO WHAT YOU LIKE. - * - * http://flowplayer.org/tools/ - * - */ -/*! jQuery v1.7.2 jquery.com | jquery.org/license */ -(function(a,b){function cy(a){return f.isWindow(a)?a:a.nodeType===9?a.defaultView||a.parentWindow:!1}function cu(a){if(!cj[a]){var b=c.body,d=f("<"+a+">").appendTo(b),e=d.css("display");d.remove();if(e==="none"||e===""){ck||(ck=c.createElement("iframe"),ck.frameBorder=ck.width=ck.height=0),b.appendChild(ck);if(!cl||!ck.createElement)cl=(ck.contentWindow||ck.contentDocument).document,cl.write((f.support.boxModel?"":"")+""),cl.close();d=cl.createElement(a),cl.body.appendChild(d),e=f.css(d,"display"),b.removeChild(ck)}cj[a]=e}return cj[a]}function ct(a,b){var c={};f.each(cp.concat.apply([],cp.slice(0,b)),function(){c[this]=a});return c}function cs(){cq=b}function cr(){setTimeout(cs,0);return cq=f.now()}function ci(){try{return new a.ActiveXObject("Microsoft.XMLHTTP")}catch(b){}}function ch(){try{return new a.XMLHttpRequest}catch(b){}}function cb(a,c){a.dataFilter&&(c=a.dataFilter(c,a.dataType));var d=a.dataTypes,e={},g,h,i=d.length,j,k=d[0],l,m,n,o,p;for(g=1; g0){if(c!=="border")for(;e