Re: finite subset of an infinite data stream

2015-11-09 Thread rss rss
Hello, thanks for the answer but windows produce periodical results. I used your example but the data source is changed to TCP stream: DataStream text = env.socketTextStream("localhost", 2015, '\n'); DataStream> wordCounts = text

Re: Missing 0.10 SNAPSHOT Download

2015-11-09 Thread Maximilian Michels
Hi Brian, We are currently in the process of releasing 0.10.0. Thus, the master version has already been updated to 1.0 which is the next scheduled release. If you want to use the latest SNAPSHOT version, you may build it from source or use the SNAPSHOT Maven artifacts. For more information,

Re: Cluster installation gives java.lang.NoClassDefFoundError for everything

2015-11-09 Thread Stephan Ewen
The distributed "start-cluster.sh" script works only, if the code is accessible under the same path on all machines, which must be the same path as on the machine where you invoke the script. Otherwise the paths for remote shell commands will be wrong, and the classpaths will be wrong as a result.

Re: Running continuously on yarn with kerberos

2015-11-09 Thread Stephan Ewen
Super nice to hear :-) On Mon, Nov 9, 2015 at 4:48 PM, Niels Basjes wrote: > Apparently I just had to wait a bit longer for the first run. > Now I'm able to package the project in about 7 minutes. > > Current status: I am now able to access HBase from within Flink on a >

Re: Running continuously on yarn with kerberos

2015-11-09 Thread Niels Basjes
Apparently I just had to wait a bit longer for the first run. Now I'm able to package the project in about 7 minutes. Current status: I am now able to access HBase from within Flink on a Kerberos secured cluster. Cleaning up the patch so I can submit it in a few days. On Sat, Nov 7, 2015 at

Checkpoints in batch processing & JDBC Output Format

2015-11-09 Thread Maximilian Bode
Hi everyone, I am considering using Flink in a project. The setting would be a YARN cluster where data is first read in from HDFS, then processed and finally written into an Oracle database using an upsert command. If I understand the documentation correctly, the DataSet API would be the

Re: Missing 0.10 SNAPSHOT Download

2015-11-09 Thread Stephan Ewen
Hi! Rather than taking an 0.10-SNAPSHOT, you could also take a 0.10 release candidate. The latest is for example in https://repository.apache.org/content/repositories/orgapacheflink-1053/ Greetings, Stephan On Mon, Nov 9, 2015 at 5:45 PM, Maximilian Michels wrote: > Hi

Flink, Kappa and Lambda

2015-11-09 Thread Welly Tambunan
Hi All, I read a couple of article about Kappa and Lambda Architecture. http://www.confluent.io/blog/real-time-stream-processing-the-next-step-for-apache-flink/ I'm convince that Flink will simplify this one with streaming. However i also stumble upon this blog post that has valid argument to

Re: Flink on EC"

2015-11-09 Thread Maximilian Michels
Hi Thomas, It appears Flink couldn't pick up the Hadoop configuration. Did you set the environment variables HADOOP_CONF_DIR or HADOOP_HOME? Best, Max On Sun, Nov 8, 2015 at 7:52 PM, Thomas Götzinger wrote: > Sorry for Confusing, > > the flink cluster throws following

Re: How best to deal with wide, structured tuples?

2015-11-09 Thread Johann Kovacs
Hi, thanks for having a look at this, Aljoscha. Not being able to read a DataSet[Row] from csv is definitively the most major issue for me right now. Everything else I could work around with Scala magic. I can create an issue for this if you'd like. Regarding the other points: 1. Oh absolutely,

Re: finite subset of an infinite data stream

2015-11-09 Thread Stephan Ewen
Hi! If you want to work on subsets of streams, the answer is usually to use windows, "stream.keyBy(...).timeWindow(Time.of(1, MINUTE))". The transformations that you want to make, do they fit into a window function? There are thoughts to introduce something like global time windows across the

Rich vs normal functions

2015-11-09 Thread Flavio Pompermaier
Hi flinkers, I have a simple question for you that I wanted to ask since the beginning of Flink: is there really the need to separate Rich and normal operators? Why not keeping just the rich version? Best, Flavio

Re: Rich vs normal functions

2015-11-09 Thread Fabian Hueske
These reason is that the non-rich function interfaces are SAM (single abstract method) interfaces. In Java 8, SAM interfaces can be specified as concise lambda functions. Cheers, Fabian 2015-11-09 10:45 GMT+01:00 Flavio Pompermaier : > Hi flinkers, > I have a simple