Re: Flink on K8s job submission best practices

2017-12-22 Thread Martin Eden
The above applies to Mesos/DCOS as well. So if someone would also share insights into automatic job deployment in that setup would very useful. Thanks. M On Fri, Dec 22, 2017 at 6:56 PM, Maximilian Bode < maximilian.b...@tngtech.com> wrote: > Hi everyone, > > We are beginning to run Flink on K8s

Re: flink eventTime, lateness, maxoutoforderness

2017-12-22 Thread chen
Thanks Gordon, Please see the rely. I use code, but the result it doesn`t like what the doc explain. -- Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/

Re: flink eventTime, lateness, maxoutoforderness

2017-12-22 Thread chen
CODE with maxOutOfOrdernesstime effect: dataStream.keyBy(row -> (String)row.getField(0)) .window(TumblingEventTimeWindows.of(Time.seconds(5))) .fold(initRow(), new FoldFunction() { @Override public Row

Re: flink eventTime, lateness, maxoutoforderness

2017-12-22 Thread chen
Hi Eron, Thanks for your help. Actually I know maxoutoforder, lateness is based on Event Time. But in my test it is not. Following is my code and test data. "key1|148325064|", "key1|1483250636000|", "key1|1483250649000|", "key1|1483250642000|",

Re: CSV writer/parser inconsistency when using the Table API?

2017-12-22 Thread Fabian Hueske
Hi Cliff, you are right. The CsvTableSink and the CsvInputFormat are not in sync. It would be great if you could open a JIRA ticket for this issue. As a workaround, you could implement your own CsvTableSink to add a delimiter after the last field. The code is straightforward, less than 150 lines

CSV writer/parser inconsistency when using the Table API?

2017-12-22 Thread Cliff Resnick
I've been trying out the Table API for some ETL using a two-stage job of CsvTableSink (DataSet) -> CsvInputFormat (Stream). I ran into an issue where the first stage produces output with trailing null values (valid), which causes a parse error in the second stage. Looking at

Flink network access control documentation

2017-12-22 Thread Elias Levy
There is a need for better documentation on what connects to what over which ports in a Flink cluster to allow users to configure network access control rules. I was under the impression that in a ZK HA configuration the Job Managers were essentially independent and only coordinated via ZK. But

Flink on K8s job submission best practices

2017-12-22 Thread Maximilian Bode
Hi everyone, We are beginning to run Flink on K8s and found the basic templates [1] as well as the example Helm chart [2] very helpful. Also the discussion about JobManager HA [3] and Patrick's talk [4] was very interesting. All in all it is delightful how easy everything can be set up and works

Re: periodic trigger

2017-12-22 Thread Plamen Paskov
I think it will not solve the problem as if i set ContinuousEventTimeTrigger to 10 seconds and allowedLateness(Time.seconds(60)) as i don't want to discard events from different users received later then i might receive more than one row for a single user based on the number of windows created

Re: entrypoint for executing job in task manager

2017-12-22 Thread Piotr Nowojski
I don’t think there is such hook in the Flink code now. You will have to walk around this issue somehow in user space. Maybe you could make a contract that every operator before touching Guice, should call static synchronized method `initializeGuiceContext`. This method could search the

Re: periodic trigger

2017-12-22 Thread Piotr Nowojski
Ok, I think now I understand your problem. Wouldn’t it be enough, if you change last global window to something like this: lastUserSession .timeWindowAll(Time.seconds(10)) .aggregate(new AverageSessionLengthAcrossAllUsers()) .print(); (As a side note, maybe you should