Hello,
thanks for the answer but windows produce periodical results. I used your
example but the data source is changed to TCP stream:
DataStream text = env.socketTextStream("localhost", 2015,
'\n');
DataStream> wordCounts =
text
Hi Brian,
We are currently in the process of releasing 0.10.0. Thus, the master
version has already been updated to 1.0 which is the next scheduled
release.
If you want to use the latest SNAPSHOT version, you may build it from
source or use the SNAPSHOT Maven artifacts. For more information,
The distributed "start-cluster.sh" script works only, if the code is
accessible under the same path on all machines, which must be the same path
as on the machine where you invoke the script.
Otherwise the paths for remote shell commands will be wrong, and the
classpaths will be wrong as a result.
Super nice to hear :-)
On Mon, Nov 9, 2015 at 4:48 PM, Niels Basjes wrote:
> Apparently I just had to wait a bit longer for the first run.
> Now I'm able to package the project in about 7 minutes.
>
> Current status: I am now able to access HBase from within Flink on a
>
Apparently I just had to wait a bit longer for the first run.
Now I'm able to package the project in about 7 minutes.
Current status: I am now able to access HBase from within Flink on a
Kerberos secured cluster.
Cleaning up the patch so I can submit it in a few days.
On Sat, Nov 7, 2015 at
Hi everyone,
I am considering using Flink in a project. The setting would be a YARN cluster
where data is first read in from HDFS, then processed and finally written into
an Oracle database using an upsert command. If I understand the documentation
correctly, the DataSet API would be the
Hi!
Rather than taking an 0.10-SNAPSHOT, you could also take a 0.10 release
candidate.
The latest is for example in
https://repository.apache.org/content/repositories/orgapacheflink-1053/
Greetings,
Stephan
On Mon, Nov 9, 2015 at 5:45 PM, Maximilian Michels wrote:
> Hi
Hi All,
I read a couple of article about Kappa and Lambda Architecture.
http://www.confluent.io/blog/real-time-stream-processing-the-next-step-for-apache-flink/
I'm convince that Flink will simplify this one with streaming.
However i also stumble upon this blog post that has valid argument to
Hi Thomas,
It appears Flink couldn't pick up the Hadoop configuration. Did you
set the environment variables HADOOP_CONF_DIR or HADOOP_HOME?
Best,
Max
On Sun, Nov 8, 2015 at 7:52 PM, Thomas Götzinger wrote:
> Sorry for Confusing,
>
> the flink cluster throws following
Hi,
thanks for having a look at this, Aljoscha.
Not being able to read a DataSet[Row] from csv is definitively the most
major issue for me right now.
Everything else I could work around with Scala magic. I can create an issue
for this if you'd like.
Regarding the other points:
1. Oh absolutely,
Hi!
If you want to work on subsets of streams, the answer is usually to use
windows, "stream.keyBy(...).timeWindow(Time.of(1, MINUTE))".
The transformations that you want to make, do they fit into a window
function?
There are thoughts to introduce something like global time windows across
the
Hi flinkers,
I have a simple question for you that I wanted to ask since the beginning
of Flink: is there really the need to separate Rich and normal operators?
Why not keeping just the rich version?
Best,
Flavio
These reason is that the non-rich function interfaces are SAM (single
abstract method) interfaces.
In Java 8, SAM interfaces can be specified as concise lambda functions.
Cheers, Fabian
2015-11-09 10:45 GMT+01:00 Flavio Pompermaier :
> Hi flinkers,
> I have a simple
13 matches
Mail list logo