Re: Combining streams with static data and using REST API as a sink

2017-06-21 Thread Nancy Estrada
Hi Josh,

I have a use-case similar to yours. I need to join a stream with data from a
database to which I have access via a REST API. Since the Side inputs API
continues begin and ongoing work. I am wondering how did you approached it,
Did you use the rich function updating it periodically?

Thank you in advance!
Nancy





--
View this message in context: 
http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Combining-streams-with-static-data-and-using-REST-API-as-a-sink-tp7083p13902.html
Sent from the Apache Flink User Mailing List archive. mailing list archive at 
Nabble.com.


Flink Docker Kubernetes Gitlab CI CDeployment

2017-05-29 Thread Nancy Estrada
Hi all,

Has someone successfully run Flink jobs with this type of setup (Gitlab CI
CD and Kubernetes)?

Since Flink jobs can’t be dockerized and deployed in a natural way as part
of the container (according to Flip-6) I am not very sure of how is the best
way of doing this.

We are thinking of having 3 Docker files:  One for the Job manager, one for
the Task Manager, and one for a container that will submit the jars and
properties File to the Job Manager. With this setup we can still use the
official Flink Docker images, and we can automatically  generate the third
Docker file with the project specific tasks. This setup allows to have
Continuous deployment. However, it implies that we will have a new Flink
cluster every time there is a new code version.

What do you think of this set up? Has someone implemented something similar?
Is there any documentation about Flink-Gitlab-CI-CD-Kubernetes?

Thank you very much,
Nancy




--
View this message in context: 
http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Flink-Docker-Kubernetes-Gitlab-CI-CDeployment-tp13371.html
Sent from the Apache Flink User Mailing List archive. mailing list archive at 
Nabble.com.


Re: flink-1.2 and unit testing / flinkspector

2017-03-23 Thread Nancy Estrada
Hi Tarandeep and Ted,

I am in this route now. I am trying to use Flinkspector with Flink 1.2 using
your instructions but failing miserably. After applying the changes, when I
try to run "mvn clean install", some Tests fail and therefore I am not able
to build successfully. 

I am wondering if there is a pull request, where I can access to a version
of Flinkspector working with Flink 1.2.

Thank you,
Nancy



--
View this message in context: 
http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/flink-1-2-and-unit-testing-flinkspector-tp12281p12377.html
Sent from the Apache Flink User Mailing List archive. mailing list archive at 
Nabble.com.


Re: Cassandra Sink version

2017-03-23 Thread Nancy Estrada
The documentation you mentioned says: "The Java client driver 3.0.7 (branch
3.0.x) is compatible with Apache Cassandra 1.2, 2.0, 2.1, 2.2 and 3.0".

Thank you Kostas!



--
View this message in context: 
http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Cassandra-Sink-version-tp12329p12340.html
Sent from the Apache Flink User Mailing List archive. mailing list archive at 
Nabble.com.


Cassandra Sink version

2017-03-22 Thread Nancy Estrada
Hi all,

Which Cassandra version is currently supported by the Flink 1.2 connector?
and I am wondering which version Flink 1.3 will be supporting?

Thank you!



--
View this message in context: 
http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Cassandra-Sink-version-tp12329.html
Sent from the Apache Flink User Mailing List archive. mailing list archive at 
Nabble.com.


readFile - Continuous file processing

2017-01-31 Thread Nancy Estrada
Hi guys,

I have the following use case. Every day a new file is created and
periodically some log records are appended to it. I am reading the file in
the following way:

executionEnvironment.readFile(format, directoryPath, PROCESS_CONTINUOUSLY,
period.toMilliseconds(),filePathFilter);

However, Flink takes modified files as new files and consequently all the
content of the modified file gets processed again. I know that a solution is
to process the file until it contains all the records of the day but I will
like to process the file continuously. Therefore, I am wondering if there is
a way of processing just the new records in a file?

Thank you in advance! :)
Nancy 







--
View this message in context: 
http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/readFile-Continuous-file-processing-tp11384.html
Sent from the Apache Flink User Mailing List archive. mailing list archive at 
Nabble.com.


Flink configuration

2017-01-23 Thread Nancy Estrada
Hi all,

I have been reading about how to configure Flink when we have a set up
consisting on a couple of VMs with more than 1 vCore.  I am a bit confused
about how to set the degree of parallelism in the
taskmanager.numberOfTaskSlots parameter:

* According to the Flink documentation[1], this value is typically
proportional to the number of/ physical CPU cores/ that the TaskManager’s
machine has.

* However the YARN documentation[2], makes reference to the number of
/Virtual CPU cores/ per TaskManager.

My question is, If my Flink Jobs will be running on VMs (without using
YARN),  the "taskmanager.numberOfTaskSlots" will depend on the number of
vCPU that mi VM has? or must be related to the physical cores?  

Thanks in advance for your help! 
Nancy

[1]https://ci.apache.org/projects/flink/flink-docs-release-0.8/config.html
[2]https://ci.apache.org/projects/flink/flink-docs-release-0.8/yarn_setup.html



--
View this message in context: 
http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Flink-configuration-tp11210.html
Sent from the Apache Flink User Mailing List archive. mailing list archive at 
Nabble.com.


Consistency guarantees on multiple sinks

2017-01-05 Thread Nancy Estrada
Hi,
 
If in a Job there is more than one sink declared, what happens when a
failure occurs? all the sink operations get aborted? (atomically as in a
transactional environment), or the exactly-once-processing consistency
guarantees are provided just when one sink is declared per job? Is it
recommended to have more than one sink per job?

Thank you!
Nancy Estrada



--
View this message in context: 
http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Consistency-guarantees-on-multiple-sinks-tp10877.html
Sent from the Apache Flink User Mailing List archive. mailing list archive at 
Nabble.com.