Re: Visualize result of Flink job

2016-05-29 Thread Palle
I know exactly what to visualize. As I wrote, it is the latest result of the Flink job I would like to visualize. There is no need to use elastic to find it first. The data I have is of such a nature that they every 10 seconds could be written into a file, meaning that the file at all times w

Write matrix/vector

2016-05-29 Thread Lydia Ickler
Hi, I would like to know how to write a Matrix or Vector (Dense/Sparse) to file? Thanks in advance! Best regards, Lydia

sparse matrix

2016-05-29 Thread Lydia Ickler
Hi all, I have two questions regarding sparse matrices: 1. I have a sparse Matrix: val sparseMatrix = SparseMatrix.fromCOO(row, col, csvInput.collect()) and now I would like to extract all values that are in a specific row X. How would I tackle that? flatMap() and filter() do not seem to be sup

Extracting Timestamp in MapFunction

2016-05-29 Thread Biplob Biswas
Hi, I have a source where I am using the collectWithTimestamp method and streaming the timestamp along with the actual data. Now I want to get the values of the timestamp in the map function, I tried looking for that in the documentation in the following link and the ExtractTimestamp method inter

Re: Visualize result of Flink job

2016-05-29 Thread Kanstantsin Kamkou
> I am thinking it may not be the best fit, because Elastic is by nature a > search engine that is good for trending and stuff like that - not entire > replacement of the current view. Why u think that the elasticsearch is not the right tool? To visualise something u have to find what to visuali

Re: Checkpoint Error in flink with Rockdb state backend

2016-05-29 Thread Aljoscha Krettek
Ah yes, if you used a local filesystem for backups this certainly was the source of the problem. On Sun, 29 May 2016 at 17:57 arpit srivastava wrote: > I think the problem was that i was using local filesystem in a cluster. > Now I have switched to hdfs. > > Thanks, > Arpit > > On Sun, May 29, 2

Visualize result of Flink job

2016-05-29 Thread Palle
Hi there I am using Flink to analyse a lot of incoming data. Every 10 seconds it makes sense to present the analysis so far as some form of visualization. Every 10 seconds I therefore will replace the current contents of the visualization/presentation with the analysis result of the most recent

Re: Checkpoint Error in flink with Rockdb state backend

2016-05-29 Thread arpit srivastava
I think the problem was that i was using local filesystem in a cluster. Now I have switched to hdfs. Thanks, Arpit On Sun, May 29, 2016 at 12:57 PM, Aljoscha Krettek wrote: > Hi, > could you please provide the code of your user function that has the > Checkpointed interface and is keeping state

Re: Result comparison from 2 DataStream Sources

2016-05-29 Thread Konstantin Knauf
Hi again, from your diagram I have put together a gist, which I think does the job. I haven't had the time to test it though :( https://gist.github.com/knaufk/d1312503b99ee51554a70c9a22abe7e5 If you have any questions, let me know. It sometimes just takes a while until I answer ;) Cheers, Kons

Context-specific step function in Iteration

2016-05-29 Thread Martin Junghanns
Hi everyone, In a step-function (bulk) I'd like to join the working set W with another data set T. The join field of T depends on the current super step. Unfortunately, W has no access to the iteration runtime context. I tried to extract the current superstep at the beginning of the step functio

Re: Reading Parameter values sent to partition

2016-05-29 Thread Biplob Biswas
Aah, thanks a lot for that insight. Pretty new to the Flink systems and learning on my own so prone to making mistakes. Thanks a lot for helping. -- View this message in context: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Reading-Parameter-values-sent-to-partition-tp7

Re: Checkpoint Error in flink with Rockdb state backend

2016-05-29 Thread Aljoscha Krettek
Hi, could you please provide the code of your user function that has the Checkpointed interface and is keeping state? This might give people a chance of understanding what is going on. Cheers, Aljoscha On Sat, 28 May 2016 at 20:55 arpit srivastava wrote: > Hi, > > I am using Flink on yarn clust

Re: Writing test for Flink streaming jobs

2016-05-29 Thread Aljoscha Krettek
Hi, using DataStreamUtils.collect() in a test is difficult due to synchronization problems, as you discovered yourself. What I propose is to write a custom Sink that collects data and verifies the results. Verification should both happen in the invoke() method and in close(). For the sink, you sho