GetSQS causes high CPU usage

2015-10-19 Thread Adam Lamar
Hi everybody! I've been testing NiFi 0.3.0 with the GetSQS processor to fetch objects from an AWS bucket as they're created. My flow looks like this: GetSQS SplitJson ExtractText FetchS3Object PutFile I noticed that GetSQS causes a high amount of CPU usage - about 90% of one core. If I turn

Re: output port

2015-10-19 Thread Bryan Bende
Hello, Just to clarify, so you are seeing the messages reach the output port and then get removed from the queue? And on the spark side the NiFi Spark receiver never receives anything? Or it receives message, but they have no content? -Bryan On Monday, October 19, 2015, Rama Krishna Manne wrote

Re: Best practice approach to set HDFS filename based on attibutes

2015-10-19 Thread Bryan Bende
gt; pretty much like: > > 1. ListenHTTP => captures attribute LogSrc from HTTP request header LogSrc > > 2. MergeContent => where Correlation Attribute Name = LogSrc / > Attribute Strategy = Keep Only Common Attributes > > 3. UpdateAttribute => Updates $filename so tha

output port

2015-10-19 Thread Rama Krishna Manne
I have an flow in which messages are emitted to an output and apache-spark will pull the messages from the port , I see the messages are pulled by spark but cannot see the data pulled(cannot do any computations) I tried a different way , the messages are pushed apache kafka and spark pulls message

Best practice approach to set HDFS filename based on attibutes

2015-10-19 Thread Andre
on Attributes 3. UpdateAttribute => Updates $filename so that it is now data-${now():format('yyyyMMdd')}.log (e.g. data-20151019.log ) 4. PutHDFS => Directory = /${LogSrc}/${now():format('/MM/dd')} (e.g. /mydevice/2015/10/19/) This, I believe, would result on a file named /mydev

Best practice approach to set HDFS filename based on attibutes

2015-10-19 Thread Andre
on Attributes 3. UpdateAttribute => Updates $filename so that it is now data-${now():format('yyyyMMdd')}.log (e.g. data-20151019.log ) 4. PutHDFS => Directory = /${LogSrc}/${now():format('/MM/dd')} (e.g. /mydevice/2015/10/19/) This, I believe, would result on a file named /mydev

how to putkafka to multiple partions

2015-10-19 Thread 彭光裕
hi, I have a test topic in kafka with 6 partitions. When I use default PutKafka of nifi to put messages(let’s say 100/batch), I found it usually only put to one partition. I have tried the kafka key property to be ‘${nextInt():mod(6)}’ or ‘${uuid()}’, but it doesn’t seems work. How do I