Re: Using ConvertRecord on compressed input

2019-07-28 Thread Koji Kawamura
Hello, Thanks for your question. I've posted my comment to the StackOverflow question. I'd avoid adding it to the core package as some of Record formats handles compressed inputs by themselves, like Avro. http://apache-avro.679487.n3.nabble.com/read-a-compressed-avro-file-td3872899.html Adding

Re: Running NiFi on Google Cloud

2019-07-28 Thread Márcio Sugar
Hi Dano, Thanks for your recommendation. I'll surely keep that in mind. >From your answer, I infer at least some of your data processing uses NiFi as >the choreographer. In my case, we use NiFi just to move data around, so it >performs a more limited role. To give you some context: My goal is

Re: Running NiFi on Google Cloud

2019-07-28 Thread dan young
Hello Márcio, We've been running NiFi clusters for almost 3 years now at Looker on AWS. We will be moving these over to GCP in the future. My main recommendation is to ensure that you're using something like Ansible to help with the deployment and configuration of the cluster. We use a lot of

Re: Re: CaptureChangeMySQL throws ServerException when configured with all nodes execution

2019-07-28 Thread wangl...@geekplus.com.cn
So the CaptureChangeMySQL processor is designed to be configured as run on Primary node? even my nifi is cluster mode. Now I have many CaptureChangeMySQL processors, connecting to different Mysql Server. If all run on the primary node, the primary node is too high loaded. Is there

Using ConvertRecord on compressed input

2019-07-28 Thread Malthe
CSV, JSON, and XML files all compress well and thus you often see them stored in a compressed format. Actually, it's not feasible or simply practical to first decompress the files and then process them into a binary format. How would I go about extending or modifying the `AbstractRecordProcessor`

Re: CaptureChangeMySQL throws ServerException when configured with all nodes execution

2019-07-28 Thread Purushotham Pushpavanthar
Hi, If CaptureChangeMySQL is configured to run on all nodes, there will duplicate events for sure. However, in your case the exception is thrown for having same "Sever ID" for all instances of the processor connecting to MySQL server. I don't see any data loss due currecurrent execution.