Greetings, My team and I have been using the SplitRecord processor to to convert CSV to JSON and to later ingest it for about a year. I’m noticing some things that I feel worked a lot better before. I just don’t know exactly when “before” was, but let’s say at initial testing and development time, my CSV ingesting to JSON used to work a lot better. I have an AVRO schema setup that gives default values to facilitate error handling. I am observing some troubling issues. We recently upgraded to nifi 1.6.0 last month, so I’m naturally suspicious of the upgrade, but do not see known issues that relate to this. Has anyone noticed these issues or are some of these items known issues? I’m dealing w/ probably over 40 CSV columns that I generate for testing and then ingest. Observations:
1. I have decimals that I’m trying to take in as Strings. Instead, we are seeing the empty string being retrieved on those fields from the CSV to JSON line-by-line conversion phase. The empty string is the default if something is not found. 2. I have integers w/ values being taken in as -1, the default, so the conversion phase is not working here properly. 3. I have strings that exist that get dropped to empty string. 4. And, I have a few integer fields that get correctly taken in as strings and match the right values in the CSV. Regards, -- Jeremy H. Taylor Software Developer ACES, Incorporated http://acesinc.net<http://acesinc.net/>
