Matt were you using 1.6?
On Thu, Apr 5, 2018 at 3:56 PM Mike Thomsen <[email protected]> wrote:

> 1.5
>
> Thanks,
>
> Mike
>
> On Thu, Apr 5, 2018 at 3:40 PM, Matt Burgess <[email protected]> wrote:
>
>> Mike,
>>
>> I can't reproduce this, I use the same DDL and Avro schema, with data
>> coming in from the SiteToSiteProvenanceReportingTask and going to
>> Postgres, and it works fine. What version of NiFi are you using?
>>
>> Regards,
>> Matt
>>
>>
>> On Thu, Apr 5, 2018 at 3:05 PM, Mike Thomsen <[email protected]>
>> wrote:
>> > Found these errors in the Docker logs:
>> >
>> > postgres_1       | 2018-04-05 18:33:22.183 UTC [51] ERROR:  column
>> > "timestamp_field" is of type timestamp without time zone but expression
>> is
>> > of type bigint at character 282
>> > postgres_1       | 2018-04-05 18:33:22.183 UTC [51] HINT:  You will
>> need to
>> > rewrite or cast the expression.
>> > postgres_1       | 2018-04-05 18:33:22.183 UTC [51] STATEMENT:  INSERT
>> INTO
>> > provenance (componentid, componentname, componenttype, details,
>> entityid,
>> > entitysize, entitytype, eventid, eventtype, processgroupid,
>> > processgroupname, record_count, schema_name, timestamp_field) VALUES
>> > ($1,$2,$3,$4,$5,$6,$7,$8,$9,$10,$11,$12,$13,$14)
>> > nifi2            | 2018-04-05 18:33:22,184 WARN [Timer-Driven Process
>> > Thread-2] o.a.n.p.standard.PutDatabaseRecord
>> > PutDatabaseRecord[id=dfdd3dd1-21ee-16e2-09cc-159b7c7f8f54] Failed to
>> process
>> >
>> StandardFlowFileRecord[uuid=8cf3b521-ac0b-4149-a60e-5fa7b2d2b3c5,claim=StandardContentClaim
>> > [resourceClaim=StandardResourceClaim[id=1522953007606-749,
>> > container=default, section=749], offset=434600,
>> > length=528],offset=0,name=4306158395541,size=528] due to
>> > java.sql.BatchUpdateException: Batch entry 0 INSERT INTO provenance
>> > (componentid, componentname, componenttype, details, entityid,
>> entitysize,
>> > entitytype, eventid, eventtype, processgroupid, processgroupname,
>> > record_count, schema_name, timestamp_field) VALUES
>> >
>> ('8d08a7d3-0162-1000-7216-7e0b426e774a','EvaluateJsonPath','EvaluateJsonPath',NULL,'26a1efa3-bcd3-4015-8a43-2a2c43c05714',66307,'org.apache.nifi.flowfile.FlowFile','e841aa4e-a7d3-4a6e-b78c-4e785f53b60e','ATTRIBUTES_MODIFIED','8c5c89d5-0162-1000-4395-a38d1c7b0b2f','Mongo
>> > ES Test',NULL,NULL,1522952968586) was aborted: ERROR: column
>> > "timestamp_field" is of type timestamp without time zone but expression
>> is
>> > of type bigint
>> >
>> > I have the following DDL and Avro schema:
>> >
>> > create table provenance (
>> > id serial,
>> > componentId varchar(128),
>> > componentName varchar(256),
>> > componentType varchar(128),
>> > details varchar(256),
>> > entityId varchar(256),
>> > entitySize int,
>> > entityType varchar(128),
>> > eventId varchar(128),
>> > eventType varchar(128),
>> > processGroupId varchar(128),
>> > processGroupName varchar(128), record_count int,
>> > schema_name varchar(64),
>> > timestamp_field timestamp
>> > )
>> >
>> > {
>> > "type": "record",
>> > "name": "ProvenanceEvent",
>> > "fields": [
>> > { "name": "componentId", "type": ["null", "string"] },
>> > { "name": "componentName", "type": ["null", "string"] },
>> > { "name": "componentType", "type": ["null", "string"] },
>> > { "name": "details", "type": ["null", "string"] },
>> > { "name": "entityId", "type": ["null", "string"] },
>> > { "name": "entitySize", "type": ["null", "int"] },
>> > { "name": "entityType", "type": ["null", "string"] },
>> > { "name": "eventId", "type": ["null", "string"] },
>> > { "name": "eventType", "type": ["null", "string"] },
>> > { "name": "processGroupId", "type": ["null", "string"] },
>> > { "name": "processGroupName", "type": ["null", "string"] },
>> > { "name": "record_count", "type": ["null", "int"] },
>> > { "name": "schema_name", "type": ["null", "string"] },
>> > { "name": "timestamp_field", "type": "long", "logicalType":
>> > "timestamp-millis" }
>> > ]
>> > }
>> >
>> > I set the JsonTreeReader to use this formatting option for reading in
>> the
>> > original ISO8601 string:
>> >
>> > yyyy-MM-dd'T'HH:mm:ssZ
>> >
>> > Everything looks fine until it gets to the processor. Any ideas?
>> >
>> > Thanks,
>> >
>> > Mike
>>
>
>

Reply via email to