[ 
https://issues.apache.org/jira/browse/NIFI-1868?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15404766#comment-15404766
 ] 

ASF GitHub Bot commented on NIFI-1868:
--------------------------------------

Github user bbende commented on the issue:

    https://github.com/apache/nifi/pull/706
  
    Error handling on the relationships is looking good, thanks for making 
those updates.
    
    One more thing I ran into, when I set concurrent tasks to 2 on the 
PutHiveStreamingProcessor and I was generating a flow file every 500 ms, I 
would occasionally get the following error:
    
    ```
    2016-08-02 20:50:17,860 ERROR [Timer-Driven Process Thread-1] 
o.a.n.processors.hive.PutHiveStreaming
    java.lang.NullPointerException: null
        at 
org.apache.hive.hcatalog.streaming.StrictJsonWriter.write(StrictJsonWriter.java:79)
 ~[na:na]
        at 
org.apache.hive.hcatalog.streaming.HiveEndPoint$TransactionBatchImpl.write(HiveEndPoint.java:632)
 ~[na:na]
        at org.apache.nifi.util.hive.HiveWriter$1.call(HiveWriter.java:113) 
~[na:na]
        at org.apache.nifi.util.hive.HiveWriter$1.call(HiveWriter.java:110) 
~[na:na]
        at java.util.concurrent.FutureTask.run(FutureTask.java:266) 
~[na:1.8.0_102]
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) 
[na:1.8.0_102]
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) 
[na:1.8.0_102]
        at java.lang.Thread.run(Thread.java:745) [na:1.8.0_102] 
    ```


> Add support for Hive Streaming
> ------------------------------
>
>                 Key: NIFI-1868
>                 URL: https://issues.apache.org/jira/browse/NIFI-1868
>             Project: Apache NiFi
>          Issue Type: New Feature
>            Reporter: Matt Burgess
>            Assignee: Matt Burgess
>             Fix For: 1.0.0
>
>
> Traditionally adding new data into Hive requires gathering a large amount of 
> data onto HDFS and then periodically adding a new partition. This is 
> essentially a “batch insertion”. Insertion of new data into an existing 
> partition is not permitted. Hive Streaming API allows data to be pumped 
> continuously into Hive. The incoming data can be continuously committed in 
> small batches of records into an existing Hive partition or table. Once data 
> is committed it becomes immediately visible to all Hive queries initiated 
> subsequently.
> This case is to add a PutHiveStreaming processor to NiFi, to leverage the 
> Hive Streaming API to allow continuous streaming of data into a Hive 
> partition/table.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to