NiFi-Neo4j Issues

2017-05-03 Thread dale.chang13
Hi All, At the bottom you can find my question. Note, I am positive this is more a network issue, but I cannot seem to figure out the solution. I tried posting this to the Neo4j Google+ board , and as a relevant NiFi-Neo4j Github issue

ExecuteProcess Question

2016-09-27 Thread dale.chang13
So I have a bash script that I am able to run from the command line, and I want to be able to let NiFi call it using the ExecuteProcess processor. The script itself runs fine from the command line, and it looks like the ExecuteProcess is executing the script as well (I have a LogAttribute

ListenHTTP Questions

2016-06-14 Thread dale.chang13
So I am having some woes with sending HTTP Post requests to my NiFi server by using curl to send a JSON object. I have a ListenHTTP Processor running that is configured as so: - BasePath = contentListener - ListeningPort = 8011 - Authorized DN Pattern = .* - Max Unconfirmed FlowFile Time = 60 secs

NiFi Version Bundled with HDF 1.1.1.0-12

2016-06-01 Thread dale.chang13
I cannot find what version of NiFi is bundled with Hortonworks Dataflow 1.1.1.0-12. Is there a way I can find out? -- View this message in context: http://apache-nifi-developer-list.39713.n7.nabble.com/NiFi-Version-Bundled-with-HDF-1-1-1-0-12-tp10857.html Sent from the Apache NiFi Developer

AttributesToJSON Multi-Valued Fields

2016-05-17 Thread dale.chang13
I don't know if this use-case is too specific to be a feature for a future release, but I would like to see the AttributesToJSON processor support multi-valued fields. In my use-case, I am storing JSON documents into Solr, and there are two ways to store multi-valued fields: using an array and

Distribute FlowFiles among Nodes in a Cluster

2016-05-16 Thread dale.chang13
I was wondering if there was a way for the NCM to distribute FlowFiles to different nodes. Currently I see that all of the nodes in my cluster run the same dataflow. I know I can restrict certain processors to the primary node, but it seems like the NCM does not distribute the FlowFiles to

Re: Purpose of Disallowing Attribute Expression

2016-05-13 Thread dale.chang13
Michael Moser wrote > NIFI-1077 [1] has discussed this a bit in the past, when > ConvertCharacterSet was improved to support expression language. A JIRA > ticket is needed to spur action on these requests. > > An interesting case to help this would be to improve the IdentifyMimeType > processor

Re: Purpose of Disallowing Attribute Expression

2016-05-12 Thread dale.chang13
Joe Witt wrote > It is generally quite easy to enable for Property Descriptors which > accept user supplied strings. And this is one that does seem like a > candidate. Were you wanting it to look at a flowfile attribute to be > the way of indicating the character set? > > Thinking through this

Purpose of Disallowing Attribute Expression

2016-05-12 Thread dale.chang13
What is the purpose of not allowing a Processor property to support expression language? Not allowing a property such as "Character set" in the ExtractText Processor is proving to be a hindrance. Would it affect NiFi under the hood if it were otherwise? -- View this message in context:

Re: Exception while restarting the Nifi Cluster

2016-05-11 Thread dale.chang13
Rahul Dahiya wrote > Hi Team, > > > I am getting below exception while trying to restart the NiFi nodes : > > > java.lang.Exception: Unable to load flow due to: java.io.IOException: > org.apache.nifi.cluster.ConnectionException: Failed to connect node to > cluster because local flow controller

Deallocation of FlowFiles

2016-05-05 Thread dale.chang13
Is there a way to free up resources (memory and disk space in repos like content_repo, flowfile_repo) at the conclusion of a NiFi flow? I would like to reclaim those resources quickly so I can reuse them for newer FlowFiles. -- View this message in context:

Re: FetchFile Cannot Allocate Enough Memory

2016-05-04 Thread dale.chang13
Mark Payne wrote > Dale, > > I think an image of the flow would be useful. Or better yet, if you can, a > template of the flow, so > that we can see all of the configuration being used. > > When you said you "get stuck at around 20 MB and then NiFi moves to a > crawl" I'm not clear on > what you

Re: FetchFile Cannot Allocate Enough Memory

2016-05-04 Thread dale.chang13
Joe Witt wrote > On May 4, 2016, at 8:56 AM, Joe Witt > joe.witt@ > wrote: > > Dale, > > Where there is a fetch file there is usually a list file. And while > the symptom of memory issues is showing up in fetch file i am curious > if the issue might actually be caused in ListFile. How many

Re: FetchFile Cannot Allocate Enough Memory

2016-04-29 Thread dale.chang13
Mark Payne wrote > Dale, > > I haven't seen this issue personally. I don't believe it has to do with > content/flowfile > repo space. Can you check the logs/nifi-app.log file and give us the exact > error message > from the logs, with the stack trace if it is provided? > > Thanks > -Mark Sure,

FetchFile Cannot Allocate Enough Memory

2016-04-29 Thread dale.chang13
I have been trying to run my data flow and I have been running into a problem with being unable to read FetchFiles. I will detail my process below and I would like some confirmation of my suspicions. First I am ingesting an initial file that is fairly large, which contains the path/filename of a

Re: NiFi Rest API Start

2016-04-25 Thread dale.chang13
Matt Burgess wrote > It's part of the "Update Processor" API, check out this thread from > the NiFi dev list archive: > > http://apache-nifi-developer-list.39713.n7.nabble.com/Reg-starting-and-Stopping-processor-tp7930p7949.html > > Regards, > Matt Nice, so I was able to change the status and

NiFi Rest API Start

2016-04-25 Thread dale.chang13
Looking through the NiFi API, I am not finding a way to start a Processor, Process Group, or an instance of NiFi. Is there a Rest API command I can invoke to start one of these? The API says there is a way to start and stop processors -- View this message in context:

Re: PutSolrContentStream Doc and DocID confusion

2016-04-20 Thread dale.chang13
/Note: I omitted a bunch of fields from the JSON document when using the Solr UI and curl./ I performed additional tests from the UI: Storing the entire JSON document using the UI resulted in an Http 400 Bad Request Response. By comparing our JSON document to the Solr schema, I saw the JSON

Re: PutSolrContentStream Doc and DocID confusion

2016-04-20 Thread dale.chang13
Hi Brian, Yes, the JSON object I am storing is a valid JSON document. The Content Payload is set to true and the value is: {"docid":"a1602677-fc7c-43ea-adba-c1ed945ede3d_1831"} I believe I would have gotten a JSON syntax error saying that the JSON object was invalid.

PutSolrContentStream Doc and DocID confusion

2016-04-20 Thread dale.chang13
While using PutSolrContentStream to store a JSON object in SolrCloud, I've been running into this issue of being unable to store a document. I've uploaded a solr schema that says that the field *docid* is required and a string. Attempting to store a document in solr, this is the error I get:

Re: Variable FlowFile Attributes Defined by FlowFile Content

2016-04-04 Thread dale.chang13
Essentially all of the information contained in a FlowFile's contents would be translated to attributes. I would like to pass in a generic delimited file with two rows: - the first row contains header names, - the second row contains values for each header corresponding to a single entry. The

Variable FlowFile Attributes Defined by FlowFile Content

2016-03-31 Thread dale.chang13
I see that the ExtractText Processor performs regular expressions on the FlowFile's content and adds the results as user-defined attributes. However, I was wondering if there was a way to avoid "hard-coding" these values. I was hoping of something along the lines where the key literal of the

Splitting Incoming FlowFile, Output Multiple FlowFiles

2016-03-31 Thread dale.chang13
My specific use-case calls for ingesting a CSV table with many rows and then storing individual rows into HBase and Solar. Additionally, I would like to avoid developing custom processors, but it seems like the SplitText and SplitContent Processors do not return individual flowfiles, each with

Re: Flow.xml.gz are Empty

2016-03-24 Thread dale.chang13
UPDATE: Here's the complete stack trace for from the slave node's app.log java.lang.Exception: Unable to load flow due to: java.io.IOException: org.apache.nifi.cluster.ConnectionException: Failed to connect node to cluster because local or cluster flow is malformed. at

Re: Flow.xml.gz are Empty

2016-03-24 Thread dale.chang13
I went through the nifi.properties for my ncm and two nodes and made sure that the "nifi.sensitive.props.key" values are all the same. However, I do not know how to make sure my ports are clear between the nodes. Could you help me or direct me to to some resources? >Make that your

Flow.xml.gz are Empty

2016-03-24 Thread dale.chang13
Good morning guys, I've been having trouble connecting two slave nodes to the ncm. In the logs, on the line before "/unable to load flow due to... ConnectionException... local or cluster flow is malformed/," I see /[Fatal Error] :3384:23: Character reference