Aldrin - regarding IRC [1]
We definitely need to get this info up on the website.
[1]
http://mail-archives.apache.org/mod_mbox/nifi-dev/201411.mbox/%3CCA%2BLyY57QFsxDxrHoNX7PGHd%2BFvX3ymQ%3DM8OtDnYxNeZEg3pAsA%40mail.gmail.com%3E
Thanks
Joe
On Sun, Nov 8, 2015 at 12:31 PM, Aldrin Piri wrote:
>
Hello
You can use provenance or a LogAttributes processor to see what the
headers are of the flow file after InvokeHTTP executes. You may find
the location header present as one of the attributes. If so then you
should be able to use that attribute to establish the URL it will hit
next time. Do
ocation attribute.
>
> If I curl the url and check the headers it does show the Location attribute.
>
> invokehttp.status.code 301
>
>
>
>
> On Sunday, November 8, 2015 4:58 PM, Joe Witt wrote:
>
>
> Hello
>
> You can use provenance or a LogAttributes pro
http://www.mkyong.com/java/java-httpurlconnection-follow-redirect-example/
> [2] https://issues.apache.org/jira/browse/NIFI-1086
>
> Joe
> - - - - - -
> Joseph Percivall
> linkedin.com/in/Percivall
> e: joeperciv...@yahoo.com
>
>
>
>
> On Sunday, November 8, 2015 9:53 PM,
Darren,
In short, yes I think NiFi can handle such a case in a generic sense quite well.
Read on for the longer response...
NiFi can process extremely large data, extremely large datasets,
extremely small data and high rates, variable sized data, etc.. It
makes this efficient by its design, how
Chakrader,
Yes the workflow today requires you to have removed those
processors/extensions from the flow before removing them from the
system classpath. In such a case it would be smooth and fine.
However, we recognize that users will not always know to do that or
remember to do that. Therefore,
Naveen
The approach Matt suggested will help. Perhaps a more complete
approach would be to be able to apply a true schema validation to JSON
documents. I did a quick search for apache friendly Java libraries
and found this [1]. Perhaps we should have a processor that lets
users validate JSON do
Darren,
A single NiFi instance (on one node or a cluster of 10+) can handle
*many* different flows.
Thanks
Joe
On Wed, Nov 11, 2015 at 10:00 AM, Darren Govoni wrote:
> Mark,
>Thanks for the tips. Appreciate it.
>
> So when I run nifi on a single server. It is essentially "one flow"?
> If I
ected linear flows have a way of growing
together and forming a true graph.
Thanks
Joe
On Wed, Nov 11, 2015 at 10:13 AM, Darren Govoni wrote:
> Thanks Joe.
>
> And it seems all the different flows would be seen on the one canvas, just
> not connected?
>
>
> On 11/11/201
Naveen,
For throughput can you state what the desired events/sec/node would be
for you and can you describe how the flowfile vs content vs prov repo
is setup on the machine it is running on?
Thanks
Joe
On Wed, Nov 11, 2015 at 1:21 PM, Madhire, Naveen
wrote:
> Thanks Mark. The workaround to have
> [1] https://issues.apache.org/jira/browse/NIFI-1155
>
>
> On Nov 12, 2015, at 2:32 AM, Christopher Hamm
> wrote:
>
> where does it dump? I cant get mine to expire fast enough. It is dev test I
> want to blow away but cant delete.
>
> On Fri, Nov 6, 2015 at 11:33 A
Right - PutHDFS definitely creates directories on the fly if necessary
(based on the path attribute of the flowfile) and if it has
permissions to do so.
Thanks
Joe
On Thu, Nov 12, 2015 at 8:34 AM, Bryan Bende wrote:
> I think PutHDFS always creates them so it isn't an option through the
> proper
Naveen,
PutFile does not support an option to append to an existing file on
the target file system at this time. It could perhaps be a reasonable
'conflict resolution strategy = append' that we would support. You
could consider in the interim as an alternative approach and one which
has much saf
We have to make this easier...
Maybe we should give someone access to an inline expression editor and see
the results. Like in regexpal...
On Nov 12, 2015 7:26 PM, "Charlie Frasure" wrote:
> Good call. I added trim() to the matches command, and it seems to have
> resolved the issue. I was che
x27;ve gone too far afield here, I've just felt the pains
> of users trying to get the right recognizers for their data fields.
>
> Cheers,
> Matt
>
> Sent from my iPhone
>
> On Nov 12, 2015, at 7:54 PM, Joe Witt wrote:
>
> We have to make this easier...
>
&
realize ScanContent worked that way, very
> cool!
>
> Sent from my iPhone
>
>> On Nov 12, 2015, at 8:40 PM, Joe Witt wrote:
>>
>> User Experience - everything we do needs to be about continually
>> improving the user experience. So yes for sure if you've got
Adam,
Are you on a recent master build?
Thanks
Joe
On Fri, Nov 13, 2015 at 12:27 AM, Adam Lamar wrote:
> Hi everybody!
>
> I'm following up from my previous thread about high CPU usage in GetSQS. I
> ran into high CPU usage while developing a patch for that processor, and
> while investigating
sorry - i see now :-)
Thanks for the analysis. Will dig in.
Joe
On Fri, Nov 13, 2015 at 12:28 AM, Joe Witt wrote:
> Adam,
>
> Are you on a recent master build?
>
> Thanks
> Joe
>
> On Fri, Nov 13, 2015 at 12:27 AM, Adam Lamar wrote:
>> Hi everybody!
>>
Indus,
Let's project out one order of magnitude. That puts you at just a bit
more than 500 lookups per second. You will want to consider the
properties of the data itself and the properties of the database.
This helps you decide on the appropriate caching and querying
strategy. If the data as-i
Naveen
You can save the flow as a template and import that. It today has
limitations such as environmental mapping is too limited and you
should know it won't copy sensitive properties.
Another option is to copy the flow.xml.gz from the current setup into
production.
We need to do more though t
Hello,
So the admittedly lame response is that we don't provide much there
yet because generally we are very flexible. Consider two vcpu, 4-8GB
ram, two disk partitions of say 100-200GB as a small/simple install.
If you want to push higher volumes perhaps the next step up is say 4
vcpu, 16GB ram
Subhash,
I don't think enough folks are familiar here to help. Please feel
free to file a JIRA for that mongo processor to add support for this
option.
Thanks
Joe
On Tue, Nov 17, 2015 at 3:19 AM, Subhash Parise
wrote:
> Hi NiFi experts,
>
> Do any one know how to pass --sslInvalidCertifcates o
and that behavior
> is not possible with Storm ( right ? pls tell me if I am wrong )
>
> Philippe
> Best regards
>
> -Message d'origine-
> De : Joe Witt [mailto:joe.w...@gmail.com]
> Envoyé : mercredi 11 novembre 2015 03:28
> À : users@nifi.apache.org
> Obj
Charlie,
The fact that this is confusing is something we agree should be more
clear and we will improve. We're tackling it based on what is
mentioned here [1].
[1]
https://cwiki.apache.org/confluence/display/NIFI/Interactive+Queue+Management
Thanks
Joe
On Thu, Nov 19, 2015 at 10:30 PM, Corey
ed is: file -b --mime-encoding ${filename}
> in the working directory: ${absolute.path}
>
> Now that the file is no longer in the source directory when the processor
> fires, the command is broken. I could PutFile somewhere temporarily; is
> there a better way?
>
> On
n run an
> external command against the physical file such as
> /opt/nifi/somedir/12345.uuid? Would that info be in an attribute somewhere?
> It just seems wasteful to make an extra copy of the file, in order to run a
> read-only command on it, then delete it. If ListFiles is still the
Chris,
Questions:
Is this is NiFi 0.2.1?
Does it always fail when trying to view content?
I just built 021 and verified it worked as expected. In addition to
the classloading issue i notice the JSP null found issues at the
bottom of what you pasted. I wonder if the work directory was
incomplete
Chris,
If you run into a case of a stuck thread we'd love to see the stack
trace. You can generate one by running 'bin/nifi.sh dump' and sending
us the logs. I believe it is bootstrap.log specifically.
If you have a processor which immediately gets stuck after startup
where it takes a thread an
lines. I guess that might be a limitation I'll have
> to find a workaround for. I've managed to recover it but I only noticed this
> trying to troubleshoot why the system stopped running the jobs properly.
>
> On Tue, 24 Nov 2015 at 1:06 AM, Joe Witt wrote:
>>
>> Chri
un "bin/nifi.sh dump" it is often easier to specify a
> filename there
> such as "bin/nifi.sh dump thread-dump.txt" so that it writes the thread dump
> to the
> filename specified, rather than writing it to the logs.
>
> Thanks
> -Mark
>
>
>> On N
Chris,
We have support for interaction with a Kerberos enabled HDFS cluster
today. For authentication within NiFi as of right now it is via 2-way
SSL. This is powerful yet cumbersome to setup unless you're in an
environment with a robust PKI infrastructure.
So for the upcoming NiFi 0.4.0 releas
Charlie,
We can definitely enter a JIRA request for you however we like to let folks
do it themselves if they're willing. You can do so here
https://issues.apache.org/jira/browse/NIFI/
It is a good way for you to fully articulate your idea, add examples
before/after, that sort of thing.
Thanks
Charlie,
Can give some pointers on how to get in the ballpark with this but
want to make sure we have a good alignment of purpose here. NiFi has
from time to time come up as an intuitive way to build an archive
management tool and it is always "not quite right" because of the
subtle differences b
Hello,
Ports on the root group automatically become remote group ports.
Input ports can receive input from external sources and output ports
can be used by external sources to pull data. So long as those
external sources are using the site-to-site client lib you should be
good.
A longer descript
Chris,
It is something which occurs automatically and behind the scenes.
Under normal circumstances there will be many FlowFiles written to the
same content claim they'll just each have different offsets. It is
more aligned with how disks work in terms of efficiently writing data,
efficiently rea
Matt,
Any chance you'd be interested in turning this into a PR or patch for
formal submission? There are some commented code blocks and todo
sections I noticed in a quick scan but if NATs is as fast as it
suggests and offers an ASLv2 client lib then this seems like a great
thing to have available
[posting on behalf of Kacem as there seems to be some issue we've not
sorted with his mailing attempts]
Hello everyone,
So my case is as the following, i'm trying to retrieve JSON files from a
social network (Facebook), and based on some values in the file of the HTTP
response i want to invoke a
All please ignore this thread i started. Kacem was able to get his
post through. Please do reply to his original post.
On Tue, Dec 1, 2015 at 10:41 AM, Joe Witt wrote:
> [posting on behalf of Kacem as there seems to be some issue we've not
> sorted with his mailing attempts]
>
&g
adding to Oleg's comment we should have this in a formal nifi release
very soon. In fact the great NIFI-655 push we've been holding on just
went down so think we're going to kick out an RC very soon.
Thanks
Joe
On Tue, Dec 1, 2015 at 11:45 AM, Oleg Zhurakousky
wrote:
> Douglass
>
> Just looked
Charlie,
You can double click on the line and it will create a bend point for
you. You can also double click on that bend point to remove it.
Not the first time someone has asked and i don't see it in the docs.
Anybody know if this is in the docs somewhere?
Thanks
Joe
On Tue, Dec 1, 2015 at 4:
Hello
So we have FetchS3 and PutHDFS and a series of interesting in between
processes to help. So that would get you most of the way there. How
to get the listing/know what to pull from S3? That part I'm not sure
about.
This would make for a great example/template for us to post (as would
the
Pompillio
Keep in mind too you can look at that summary page and it will give
you a tabular listing of various types of components. You can sort
processors by which is using the most threads (at that very moment
anyway), or has run the most tasks over 5 minutes or taken the most
time over 5 minut
Michael,
I *think* you could just replace the instance of the distributed cache
service it is using to be a new one.
That feels a bit cumbersome off the cuff but what do you think?
Thanks
Joe
On Thu, Dec 3, 2015 at 7:50 PM, Michael Dyer wrote:
> Is there a way to invalidate the cache for ListH
Understood. As you look at the app and you're thinking 'hey i'd like
to wipe this cache' how would you like that to work? Just looking for
ideas on how to do this in an intuitive way.
On Thu, Dec 3, 2015 at 8:00 PM, Michael Dyer wrote:
> Yeah, I was trying to avoid that.
>
> I'm attempting to r
Chandu,
In building a template to support this I'm finding it more involved than
desired. If I get that done will post it unless someone else does so
sooner. However, the only reason it is more involved than I'd like is that
the ReplaceText processor does almost everything we need here. It can
Your use case description was excellent and helped expose a powerful and
simple feature enhancement.
https://issues.apache.org/jira/browse/NIFI-1249
Thanks
Joe
On Thu, Dec 3, 2015 at 9:21 PM, Joe Witt wrote:
> Chandu,
>
> In building a template to support this I'm finding it
Manish,
As you have laid it out to work i think it would be harder than it
should be. However, that is in part due to not taking advantage of
what NiFi's strengths are in building something reliable for that.
The JS being called - it may be possible to simply implement its logic
to pull data dire
h, group2 is date and group1
> is ddMMM
>
>
>
>
>
>
>
> *From:* Joe Witt [mailto:joe.w...@gmail.com]
> *Sent:* Thursday, December 03, 2015 7:00 PM
> *To:* users@nifi.apache.org
> *Subject:* Re: Content replacement
>
>
>
> Your use case description w
Chris,
I think there are some important details missing. In the case of auto
resume being false on restart there is no opportunity for those
processors to be stuck as they'll never receive a thread. Can you
share a thread dump? Feel free to create a JIRA for what you're
seeing and attach the th
Chris,
This is a really common approach and it is backpressure that makes it
work and be so successful. So definitely want to help you get to the
bottom of it. Can you perhaps make a template of your config and
share that so we can see what might be happening?
Thanks
Joe
On Fri, Dec 4, 2015 at
Is anyone aware of any reasons not to do this?
On Fri, Dec 4, 2015 at 8:59 AM, Nigel Jones wrote:
> Would it be ok if I asked gmane to mirror these nifi lists on the gmane nntp
> server?
>
> I personally find it a little easier to work through & catch up with
> discussions - though doing this sh
Charlie,
"1. Would it be more efficient to let the files queue up, or to try to
match the process rate with timing or back pressure?"
From an efficiency perspective each transaction has a cost. What you
have to consider is whether you seek lower latency or higher throughput.
The default settin
Louis-Etienne,
My initial thought is your idea with MergeContent is the right one.
However, the issue there is not just the combining of the data but the
'what does merging truly mean in that case'. So it is a bit undefined
what the next step will be. Merge the content? If so, how? What is
the
> message from Topic2.
>
> I'll dig in the code of the PR and the MergeContent processor in order to
> have a better understanding.
>
> Was that JIRA scheduled for a specific milestone? It would probably be a
> great addition but maybe it require a lot of change that I
Is there any reason we'd not just let the user enter an override URL
to pull from?
On Tue, Dec 8, 2015 at 9:02 PM, Bob Zhao wrote:
> Thanks for the quick response, Aldrin!
>
> I know we are not NSA. But there are at least 2 3rd parties who provide the
> Firehose stream(paid).
> Does NIFI have any
Hello
The GetFTP processor itself doesn't support idempotent behavior.
We've moved to a 'List' and 'Fetch' model for processors where that is
useful as you'll see in ListSFTP and FetchSFTP. We have not built
that pattern for GetFTP. It was originally built with the idea that
it would be taking t
Hello,
The Apache NiFi team would like to announce the release of Apache NiFi
0.4.0.
Apache NiFi is an easy to use, powerful, and reliable system to process and
distribute data. Apache NiFi was made for dataflow. It supports highly
configurable directed graphs of data routing, transformation, and
4 AM, Nigel Jones
> wrote:
>
>
> I have no experience of nabble but I'm also open to any similar
> suggestions? . I see some apache mailing lists there too
>
> On Sat, 5 Dec 2015 03:55 Tony Kurc wrote:
>
> Similar to nabble, right?
> On Dec 4, 2015 10:18 PM, "Joe
In addition to Alan's question to understand what the actual error is
please note that when i try your config I get complains about the target
dataset uri. If i have for instance
dataset:hive:default/ctest1
As shown in your screenshot, the validation error I see is:
'Target dataset URI' valida
t;>> when an object arrives via SQS. Then have GetSQS get that notification
>>> and then pull the data via FetchS3Object.
>>> So you could fairly easily setup a GetSQS -> EvaluateJSONPath ->
>>> FetchS3Object -> PutHDFS. That would require that SQS be setup tho
y searches Nabble and
>> Markmail come up much more often and more pleasant to use.
>>
>> Nabble can be allowed to do 2-way posts and is also embeddable, nice light
>> theme.
>>
>> Nabble (forum-like) http://apache-nifi.1125220.n5.nabble.com/
>> Markmail is a
gt;>
>> ·Hadoop Configuration files:
>> /opt/cloudera/parcels/CDH-5.4.0-1.cdh5.4.0.p0.27/lib/hadoop/etc/hadoop/hdfs-site.xml,/opt/cloudera/parcels/CDH-5.4.0-1.cdh5.4.0.p0.27/lib/hadoop/etc/hadoop/core-site.xml,/opt/cloudera/parcels/CDH-5.4.0-1.cdh5.4.0.p0.27/etc/hive/conf.dist/hi
Hello,
The Apache NiFi team would like to announce the release of Apache NiFi 0.4.1.
This is a bug/incremental release building on Apache NiFi 0.4.0.
Apache NiFi is an easy to use, powerful, and reliable system to
process and distribute data. Apache NiFi was made for dataflow. It
supports highly
e-3.2.10.jar
>> datanucleus-rdbms-3.2.9.jar
>> hadoop-mapreduce-client-common-2.6.0-cdh5.4.2.jar
>> hadoop-mapreduce-client-core-2.6.0-cdh5.4.2.jar
>> hive-exec-1.1.0-cdh5.4.2.jar
>> hive-metastore-1.1.0-cdh5.4.2.jar
>> hive-shims-1.1.0-cdh5.4.2.jar
>> jdo-
Hello Chakri
I updated the downloads page to provide a convenient link to the
release archives.
Thanks Chris for the pointer to that!
Joe
On Mon, Dec 28, 2015 at 6:36 PM, Chakrader Dewaragatla
wrote:
> Thanks Chris.
>
> From: Chris Lim
> Reply-To: "users@nifi.apache.org"
> Date: Monday, Dece
Hello Obaid,
At 6 TB/day and average size of 2-3GB per dataset you're looking at a
sustained rate of 70+MB/s and a pretty low transaction rate. So well
within a good range to work with on a single system.
'I's there any way to by pass writing flow files on disk or directly
pass those files to HD
kup data does not change frequently so
> your built-in caching strategy sounds great. Would you expand on how to
> implement the built-in caching mechanism as well as refreshing the cache in
> NiFi please?
>
> Thanks,
>
> Indus
>
> On Sat, Nov 14, 2015 at 12:24 PM, Joe Witt
Mans
The stats shown on the component are fixed at a 5 minute window. The
request to be able to alter that does come up from time to time. The
idea there is that from a user experience perspective those numbers
are the most relevant to show as a recurring value because over time
you become famil
cessing.
> Then I randomly changed the content of the file and it was immediately
> processed.
>
> My question is what is the new file selection criteria for "ListFile" ? Can
> I change it only to file name ?
>
> Thanks in advance.
>
> -Obaid
>
>
>
>
ate(as current one) unique file names, checksum etc.
>
> Thanks again man.
> -Obaid
>
>
> On Monday, January 4, 2016, Joe Witt wrote:
>>
>> Hello Obaid,
>>
>> The default behavior of the ListFile processor is to keep track of the
>> last modified time of th
Hello Chakri,
Take a look here
https://cwiki.apache.org/confluence/display/NIFI/Upgrading+NiFi
And here
https://cwiki.apache.org/confluence/display/NIFI/Migration+Guidance
Thanks
Joe
On Mon, Jan 4, 2016 at 8:02 PM, Chakrader Dewaragatla
wrote:
> Hi – What is the best way to upgrade from 0.
Mon, Jan 4, 2016 at 9:11 AM, obaidul karim wrote:
>>
>> Hi Joe,
>>
>> Yes, symlink is another option I was thinking when I was trying to use
>> getfile.
>> Thanks for your insights, I will update you on this mail chain when my
>> entire workflow completes
Vincent,
As Aldrin mentioned there is an effort underway to help with
integration testing. But I'm not sure yet if it will help with what
you're looking to do from a performance baseline perspective. I
totally get Mark's cautionary comments but given your response it does
seem like there is some
Christopher,
Is there any error/feedback showing up in the UI or in the logs?
Thanks
Joe
On Wed, Jan 6, 2016 at 5:19 PM, Christopher Hamm
wrote:
> What am I doing wrong with hooking up my activemq jms get template? I put
> stuff into the activeMQ and NIFI wont get it. Using 0.4.1.
>
> --
> Sinc
s the condition that triggers it. May be if you shut down NIFI,
>> delete all the logs and restart so you can get a fresh data. . .
>>
>> Cheers
>> Oleg
>>
>> On Jan 9, 2016, at 1:42 AM, Christopher Hamm
>> wrote:
>>
>> Here are the
Hello Sudeep,
"Which NiFi processor can I use to split each record (separated by a
new line character)"
For this the SplitText processor is rather helpful if you want to
split each line. I recommend you do two SplitText processors in a
chain where one splits on every 1000 lines for example and
Lars,
First, welcome to NiFi.
So you'll initiate the flow from NiFi through some manual trigger and
that part makes sense/sounds straightforward. But then the question
is 'once the flow is done' I'd like it to shut down. What are the
processors/things that begin the flow? We've recently had so
Chakri,
Would love to hear what you've learned and how that differed from the
docs themselves. Site-to-site has proven difficult to setup so we're
clearly not there yet in having the right operator/admin experience.
Thanks
Joe
On Sun, Jan 10, 2016 at 3:41 PM, Chakrader Dewaragatla
wrote:
> I w
Hello Roland
I don't believe we have out of the box support for that but it would
make for likely a well scoped custom processor.
You mention support for sort syntax. Can you share what that might look like?
Thanks
Joe
On Sun, Jan 10, 2016 at 9:45 PM, 彭光裕 wrote:
> Hi,
>
> I have a jso
while ((line = reader.readLine()) != null) {
> if(line.trim().length() > 0 ) {
> out.write( parseLine(line, seperator, quote, escape,
> maskColumns).getBytes() );
> }
> };
> out.flush();
>}
>});
>
>
>
>
> Thanks in advance.
> -Obaid
>
&g
Sudeep,
Hello. At this time there are no apache nifi redis processors to
push/pull data with Redis that I am aware of. Something you might be
interested in contributing or contributing to?
Thanks
Joe
On Wed, Jan 13, 2016 at 12:13 AM, sudeep mishra
wrote:
> Hi,
>
> Do we have any processor to
ask, just pass through
> if(reader.ready()) {
> writer.write(reader.readLine());
> writer.write(lineEndingBuilder.toString());
> }
> }
> // decide about empty line earlier
> String line;
> while ((line = reader.readLine()) != null) {
> writer.write(parseLine(line, seperator, quote, escap
attached jstat & iostat output.
>
> So far it seems to me that it is CPU bound. However, your eyes are better
> tan mine :).
>
> -Obaid
>
> On Thu, Jan 14, 2016 at 11:51 AM, Joe Witt wrote:
>>
>> Hello
>>
>> Let's narrow in on potential issue
gt; java.arg.2=-Xms5120m
> java.arg.3=-Xmx10240m
>
> latest jstate & iostate output are attached.
> To me it is still slow, no significant improvements.
>
> -Obaid
>
> On Thu, Jan 14, 2016 at 12:41 PM, Joe Witt wrote:
>>
>> Obaid,
>>
>> Great so this is
see in UI.
>> >> Also, the fact that you see NPE is definitely a bug that we have to fix
>> >> (user’s should never see NPE), so that can be filed. What I am trying
>> >> to
>> >> figure out is the condition that triggers it. May be if you shut down
>
Chakri,
The most telling bits of the logs are probably below that. Could
possibly attach the full nifi-app.log (assuming nothing sensitive is
in it) or search lower for the root cause of the exception.
Even if provenance data is corrupted the problematic sections should
be safely ignored so let'
Hello
Sounds like you're on track now. Just wanted to correct one comment made
ListenHTTP by default does not set an input port and instead it is a
required input.
Thanks
Joe
On Tue, Jan 19, 2016 at 7:49 AM, sudeep mishra wrote:
> Listening on port 81 on Linux RHEL 6.5
>
> On Tue, Jan 19, 2
Chakri
A user is associated with one or more roles. One of those roles a
user could have is simply read-only. In a typical organization the
majority of users actually have 'read-only' like a watch person or
monitoring service and so on. Very few users would have roles such as
provenance or dfm
Frank,
Absolutely. Have seen several of these of various types of the years.
It usually works best if the processor itself calls out/polls some
endpoint to get the latest instruction set. This approach tends to
scale better, means clustered systems work smoothly as well, and it is
nice from a re
14 PM, obaidul karim wrote:
>>
>> Joe,
>> I am doing some optimizations on my csv processing.
>> Let clear them out then I will share the final version.
>>
>> -Obaid
>>
>>
>> On Thursday, January 14, 2016, Joe Witt wrote:
>>>
>>>
Kyle
Let us know if that doesn't get you what you need. We have a decent set of
templates but I didn't see one that demonstrates interaction with amazon
services.
Thanks
Joe
On Jan 30, 2016 12:56 PM, "Joey Frazee" wrote:
> Kyle,
>
> I think you can do what you want right now without ListS3 by
Ian
Makes sense. Happy to help put together a template to see if we can
get you closer. Working on some 050 release items at the moment but
will flag this to come back
Thanks
Joe
On Fri, Jan 29, 2016 at 8:43 AM, Hellstroem Christian (AE/PJ-ASC)
wrote:
> Hi,
>
>
>
> I’m just starting out with
; IgnitionOne - Marketing Technology. Simplified.
> Office: 1545 Peachtree St NE, Suite 500 | Atlanta, GA | 30309
>
>
> From: Joe Witt
> Reply-To: "users@nifi.apache.org"
> Date: Saturday, January 30, 2016 at 2:06 PM
> To: "users@nifi.apache.org"
> Subject: Re
github.io/2016-01-30-monitoring-an-s3-bucket-in-apache-nifi/
>
> Hope that helps,
> Adam
>
>
> On 1/30/16 9:29 PM, Joe Witt wrote:
>>
>> Kyle,
>>
>> The ideal case for communicating how to do this would be both a
>> template and an associated doc. Great
Jeroen
Definitely empathize with how frustrating it can be to have components
which get you so close but not quite there. Very happy to suggest
that NIFI-210 (due for inclusion for the upcoming 050 release - in
fact merging it in as we speak) will give you the ability to write
scripts in-line in
?
>
>
> Kind regards,
>
>
> Jeroen.
>
>
>
>
> From: Joe Witt
> Sent: Tuesday, February 2, 2016 3:01 AM
> To: users@nifi.apache.org
> Subject: Re: Nesting and AttributesToJSON
>
> Jeroen
>
> Definitely empathize with how frustrating it can be to have com
Andre
I believe until this next release which is 0.5.0 merge content did not
allow expression language statements as the correlation attribute. By
using an expression language statement there it is matching everything
basically.
For now you just put 'syslog.hostname' there instead.
Make sense?
result.
>
> I wonder if this is linked to
> https://issues.apache.org/jira/browse/NIFI-1438 ?
>
>
>
>
>
> On Mon, Feb 8, 2016 at 12:55 AM, Joe Witt wrote:
>
>> Andre
>>
>> I believe until this next release which is 0.5.0 merge content did not
>> allow expres
There is also the provenance data itself which captures this and there
is the LogAttributes processor.
Multiple options to get the information about the flow out.
Thanks
Joe
On Thu, Feb 11, 2016 at 10:34 AM, Andrew Grande wrote:
> Take a look at ReportingTask, this sounds like a good fit. Much h
Jeff,
This is definitely a strong use case for nifi.
It might be that InvokeHTTP is the better choice here.
If what you'd like to do is effectively throttle the rate at which you
hit the web service with the InvokeHttp calls you can schedule that
processor to run as often as you like (for exampl
401 - 500 of 1117 matches
Mail list logo