Right.  So NiFi is giving access to an InputStream.  Things like mark and
reset type operations will need to be wrapped.  If you have access to the
client code being used or a pointer to it that would be helpful in terms of
figuring out what it is doing.  I am confident this is not any sort of
issue with NiFi itself but we can still try to help find the root issue.

thanks

On Wed, Sep 7, 2016 at 10:05 AM, Tony Kurc <[email protected]> wrote:

> I won't have access to an IDE for the most of today, or an active azure
> account so I won't be able to debug myself in short order, but I'm
> suspecting the same thing joe alluded to at this point - this being an
> issue with the client API, possibly complicated by with the way NiFi allows
> access to the underlying bytes to the file.
>
> On Tue, Sep 6, 2016 at 6:21 PM, Kumiko Yada <[email protected]> wrote:
>
>> Lorenzo was not clear how we tested.  I wrote the sample java program
>> using the Azure SDK, then uploaded the 50 MB file and it’s worked without
>> error.   Nifi custom processor used the same SDK code; however, it’s
>> failing when the custom processor is tried to uploaded 50 MB file.
>>
>>
>>
>> Thanks
>>
>> Kumiko
>>
>>
>>
>> *From:* Tony Kurc [mailto:[email protected]]
>> *Sent:* Tuesday, September 6, 2016 3:02 PM
>> *To:* Kumiko Yada <[email protected]>
>> *Cc:* [email protected]; Joe Witt <[email protected]>; #Operations
>> Automation and Tools <#[email protected]>; Kevin Verhoeven <
>> [email protected]>
>>
>> *Subject:* RE: Issue writing file (~50mb) to azure data lake with Nifi
>>
>>
>>
>> I apologize if I'm missing something, as I'm trying to read the code on
>> my phone, but it looks like the test script is using a different api call
>> to perform the upload - did you already test using the same call in your
>> script?
>>
>>
>>
>> On Sep 6, 2016 5:48 PM, "Kumiko Yada" <[email protected]> wrote:
>>
>> Here is the code:  https://github.com/kyada1/ADL_UploadFile.
>>
>>
>>
>> I removed the following values for the security reason.
>>
>>
>>
>> *final static *String *ADLS_ACCOUNT_NAME *= *""*;
>> *final static *String *RESOURCE_GROUP_NAME *= *""*;
>> *final static *String *LOCATION *= *""*;
>> *final static *String *TENANT_ID *= *""*;
>> *final static *String *SUBSCRIPTION_ID *=  *""*;
>> *final static *String *CLIENT_ID *= *""*;
>> *final static *String *CLIENT_SECRET *= *""*; */*
>>
>>
>>
>> Thanks
>>
>> Kumiko
>>
>>
>>
>> *From:* Tony Kurc [mailto:[email protected]]
>> *Sent:* Tuesday, September 6, 2016 2:41 PM
>> *To:* Kumiko Yada <[email protected]>
>> *Cc:* Joe Witt <[email protected]>; [email protected]; #Operations
>> Automation and Tools <#[email protected]>; Kevin Verhoeven <
>> [email protected]>
>> *Subject:* RE: Issue writing file (~50mb) to azure data lake with Nifi
>>
>>
>>
>> I was referring to this: "Then we wrote a java sdk script to upload this
>> same file without Nifi into the data lake and it worked successfully."
>>
>> Is that code somewhere?
>>
>>
>>
>> On Sep 6, 2016 5:38 PM, "Kumiko Yada" <[email protected]> wrote:
>>
>> I didn’t add any test code.  This custom controller and processor is
>> working for a small size file.
>>
>>
>>
>> Thanks
>>
>> Kumiko
>>
>>
>>
>> *From:* Tony Kurc [mailto:[email protected]]
>> *Sent:* Tuesday, September 6, 2016 2:32 PM
>> *To:* [email protected]
>> *Cc:* Joe Witt <[email protected]>; #Operations Automation and Tools <#
>> [email protected]>; Kevin Verhoeven <[email protected]>
>> *Subject:* RE: Issue writing file (~50mb) to azure data lake with Nifi
>>
>>
>>
>> I didn't see the test script that worked in the source code - did I miss
>> it, or is it not in the tree?
>>
>>
>>
>> On Sep 6, 2016 3:17 PM, "Kumiko Yada" <[email protected]> wrote:
>>
>> Joe,
>>
>>
>>
>> Here is the log (there was no callstack related to this error) and code,
>> https://github.com/kyada1/dl_sdkworkaround/tree/master/nifi-
>> azure-dlstore-bundle.
>>
>>
>>
>> 2016-09-06 12:06:50,508 INFO [NiFi Web Server-19]
>> c.s.j.s.i.application.WebApplicationImpl Initiating Jersey application,
>> version 'Jersey: 1.19 02/11/2015 03:25 AM'
>>
>> 2016-09-06 12:07:00,991 INFO [StandardProcessScheduler Thread-1]
>> o.a.n.c.s.TimerDrivenSchedulingAgent Scheduled
>> PutFileAzureDLStore[id=00dd95dc-0157-1000-8ab1-2de88c159b55] to run with
>> 1 threads
>>
>> 2016-09-06 12:07:01,545 INFO [Flow Service Tasks Thread-1]
>> o.a.nifi.controller.StandardFlowService Saved flow controller
>> org.apache.nifi.controller.FlowController@28f414a1 // Another save
>> pending = false
>>
>> 2016-09-06 12:07:01,904 INFO [pool-27-thread-1]
>> c.m.aad.adal4j.AuthenticationAuthority [Correlation ID:
>> 564fb5ec-643b-43e6-ab68-59f259a4843a] Instance discovery was successful
>>
>> 2016-09-06 12:08:05,988 ERROR [Timer-Driven Process Thread-1]
>> n.a.d.processors.PutFileAzureDLStore 
>> PutFileAzureDLStore[id=00dd95dc-0157-1000-8ab1-2de88c159b55]
>> File was not created: /kumiko/test/20160906120701022.txt
>> com.microsoft.azure.management.datalake.store.models.AdlsErrorException:
>> Invalid status code 404
>>
>> 2016-09-06 12:08:12,541 INFO [Provenance Maintenance Thread-1]
>> o.a.n.p.PersistentProvenanceRepository Created new Provenance Event
>> Writers for events starting with ID 618
>>
>> 2016-09-06 12:08:12,838 INFO [Provenance Repository Rollover Thread-1]
>> o.a.n.p.PersistentProvenanceRepository Successfully merged 16 journal
>> files (1 records) into single Provenance Log File .\provenance_repository\
>> 617.prov in 305 milliseconds
>>
>> 2016-09-06 12:08:12,838 INFO [Provenance Repository Rollover Thread-1]
>> o.a.n.p.PersistentProvenanceRepository Successfully Rolled over
>> Provenance Event file containing 1 records
>>
>> 2016-09-06 12:08:21,148 INFO [NiFi Web Server-28]
>> o.a.n.controller.StandardProcessorNode Stopping processor: class
>> nifi.azure.dlstore.processors.PutFileAzureDLStore
>>
>>
>>
>> Thanks
>>
>> Kumiko
>>
>>
>>
>> *From:* Joe Witt [mailto:[email protected]]
>> *Sent:* Friday, September 2, 2016 2:56 PM
>> *To:* [email protected]
>> *Cc:* #Operations Automation and Tools <#[email protected]>; Kevin
>> Verhoeven <[email protected]>
>> *Subject:* Re: Issue writing file (~50mb) to azure data lake with Nifi
>>
>>
>>
>> Lorenzo
>>
>> Without seeing the code and logs it would be very difficult to help.
>> nifi has no trouble by design writing large files (GBs) to many things
>> including hdfs so the issue is probably in how this client library
>> interacts with the data stream.
>>
>>
>>
>> On Sep 2, 2016 4:19 PM, "Lorenzo Peder" <[email protected]> wrote:
>>
>> Hi All,
>>
>>
>>
>> We’ve run into an issue uploading a larger file (~50Mb) into an Azure
>> Data Lake using a custom processor in nifi 0.7-1.0. This custom processor
>> has worked consistently for smaller files, but once encountered with this
>> larger file, it spits http error 404 (file not found). Eventually a minor
>> portion of the file wrote to the data lake.
>>
>> We used fiddler to capture network traffic between Nifi and the Azure
>> data lake while the processor was running and captured http error 204 (no
>> contents).
>>
>> Then we wrote a java sdk script to upload this same file without Nifi
>> into the data lake and it worked successfully.
>>
>> These findings lead us to believe that this issue is occurring within
>> Nifi, if someone could please point us in the right direction in resolving
>> this issue it would be greatly appreciated.
>>
>>
>>
>> Thank you,
>>
>>
>>
>> Lorenzo Peder
>>
>> Operations Analyst, Campaign Operations & Services
>>
>>
>>
>> 425.974.1363 : Office
>>
>> 425.260.5027 : Mobile
>>
>> www.ds-iq.com
>>
>>
>>
>> Dynamic Shopper Intelligence
>>
>>
>>
>> This e-mail may contain confidential or privileged information.
>>
>> If you are not the intended recipient, please notify the sender
>> immediately and then delete this message.
>>
>>
>>
>>
>>
>>
>

Reply via email to