Thanks James.. I am looking into permission issue and update the thread. I will also make the changes as you per your recommendation.
On Fri, Oct 28, 2016 at 10:23 AM, James Wing <jvw...@gmail.com> wrote: > From the screenshot and the error message, I interpret the sequence of > events to be something like this: > > 1.) ListS3 succeeds and generates flowfiles with attributes referencing S3 > objects, but no content (0 bytes) > 2.) FetchS3Object fails to pull the S3 object content with an Access > Denied error, but the failed flowfiles are routed on to PutS3Object (35,179 > files / 0 bytes in the "putconnector" queue) > 3.) PutS3Object is succeeding, writing the 0 byte content from ListS3 > > I recommend a couple thing for FetchS3Object: > > * Only allow the "success" relationship to continue to PutS3Object. > Separate the "failure" relationship to either loop back to FetchS3Object or > go to a LogAttibute processor, or other handling path. > * It looks like the permissions aren't working, you might want to > double-check the access keys or try a sample file with the AWS CLI. > > Thanks, > > James > > > On Fri, Oct 28, 2016 at 10:01 AM, Gop Krr <gop....@gmail.com> wrote: > >> This is how my nifi flow looks like. >> >> On Fri, Oct 28, 2016 at 9:57 AM, Gop Krr <gop....@gmail.com> wrote: >> >>> Thanks Bryan, Joe, Adam and Pierre. I went past this issue by switching >>> to 0.71. Now it is able to list the files from buckets and create those >>> files in the another bucket. But write is not happening and I am getting >>> the permission issue ( I have attached below for the reference) Could this >>> be the setting of the buckets or it has more to do with the access key. All >>> the files which are creaetd in the new bucket are of 0 byte. >>> Thanks >>> Rai >>> >>> 2016-10-28 16:45:25,438 ERROR [Timer-Driven Process Thread-3] >>> o.a.nifi.processors.aws.s3.FetchS3Object FetchS3Object[id=xxxxx] Failed >>> to retrieve S3 Object for StandardFlowFileRecord[uuid=yy >>> yyy,claim=,offset=0,name=xxxxx.gz,size=0]; routing to failure: >>> com.amazonaws.services.s3.model.AmazonS3Exception: Access Denied >>> (Service: Amazon S3; Status Code: 403; Error Code: AccessDenied; >>> Request ID: xxxxxxx), S3 Extended Request ID: lu8tAqRxu+ouinnVvJleHkUUyK6J6r >>> IQCTw0G8G6DB6NOPGec0D1KB6cfUPsj08IQXI8idtiTp4= >>> >>> 2016-10-28 16:45:25,438 ERROR [Timer-Driven Process Thread-3] >>> o.a.nifi.processors.aws.s3.FetchS3Object >>> >>> com.amazonaws.services.s3.model.AmazonS3Exception: Access Denied >>> (Service: Amazon S3; Status Code: 403; Error Code: AccessDenied; >>> Request ID: 0F34E71C0697B1D8) >>> >>> at >>> com.amazonaws.http.AmazonHttpClient.handleErrorResponse(AmazonHttpClient.java:1219) >>> ~[aws-java-sdk-core-1.10.32.jar:na] >>> >>> at >>> com.amazonaws.http.AmazonHttpClient.executeOneRequest(AmazonHttpClient.java:803) >>> ~[aws-java-sdk-core-1.10.32.jar:na] >>> >>> at >>> com.amazonaws.http.AmazonHttpClient.executeHelper(AmazonHttpClient.java:505) >>> ~[aws-java-sdk-core-1.10.32.jar:na] >>> >>> at >>> com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:317) >>> ~[aws-java-sdk-core-1.10.32.jar:na] >>> >>> at >>> com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:3595) >>> ~[aws-java-sdk-s3-1.10.32.jar:na] >>> >>> at >>> com.amazonaws.services.s3.AmazonS3Client.getObject(AmazonS3Client.java:1116) >>> ~[aws-java-sdk-s3-1.10.32.jar:na] >>> >>> at >>> org.apache.nifi.processors.aws.s3.FetchS3Object.onTrigger(FetchS3Object.java:106) >>> ~[nifi-aws-processors-0.7.1.jar:0.7.1] >>> >>> at >>> org.apache.nifi.processor.AbstractProcessor.onTrigger(AbstractProcessor.java:27) >>> [nifi-api-0.7.1.jar:0.7.1] >>> >>> at org.apache.nifi.controller.StandardProcessorNode.onTrigger(S >>> tandardProcessorNode.java:1054) [nifi-framework-core-0.7.1.jar:0.7.1] >>> >>> at org.apache.nifi.controller.tasks.ContinuallyRunProcessorTask >>> .call(ContinuallyRunProcessorTask.java:136) >>> [nifi-framework-core-0.7.1.jar:0.7.1] >>> >>> at org.apache.nifi.controller.tasks.ContinuallyRunProcessorTask >>> .call(ContinuallyRunProcessorTask.java:47) >>> [nifi-framework-core-0.7.1.jar:0.7.1] >>> >>> at org.apache.nifi.controller.scheduling.TimerDrivenSchedulingA >>> gent$1.run(TimerDrivenSchedulingAgent.java:127) >>> [nifi-framework-core-0.7.1.jar:0.7.1] >>> >>> at >>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) >>> [na:1.8.0_101] >>> >>> at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308) >>> [na:1.8.0_101] >>> >>> at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFu >>> tureTask.access$301(ScheduledThreadPoolExecutor.java:180) [na:1.8.0_101] >>> >>> at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFu >>> tureTask.run(ScheduledThreadPoolExecutor.java:294) [na:1.8.0_101] >>> >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >>> [na:1.8.0_101] >>> >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >>> [na:1.8.0_101] >>> >>> at java.lang.Thread.run(Thread.java:745) [na:1.8.0_101] >>> >>> On Fri, Oct 28, 2016 at 6:31 AM, Pierre Villard < >>> pierre.villard...@gmail.com> wrote: >>> >>>> Quick remark: the fix has also been merged in master and will be in >>>> release 1.1.0. >>>> >>>> Pierre >>>> >>>> 2016-10-28 15:22 GMT+02:00 Gop Krr <gop....@gmail.com>: >>>> >>>>> Thanks Adam. I will try 0.7.1 and update the community on the outcome. >>>>> If it works then I can create a patch for 1.x >>>>> Thanks >>>>> Rai >>>>> >>>>> On Thu, Oct 27, 2016 at 7:41 PM, Adam Lamar <adamond...@gmail.com> >>>>> wrote: >>>>> >>>>>> Hey All, >>>>>> >>>>>> I believe OP is running into a bug fixed here: >>>>>> https://issues.apache.org/jira/browse/NIFI-2631 >>>>>> >>>>>> Basically, ListS3 attempts to commit all the files it finds >>>>>> (potentially 100k+) at once, rather than in batches. NIFI-2631 >>>>>> addresses the issue. Looks like the fix is out in 0.7.1 but not yet in >>>>>> a 1.x release. >>>>>> >>>>>> Cheers, >>>>>> Adam >>>>>> >>>>>> >>>>>> On Thu, Oct 27, 2016 at 7:59 PM, Joe Witt <joe.w...@gmail.com> wrote: >>>>>> > Looking at this line [1] makes me think the FetchS3 processor is >>>>>> > properly streaming the bytes directly to the content repository. >>>>>> > >>>>>> > Looking at the screenshot showing nothing out of the ListS3 >>>>>> processor >>>>>> > makes me think the bucket has so many things in it that the >>>>>> processor >>>>>> > or associated library isn't handling that well and is just listing >>>>>> > everything with no mechanism of max buffer size. Krish please try >>>>>> > with the largest heap you can and let us know what you see. >>>>>> > >>>>>> > [1] https://github.com/apache/nifi/blob/master/nifi-nar-bundles/ >>>>>> nifi-aws-bundle/nifi-aws-processors/src/main/java/org/apache >>>>>> /nifi/processors/aws/s3/FetchS3Object.java#L107 >>>>>> > >>>>>> > On Thu, Oct 27, 2016 at 9:37 PM, Joe Witt <joe.w...@gmail.com> >>>>>> wrote: >>>>>> >> moving dev to bcc >>>>>> >> >>>>>> >> Yes I believe the issue here is that FetchS3 doesn't do chunked >>>>>> >> transfers and so is loading all into memory. I've not verified >>>>>> this >>>>>> >> in the code yet but it seems quite likely. Krish if you can verify >>>>>> >> that going with a larger heap gets you in the game can you please >>>>>> file >>>>>> >> a JIRA. >>>>>> >> >>>>>> >> Thanks >>>>>> >> Joe >>>>>> >> >>>>>> >> On Thu, Oct 27, 2016 at 9:34 PM, Bryan Bende <bbe...@gmail.com> >>>>>> wrote: >>>>>> >>> Hello, >>>>>> >>> >>>>>> >>> Are you running with all of the default settings? >>>>>> >>> >>>>>> >>> If so you would probably want to try increasing the memory >>>>>> settings in >>>>>> >>> conf/bootstrap.conf. >>>>>> >>> >>>>>> >>> They default to 512mb, you may want to try bumping it up to >>>>>> 1024mb. >>>>>> >>> >>>>>> >>> -Bryan >>>>>> >>> >>>>>> >>> On Thu, Oct 27, 2016 at 5:46 PM, Gop Krr <gop....@gmail.com> >>>>>> wrote: >>>>>> >>>> >>>>>> >>>> Hi All, >>>>>> >>>> >>>>>> >>>> I have very simple data flow, where I need to move s3 data from >>>>>> one bucket >>>>>> >>>> in one account to another bucket under another account. I have >>>>>> attached my >>>>>> >>>> processor configuration. >>>>>> >>>> >>>>>> >>>> >>>>>> >>>> 2016-10-27 20:09:57,626 ERROR [Flow Service Tasks Thread-2] >>>>>> >>>> org.apache.nifi.NiFi An Unknown Error Occurred in Thread >>>>>> Thread[Flow Service >>>>>> >>>> Tasks Thread-2,5,main]: java.lang.OutOfMemoryError: Java heap >>>>>> space >>>>>> >>>> >>>>>> >>>> I am very new to NiFi and trying ot get few of the use cases >>>>>> going. I need >>>>>> >>>> help from the community. >>>>>> >>>> >>>>>> >>>> Thanks again >>>>>> >>>> >>>>>> >>>> Rai >>>>>> >>>> >>>>>> >>>> >>>>>> >>>> >>>>>> >>> >>>>>> >>>>> >>>>> >>>> >>> >> >