On Mon, May 27, 2019 at 2:24 PM 青雉(祁明良) <[email protected]> wrote:
>
> Just now I try to use the PROCESS environment type, the Flink taskmanager 
> complains about "/tmp/beam-artifact-staging/job_xxx" not found. And I found 
> this directory is only created on the machine with beam job endpoint. I guess 
> maybe I should set the artifact-dir to a hdfs location, but no luck for me:(

Yes, you need to set your artifact staging directory (the
--artifacts-dir flag) to something visible to both the job server and
the workers. Did you try that?

> I don’t know if the following error message from job endpoint is related when 
> submitting the job.
>
> Error from job endpoint:
> ---------------
> [grpc-default-executor-0] ERROR 
> org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService - 
> Encountered Unexpected Exception for Invocation 
> job_09aa2abd-0bc0-4994-a8b7-130156e4c13c
> org.apache.beam.vendor.grpc.v1p13p1.io.grpc.StatusException: NOT_FOUND
> at 
> org.apache.beam.vendor.grpc.v1p13p1.io.grpc.Status.asException(Status.java:534)
> at 
> org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService.getInvocation(InMemoryJobService.java:341)
> at 
> org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService.getStateStream(InMemoryJobService.java:262)
> at 
> org.apache.beam.model.jobmanagement.v1.JobServiceGrpc$MethodHandlers.invoke(JobServiceGrpc.java:770)
> at 
> org.apache.beam.vendor.grpc.v1p13p1.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
> at 
> org.apache.beam.vendor.grpc.v1p13p1.io.grpc.PartialForwardingServerCallListener.onHalfClose(PartialForwardingServerCallListener.java:35)
> at 
> org.apache.beam.vendor.grpc.v1p13p1.io.grpc.ForwardingServerCallListener.onHalfClose(ForwardingServerCallListener.java:23)
> at 
> org.apache.beam.vendor.grpc.v1p13p1.io.grpc.ForwardingServerCallListener$SimpleForwardingServerCallListener.onHalfClose(ForwardingServerCallListener.java:40)
> at 
> org.apache.beam.vendor.grpc.v1p13p1.io.grpc.Contexts$ContextualizedServerCallListener.onHalfClose(Contexts.java:86)
> at 
> org.apache.beam.vendor.grpc.v1p13p1.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
> at 
> org.apache.beam.vendor.grpc.v1p13p1.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:707)
> at 
> org.apache.beam.vendor.grpc.v1p13p1.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
> at 
> org.apache.beam.vendor.grpc.v1p13p1.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
>
> On 27 May 2019, at 6:53 PM, Maximilian Michels <[email protected]> wrote:
>
> Hi Mingliang,
>
> The environment is created for each TaskManager.
>
>    For docker, will it create one docker per flink taskmanager?
>
>
> Yes.
>
>    For process, does it mean start a python process to run the user code? And 
> it seems "command" should be set in the environment config, but what should 
> it be?
>
>
> You will have to start the same Python SDK Harness which would run inside a 
> Docker container if you had chosen Docker. This is a more manual approach 
> which should only be chosen if you cannot use Docker.
>
>    For external(loopback), does it mean let flink operator to call an 
> external service and by default set to the place where I submit the beam job? 
> This looks like all the data will be shift to a single machine and processed 
> there.
>
>
> This intended for a long-running SDK Harness which is already running when 
> you run your pipeline. Thus, you only provide the address to the already 
> running SDK Harness.
>
> Cheers,
> Max
>
> On 26.05.19 13:51, 青雉(祁明良) wrote:
>
> Hi All,
> I'm currently trying python portable runner with Flink. I see there are 3 
> kinds of environment_type available "docker/process/external(loopback)" when 
> submit a job. But I didn't find any material explain more.
> 1. For docker, will it create one docker per flink taskmanager?
> 2. For process, does it mean start a python process to run the user
>    code? And it seems "command" should be set in the environment
>    config, but what should it be?
> 3. For external(loopback), does it mean let flink operator to call an
>    external service and by default set to the place where I submit the
>    beam job? This looks like all the data will be shift to a single
>    machine and processed there.
> Thanks,
> Mingliang
> 本邮件及其附件含有小红书公司的保密信息,仅限于发送给以上收件人或群组。禁 止任何其他人以任何形式使用(包括但不限于全部或部分地泄露、复制、或散发) 
> 本邮件中的信息。如果您错收了本邮件,请您立即电话或邮件通知发件人并删除本 邮件!
> This communication may contain privileged or other confidential information 
> of Red. If you have received it in error, please advise the sender by reply 
> e-mail and immediately delete the message and any attachments without copying 
> or disclosing the contents. Thank you.
>
>
>
> 本邮件及其附件含有小红书公司的保密信息,仅限于发送给以上收件人或群组。禁止任何其他人以任何形式使用(包括但不限于全部或部分地泄露、复制、或散发)本邮件中的信息。如果您错收了本邮件,请您立即电话或邮件通知发件人并删除本邮件!
> This communication may contain privileged or other confidential information 
> of Red. If you have received it in error, please advise the sender by reply 
> e-mail and immediately delete the message and any attachments without copying 
> or disclosing the contents. Thank you.

Reply via email to