You can also try without --streaming option. There's a separate streaming
wordcount example in the same directory.

If you want to look into the output files, it would be easier to use
external target like gs:// instead of local file.

python -m apache_beam.examples.wordcount --input=/etc/profile
--output=gs://tmp_location/py-wordcount --runner=PortableRunner
--job_endpoint=localhost:8099 --parallelism=1

On Tue, Jan 22, 2019 at 11:44 AM junwa...@gmail.com <junwa...@gmail.com>
wrote:

> Hello,
>
> I tried to follow the instructions at
> https://beam.apache.org/roadmap/portability/#python-on-flink,
>
> 1. I installed Flink local cluster, and followed their
> SocketWindowWordCount example and confirmed  the cluster works properly.
>
> 2. Start Flink job server:
> ./gradlew :beam-runners-flink_2.11-job-server:runShadow
> -PflinkMasterUrl=localhost:8081
>
> 3. Subject the job as suggested by an earlier thread:
> python -m apache_beam.examples.wordcount --input=/etc/profile
> --output=/tmp/py-wordcount-direct --runner=PortableRunner
> --job_endpoint=localhost:8099 --parallelism=1
> --OPTIONALflink_master=localhost:8081 --streaming
>
> But got the following NullPointerException error (sorry for the long text
> below), any ideas? Thanks
>
> Jun Wan
>
> ---- log starts ----
> [grpc-default-executor-2] INFO
> org.apache.beam.runners.flink.FlinkJobInvoker - Invoking job
> BeamApp-jwan-0121192804-387b3baa_1d32eea3-d71a-45a9-afa8-edbc66bc1d6b
> [grpc-default-executor-2] INFO
> org.apache.beam.runners.flink.FlinkJobInvocation - Starting job invocation
> BeamApp-jwan-0121192804-387b3baa_1d32eea3-d71a-45a9-afa8-edbc66bc1d6b
> [flink-runner-job-server] INFO
> org.apache.beam.runners.flink.FlinkJobInvocation - Translating pipeline to
> Flink program.
> [flink-runner-job-server] INFO
> org.apache.beam.runners.flink.FlinkExecutionEnvironments - Creating a
> Streaming Environment.
> [flink-runner-job-server] INFO
> org.apache.beam.runners.flink.FlinkExecutionEnvironments - Using Flink
> Master URL localhost:8081.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - class
> org.apache.beam.sdk.transforms.join.RawUnionValue does not contain a setter
> for field unionTag
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - Class class
> org.apache.beam.sdk.transforms.join.RawUnionValue cannot be used as a POJO
> type because not all fields are valid POJO fields, and must be processed as
> GenericType. Please read the Flink documentation on "Data Types &
> Serialization" for details of the effect on performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - class
> org.apache.beam.sdk.transforms.join.RawUnionValue does not contain a setter
> for field unionTag
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - Class class
> org.apache.beam.sdk.transforms.join.RawUnionValue cannot be used as a POJO
> type because not all fields are valid POJO fields, and must be processed as
> GenericType. Please read the Flink documentation on "Data Types &
> Serialization" for details of the effect on performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - class
> org.apache.beam.sdk.transforms.join.RawUnionValue does not contain a setter
> for field unionTag
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - Class class
> org.apache.beam.sdk.transforms.join.RawUnionValue cannot be used as a POJO
> type because not all fields are valid POJO fields, and must be processed as
> GenericType. Please read the Flink documentation on "Data Types &
> Serialization" for details of the effect on performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - class
> org.apache.beam.sdk.transforms.join.RawUnionValue does not contain a setter
> for field unionTag
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - Class class
> org.apache.beam.sdk.transforms.join.RawUnionValue cannot be used as a POJO
> type because not all fields are valid POJO fields, and must be processed as
> GenericType. Please read the Flink documentation on "Data Types &
> Serialization" for details of the effect on performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - class
> org.apache.beam.sdk.transforms.join.RawUnionValue does not contain a setter
> for field unionTag
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - Class class
> org.apache.beam.sdk.transforms.join.RawUnionValue cannot be used as a POJO
> type because not all fields are valid POJO fields, and must be processed as
> GenericType. Please read the Flink documentation on "Data Types &
> Serialization" for details of the effect on performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - No fields were detected
> for class org.apache.beam.sdk.util.WindowedValue so it cannot be used as a
> POJO type and must be processed as GenericType. Please read the Flink
> documentation on "Data Types & Serialization" for details of the effect on
> performance.
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - class
> org.apache.beam.sdk.transforms.join.RawUnionValue does not contain a setter
> for field unionTag
> [flink-runner-job-server] INFO
> org.apache.flink.api.java.typeutils.TypeExtractor - Class class
> org.apache.beam.sdk.transforms.join.RawUnionValue cannot be used as a POJO
> type because not all fields are valid POJO fields, and must be processed as
> GenericType. Please read the Flink documentation on "Data Types &
> Serialization" for details of the effect on performance.
> [flink-runner-job-server] INFO
> org.apache.beam.runners.flink.FlinkExecutionEnvironments - Running remotely
> at localhost:8081
> [flink-runner-job-server] WARN
> org.apache.flink.configuration.Configuration - Config uses deprecated
> configuration key 'jobmanager.rpc.address' instead of proper key
> 'rest.address'
> [flink-runner-job-server] INFO org.apache.flink.runtime.rest.RestClient -
> Rest client endpoint started.
> [flink-runner-job-server] INFO
> org.apache.flink.client.program.rest.RestClusterClient - Submitting job
> 9a91889c469db1d88ec8f6a6d04a67b7 (detached: false).
> [flink-runner-job-server] INFO org.apache.flink.runtime.rest.RestClient -
> Shutting down rest endpoint.
> [flink-runner-job-server] INFO org.apache.flink.runtime.rest.RestClient -
> Rest endpoint shutdown complete.
> [flink-runner-job-server] ERROR
> org.apache.beam.runners.flink.FlinkJobInvocation - Error during job
> invocation
> BeamApp-jwan-0121192804-387b3baa_1d32eea3-d71a-45a9-afa8-edbc66bc1d6b.
> org.apache.flink.client.program.ProgramInvocationException: Could not
> retrieve the execution result.
>         at
> org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:258)
>         at
> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:464)
>         at
> org.apache.beam.runners.flink.FlinkExecutionEnvironments$BeamFlinkRemoteStreamEnvironment.executeRemotely(FlinkExecutionEnvironments.java:355)
>         at
> org.apache.flink.streaming.api.environment.RemoteStreamEnvironment.execute(RemoteStreamEnvironment.java:179)
>         at
> org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator$StreamingTranslationContext.execute(FlinkStreamingPortablePipelineTranslator.java:158)
>         at
> org.apache.beam.runners.flink.FlinkJobInvocation.runPipelineWithTranslator(FlinkJobInvocation.java:142)
>         at
> org.apache.beam.runners.flink.FlinkJobInvocation.runPipeline(FlinkJobInvocation.java:112)
>         at org.apache.beam.repackaged.beam_runners_flink_2.11.com
> .google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:111)
>         at org.apache.beam.repackaged.beam_runners_flink_2.11.com
> .google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:58)
>         at org.apache.beam.repackaged.beam_runners_flink_2.11.com
> .google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:75)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>         at java.lang.Thread.run(Thread.java:748)
> Caused by: org.apache.flink.runtime.client.JobSubmissionException: Failed
> to submit JobGraph.
>         at
> org.apache.flink.client.program.rest.RestClusterClient.lambda$submitJob$8(RestClusterClient.java:371)
>         at
> java.util.concurrent.CompletableFuture.uniExceptionally(CompletableFuture.java:870)
>         at
> java.util.concurrent.CompletableFuture$UniExceptionally.tryFire(CompletableFuture.java:852)
>         at
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>         at
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
>         at
> org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$5(FutureUtils.java:203)
>         at
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>         at
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>         at
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>         at
> java.util.concurrent.CompletableFuture.postFire(CompletableFuture.java:561)
>         at
> java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:929)
>         at
> java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:442)
>         ... 3 more
> Caused by: org.apache.flink.runtime.rest.util.RestClientException:
> [Internal server error., <Exception on server side:
> java.util.concurrent.CompletionException: java.lang.NullPointerException
>         at
> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)
>         at
> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280)
>         at
> java.util.concurrent.CompletableFuture.biApply(CompletableFuture.java:1107)
>         at
> java.util.concurrent.CompletableFuture$BiApply.tryFire(CompletableFuture.java:1070)
>         at
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>         at
> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1595)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>         at java.lang.Thread.run(Thread.java:748)
> Caused by: java.lang.NullPointerException
>         at
> org.apache.flink.runtime.jobgraph.JobGraph.writeUserArtifactEntriesToConfiguration(JobGraph.java:586)
>         at
> org.apache.flink.runtime.client.ClientUtils.setUserArtifactBlobKeys(ClientUtils.java:140)
>         at
> org.apache.flink.runtime.client.ClientUtils.uploadAndSetUserArtifacts(ClientUtils.java:121)
>         at
> org.apache.flink.runtime.client.ClientUtils.uploadJobGraphFiles(ClientUtils.java:78)
>         at
> org.apache.flink.runtime.rest.handler.job.JobSubmitHandler.lambda$uploadJobGraphFiles$4(JobSubmitHandler.java:168)
>         at
> java.util.concurrent.CompletableFuture.biApply(CompletableFuture.java:1105)
>         ... 6 more
>
> End of exception on server side>]
>         at
> org.apache.flink.runtime.rest.RestClient.parseResponse(RestClient.java:349)
>         at
> org.apache.flink.runtime.rest.RestClient.lambda$submitRequest$3(RestClient.java:333)
>         at
> java.util.concurrent.CompletableFuture.uniCompose(CompletableFuture.java:952)
>         at
> java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:926)
>         ... 4 more
>
> ---- log ends ----
>
> On 2018/11/15 11:01:28, Maximilian Michels <m...@apache.org> wrote:
> > Hi Ruoyun,
> >
> > The output file will be within the container which is deleted after
> > shutdown by default. You can keep the containers if you add the flag
> >
> >    --retain_docker_containers
> >
> > Note, this is from ManualDockerEnvironmentOptions.
> >
> > The problem with batch is that it executes staged and will create
> > multiple containers [1] which don't share the same local file system. So
> > the wordcount only works reliably if you use a distributed file system.
> >
> > Cheers,
> > Max
> >
> > [1] You can prevent multiple containers by using
> >      --environment_cache_millis=10000
> >
> > On 14.11.18 20:44, Ruoyun Huang wrote:
> > > Thanks Thomas!
> > >
> > > My desktop runs Linux.  I was using gradle to run wordcount, and that
> > > was how I got the job hanging. Since both of you get it working, I
> guess
> > > more likely sth is wrong with my setup.
> > >
> > >
> > > By using Thmoas's python command line exactly as is, I am able to see
> > > the job run succeeds, however two questions:
> > >
> > > 1)  Did you check whether output file "/tmp/py-wordcount-direct"
> exists
> > > or not?  I expect there should be a text output, but I don't see this
> > > file afterwards.   (I am still in the stage building confidence in
> > > telling what a succeeded run is.  Maybe I will try DataflowRunner and
> > > cross check outputs).
> > >
> > > 2)  Why it needs a "--streaming" arg?  Isn't this a static batch
> input,
> > > by feeding a txt file input?  In fact, I got failure message if I
> remove
> > > '--streaming', not sure if it is due to my setup again.
> > >
> > >
> > > On Wed, Nov 14, 2018 at 7:51 AM Thomas Weise <t...@apache.org
> > > <mailto:t...@apache.org>> wrote:
> > >
> > >     Works for me on macOS as well.
> > >
> > >     In case you don't launch the pipeline through Gradle, this would be
> > >     the command:
> > >
> > >     python -m apache_beam.examples.wordcount \
> > >        --input=/etc/profile \
> > >        --output=/tmp/py-wordcount-direct \
> > >        --runner=PortableRunner \
> > >        --job_endpoint=localhost:8099 \
> > >        --parallelism=1 \
> > >        --OPTIONALflink_master=localhost:8081 \
> > >        --streaming
> > >
> > >     We talked about adding the wordcount to pre-commit..
> > >
> > >     Regarding using ULR vs. Flink runner: There seems to be confusion
> > >     between PortableRunner using the user supplied endpoint vs. trying
> > >     to launch a job server. I commented in the doc.
> > >
> > >     Thomas
> > >
> > >
> > >
> > >     On Wed, Nov 14, 2018 at 3:30 AM Maximilian Michels <m...@apache.org
> > >     <mailto:m...@apache.org>> wrote:
> > >
> > >         Hi Ruoyun,
> > >
> > >         I just ran the wordcount locally using the instructions on the
> > >         page.
> > >         I've tried the local file system and GCS. Both times it ran
> > >         successfully
> > >         and produced valid output.
> > >
> > >         I'm assuming there is some problem with your setup. Which
> > >         platform are
> > >         you using? I'm on MacOS.
> > >
> > >         Could you expand on the planned merge? From my understanding we
> > >         will
> > >         always need PortableRunner in Python to be able to submit
> > >         against the
> > >         Beam JobServer.
> > >
> > >         Thanks,
> > >         Max
> > >
> > >         On 14.11.18 00:39, Ruoyun Huang wrote:
> > >          > A quick follow-up on using current PortableRunner.
> > >          >
> > >          > I followed the exact three steps as Ankur and Maximilian
> > >         shared in
> > >          >
> https://beam.apache.org/roadmap/portability/#python-on-flink
> > >         ;   The
> > >          > wordcount example keeps hanging after 10 minutes.  I also
> tried
> > >          > specifying explicit input/output args, either using gcs
> > >         folder or local
> > >          > file system, but none of them works.
> > >          >
> > >          > Spent some time looking into it but conclusion yet.  At this
> > >         point
> > >          > though, I guess it does not matter much any more, given we
> > >         already have
> > >          > the plan of merging PortableRunner into using java reference
> > >         runner
> > >          > (i.e. :beam-runners-reference-job-server).
> > >          >
> > >          > Still appreciated if someone can try out the python-on-flink
> > >          >
> > >         <https://beam.apache.org/roadmap/portability/#python-on-flink
> >instructions
> > >
> > >          > in case it is just due to my local machine setup.  Thanks!
> > >          >
> > >          >
> > >          >
> > >          > On Thu, Nov 8, 2018 at 5:04 PM Ruoyun Huang
> > >         <ruo...@google.com <mailto:ruo...@google.com>
> > >          > <mailto:ruo...@google.com <mailto:ruo...@google.com>>>
> wrote:
> > >          >
> > >          >     Thanks Maximilian!
> > >          >
> > >          >     I am working on migrating existing PortableRunner to
> > >         using java ULR
> > >          >     (Link to Notes
> > >          >
> > >           <
> https://docs.google.com/document/d/1S86saZqiDaE_M5wxO0zOQ_rwC6QHv7sp1BmGTm0dLNE/edit#
> >).
> > >          >     If this issue is non-trivial to solve, I would vote for
> > >         removing
> > >          >     this default behavior as part of the consolidation.
> > >          >
> > >          >     On Thu, Nov 8, 2018 at 2:58 AM Maximilian Michels
> > >         <m...@apache.org <mailto:m...@apache.org>
> > >          >     <mailto:m...@apache.org <mailto:m...@apache.org>>> wrote:
> > >          >
> > >          >         In the long run, we should get rid of the
> > >         Docker-inside-Docker
> > >          >         approach,
> > >          >         which was only intended for testing anyways. It
> would
> > >         be cleaner to
> > >          >         start the SDK harness container alongside with
> > >         JobServer container.
> > >          >
> > >          >         Short term, I think it should be easy to either fix
> the
> > >          >         permissions of
> > >          >         the mounted "docker" executable or use a Docker
> image
> > >         for the
> > >          >         JobServer
> > >          >         which comes with Docker pre-installed.
> > >          >
> > >          >         JIRA:
> https://issues.apache.org/jira/browse/BEAM-6020
> > >          >
> > >          >         Thanks for reporting this Ruoyun!
> > >          >
> > >          >         -Max
> > >          >
> > >          >         On 08.11.18 00:10, Ruoyun Huang wrote:
> > >          >          > Thanks Ankur and Maximilian.
> > >          >          >
> > >          >          > Just for reference in case other people
> > >         encountering the same
> > >          >         error
> > >          >          > message, the "permission denied" error in my
> > >         original email
> > >          >         is exactly
> > >          >          > due to dockerinsidedocker issue that Ankur
> mentioned.
> > >          >         Thanks Ankur!
> > >          >          > Didn't make the link when you said it, had to
> > >         discover that
> > >          >         in a hard
> > >          >          > way (I thought it is due to my docker
> installation
> > >         messed up).
> > >          >          >
> > >          >          > On Tue, Nov 6, 2018 at 1:53 AM Maximilian Michels
> > >          >         <m...@apache.org <mailto:m...@apache.org>
> > >         <mailto:m...@apache.org <mailto:m...@apache.org>>
> > >          >          > <mailto:m...@apache.org <mailto:m...@apache.org>
> > >         <mailto:m...@apache.org <mailto:m...@apache.org>>>> wrote:
> > >          >          >
> > >          >          >     Hi,
> > >          >          >
> > >          >          >     Please follow
> > >          >          >
> > >         https://beam.apache.org/roadmap/portability/#python-on-flink
> > >          >          >
> > >          >          >     Cheers,
> > >          >          >     Max
> > >          >          >
> > >          >          >     On 06.11.18 01:14, Ankur Goenka wrote:
> > >          >          >      > Hi,
> > >          >          >      >
> > >          >          >      > The Portable Runner requires a job server
> > >         uri to work
> > >          >         with. The
> > >          >          >     current
> > >          >          >      > default job server docker image is broken
> > >         because of
> > >          >         docker inside
> > >          >          >      > docker issue.
> > >          >          >      >
> > >          >          >      > Please refer to
> > >          >          >      >
> > >          >
> https://beam.apache.org/roadmap/portability/#python-on-flink for
> > >          >          >     how to
> > >          >          >      > run a wordcount using Portable Flink
> Runner.
> > >          >          >      >
> > >          >          >      > Thanks,
> > >          >          >      > Ankur
> > >          >          >      >
> > >          >          >      > On Mon, Nov 5, 2018 at 3:41 PM Ruoyun
> Huang
> > >          >         <ruo...@google.com <mailto:ruo...@google.com>
> > >         <mailto:ruo...@google.com <mailto:ruo...@google.com>>
> > >          >          >     <mailto:ruo...@google.com
> > >         <mailto:ruo...@google.com> <mailto:ruo...@google.com
> > >         <mailto:ruo...@google.com>>>
> > >          >          >      > <mailto:ruo...@google.com
> > >         <mailto:ruo...@google.com> <mailto:ruo...@google.com
> > >         <mailto:ruo...@google.com>>
> > >          >         <mailto:ruo...@google.com <mailto:ruo...@google.com
> >
> > >         <mailto:ruo...@google.com <mailto:ruo...@google.com>>>>>
> wrote:
> > >          >          >      >
> > >          >          >      >     Hi, Folks,
> > >          >          >      >
> > >          >          >      >           I want to try out Python
> > >         PortableRunner, by
> > >          >         using following
> > >          >          >      >     command:
> > >          >          >      >
> > >          >          >      >     *sdk/python: python -m
> > >         apache_beam.examples.wordcount
> > >          >          >      >       --output=/tmp/test_output   --runner
> > >         PortableRunner*
> > >          >          >      >
> > >          >          >      >           It complains with following
> error
> > >         message:
> > >          >          >      >
> > >          >          >      >     Caused by: java.lang.Exception: The
> > >         user defined
> > >          >         'open()' method
> > >          >          >      >     caused an exception:
> > >         java.io.IOException: Cannot
> > >          >         run program
> > >          >          >      >     "docker": error=13, Permission denied
> > >          >          >      >     at
> > >          >          >
> > >          >
> > >
>  org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:498)
> > >          >          >      >     at
> > >          >          >      >
> > >          >          >
> > >          >
> > >
>  org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:368)
> > >          >          >      >     at
> > >          >
> > >           org.apache.flink.runtime.taskmanager.Task.run(Task.java:712)
> > >          >          >      >     ... 1 more
> > >          >          >      >     Caused by:
> > >          >          >      >
> > >          >          >
> > >          >
> > >
>  
> org.apache.beam.repackaged.beam_runners_java_fn_execution.com.google.common.util.concurrent.UncheckedExecutionException:
> > >          >          >      >     java.io.IOException: Cannot run
> program
> > >         "docker":
> > >          >         error=13,
> > >          >          >      >     Permission denied
> > >          >          >      >     at
> > >          >          >      >
> > >          >          >
> > >          >
> > >
>  
> org.apache.beam.repackaged.beam_runners_java_fn_execution.com.google.common.cache.LocalCache$LocalLoadingCache.getUnchecked(LocalCache.java:4994)
> > >          >          >      >
> > >          >          >      >     ... 7 more
> > >          >          >      >
> > >          >          >      >
> > >          >          >      >
> > >          >          >      >     My py2 environment is properly
> > >         configured, because
> > >          >         DirectRunner
> > >          >          >      >     works.  Also I tested my docker
> > >         installation by
> > >          >         'docker run
> > >          >          >      >     hello-world ', no issue.
> > >          >          >      >
> > >          >          >      >
> > >          >          >      >     Thanks.
> > >          >          >      >     --
> > >          >          >      >     ================
> > >          >          >      >     Ruoyun  Huang
> > >          >          >      >
> > >          >          >
> > >          >          >
> > >          >          >
> > >          >          > --
> > >          >          > ================
> > >          >          > Ruoyun  Huang
> > >          >          >
> > >          >
> > >          >
> > >          >
> > >          >     --
> > >          >     ================
> > >          >     Ruoyun  Huang
> > >          >
> > >          >
> > >          >
> > >          > --
> > >          > ================
> > >          > Ruoyun  Huang
> > >          >
> > >
> > >
> > >
> > > --
> > > ================
> > > Ruoyun  Huang
> > >
> >
>

Reply via email to