Hi Anil,

It looks to me like you are not using Apache Beam -- e.g., I
see com.google.cloud.dataflow.sdk.util.UserCodeException which is not a
Beam class. Additionally, the documentation on cloud.google.com is not for
Apache Beam -- for that, see https://beam.apache.org -- e.g., the Beam Java
Quickstart <https://beam.apache.org/get-started/quickstart-java/>.

Can you please try with Apache Beam 0.6.0 release? And make sure that your
pipeline works in the DirectRunner before trying it on any of the other
runners, including DataflowRunner.

Thanks,
Dan

On Wed, Apr 12, 2017 at 5:07 AM, Anil Srinivas <anil.b...@impactanalytics.co
> wrote:

> Hi,
>  Yes I have activated the Dataflow api and it is getting executed in the
> local. The problem is when I try to execute on the Cloud Dataflow service.
>
>
>
> Thanks and Regards
> Anil
>
> On Apr 12, 2017 17:33, "Alexandre Crayssac" <alexandre.crays...@polynom.io>
> wrote:
>
>> Hi,
>>
>> Have you activated the Dataflow API in the GCP console ?
>>
>> Regards,
>>
>> Alexandre
>>
>> On Wed, Apr 12, 2017 at 11:52 AM, Anil Srinivas <
>> anil.b...@impactanalytics.co> wrote:
>>
>>> Hi,
>>> I have been trying to execute the example word-count on the cloud
>>> dataflow service as per the documentation in https://cloud.google.com/da
>>> taflow/docs/quickstarts/quickstart-java-maven. Even after following the
>>> exact same steps as mentioned in the document, I am getting the following
>>> error:
>>>
>>> Caused by: com.google.cloud.dataflow.sdk.util.UserCodeException:
>>> java.lang.RuntimeException: java.lang.RuntimeException:
>>> com.google.cloud.dataflow.sdk.util.UserCodeException:
>>> java.io.IOException: com.google.api.client.googleap
>>> is.json.GoogleJsonResponseException: 400 Bad Request
>>> {
>>>   "code" : 400,
>>>   "errors" : [ {
>>>     "domain" : "global",
>>>     "message" : "No object name",
>>>     "reason" : "required"
>>>   } ],
>>>   "message" : "No object name"
>>> }
>>> at com.google.cloud.dataflow.sdk.util.UserCodeException.wrap(Us
>>> erCodeException.java:35)
>>> at com.google.cloud.dataflow.sdk.util.UserCodeException.wrapIf(
>>> UserCodeException.java:40)
>>> at com.google.cloud.dataflow.sdk.util.DoFnRunnerBase.wrapUserCo
>>> deException(DoFnRunnerBase.java:369)
>>> at com.google.cloud.dataflow.sdk.util.SimpleDoFnRunner.invokePr
>>> ocessElement(SimpleDoFnRunner.java:51)
>>> at com.google.cloud.dataflow.sdk.util.DoFnRunnerBase.processEle
>>> ment(DoFnRunnerBase.java:139)
>>> at com.google.cloud.dataflow.sdk.runners.worker.SimpleParDoFn.p
>>> rocessElement(SimpleParDoFn.java:188)
>>> at com.google.cloud.dataflow.sdk.runners.worker.ForwardingParDo
>>> Fn.processElement(ForwardingParDoFn.java:42)
>>> at com.google.cloud.dataflow.sdk.runners.worker.DataflowWorkerL
>>> oggingParDoFn.processElement(DataflowWorkerLoggingParDoFn.java:47)
>>> at com.google.cloud.dataflow.sdk.util.common.worker.ParDoOperat
>>> ion.process(ParDoOperation.java:55)
>>> at com.google.cloud.dataflow.sdk.util.common.worker.OutputRecei
>>> ver.process(OutputReceiver.java:52)
>>> at com.google.cloud.dataflow.sdk.runners.worker.SimpleParDoFn$1
>>> .output(SimpleParDoFn.java:158)
>>> at com.google.cloud.dataflow.sdk.util.DoFnRunnerBase$DoFnContex
>>> t.outputWindowedValue(DoFnRunnerBase.java:288)
>>> at com.google.cloud.dataflow.sdk.util.DoFnRunnerBase$DoFnProces
>>> sContext.output(DoFnRunnerBase.java:450)
>>> at com.google.cloud.dataflow.sdk.runners.worker.CombineValuesFn
>>> Factory$ExtractOutputDoFn.processElement(CombineValuesFnFact
>>> ory.java:270)
>>> at com.google.cloud.dataflow.sdk.util.SimpleDoFnRunner.invokePr
>>> ocessElement(SimpleDoFnRunner.java:49)
>>> at com.google.cloud.dataflow.sdk.util.DoFnRunnerBase.processEle
>>> ment(DoFnRunnerBase.java:139)
>>> at com.google.cloud.dataflow.sdk.runners.worker.SimpleParDoFn.p
>>> rocessElement(SimpleParDoFn.java:188)
>>> at com.google.cloud.dataflow.sdk.runners.worker.ForwardingParDo
>>> Fn.processElement(ForwardingParDoFn.java:42)
>>> at com.google.cloud.dataflow.sdk.runners.worker.DataflowWorkerL
>>> oggingParDoFn.processElement(DataflowWorkerLoggingParDoFn.java:47)
>>> at com.google.cloud.dataflow.sdk.util.common.worker.ParDoOperat
>>> ion.process(ParDoOperation.java:55)
>>> at com.google.cloud.dataflow.sdk.util.common.worker.OutputRecei
>>> ver.process(OutputReceiver.java:52)
>>> at com.google.cloud.dataflow.sdk.runners.worker.SimpleParDoFn$1
>>> .output(SimpleParDoFn.java:158)
>>> at com.google.cloud.dataflow.sdk.util.DoFnRunnerBase$DoFnContex
>>> t.outputWindowedValue(DoFnRunnerBase.java:288)
>>> at com.google.cloud.dataflow.sdk.util.DoFnRunnerBase$DoFnContex
>>> t.outputWindowedValue(DoFnRunnerBase.java:284)
>>> at com.google.cloud.dataflow.sdk.util.DoFnRunnerBase$DoFnProces
>>> sContext$1.outputWindowedValue(DoFnRunnerBase.java:508)
>>> at com.google.cloud.dataflow.sdk.util.GroupAlsoByWindowsAndComb
>>> ineDoFn.closeWindow(GroupAlsoByWindowsAndCombineDoFn.java:203)
>>> at com.google.cloud.dataflow.sdk.util.GroupAlsoByWindowsAndComb
>>> ineDoFn.processElement(GroupAlsoByWindowsAndCombineDoFn.java:190)
>>> at com.google.cloud.dataflow.sdk.util.SimpleDoFnRunner.invokePr
>>> ocessElement(SimpleDoFnRunner.java:49)
>>> at com.google.cloud.dataflow.sdk.util.DoFnRunnerBase.processEle
>>> ment(DoFnRunnerBase.java:139)
>>> at com.google.cloud.dataflow.sdk.runners.worker.SimpleParDoFn.p
>>> rocessElement(SimpleParDoFn.java:188)
>>> at com.google.cloud.dataflow.sdk.runners.worker.ForwardingParDo
>>> Fn.processElement(ForwardingParDoFn.java:42)
>>> at com.google.cloud.dataflow.sdk.runners.worker.DataflowWorkerL
>>> oggingParDoFn.processElement(DataflowWorkerLoggingParDoFn.java:47)
>>> at com.google.cloud.dataflow.sdk.util.common.worker.ParDoOperat
>>> ion.process(ParDoOperation.java:55)
>>> at com.google.cloud.dataflow.sdk.util.common.worker.OutputRecei
>>> ver.process(OutputReceiver.java:52)
>>> at com.google.cloud.dataflow.sdk.util.common.worker.ReadOperati
>>> on.runReadLoop(ReadOperation.java:221)
>>> at com.google.cloud.dataflow.sdk.util.common.worker.ReadOperati
>>> on.start(ReadOperation.java:182)
>>> at com.google.cloud.dataflow.sdk.util.common.worker.MapTaskExec
>>> utor.execute(MapTaskExecutor.java:69)
>>> at com.google.cloud.dataflow.sdk.runners.worker.Datafl
>>> 2017-04-12T09:30:48.972Z: Detail:  (7323341af8d0122c): Cleaning up.
>>> 2017-04-12T09:30:48.978Z: Basic:  (7323341af8d010f8): Stopping worker
>>> pool...
>>> 2017-04-12T09:31:53.989Z: Basic:  (7323341af8d01811): Worker pool
>>> stopped.
>>> Apr 12, 2017 3:01:59 PM com.google.cloud.dataflow.sdk.
>>> runners.BlockingDataflowPipelineRunner run
>>> INFO: Job finished with status FAILED
>>> [WARNING]
>>> java.lang.reflect.InvocationTargetException
>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce
>>> ssorImpl.java:62)
>>> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe
>>> thodAccessorImpl.java:43)
>>> at java.lang.reflect.Method.invoke(Method.java:498)
>>> at org.codehaus.mojo.exec.ExecJavaMojo$1.run(ExecJavaMojo.java:293)
>>> at java.lang.Thread.run(Thread.java:745)
>>> Caused by: 
>>> com.google.cloud.dataflow.sdk.runners.DataflowJobExecutionException:
>>> Job 2017-04-12_02_26_24-7430037228537323212 failed with status FAILED
>>> at com.google.cloud.dataflow.sdk.runners.BlockingDataflowPipeli
>>> neRunner.run(BlockingDataflowPipelineRunner.java:158)
>>> at com.google.cloud.dataflow.sdk.runners.BlockingDataflowPipeli
>>> neRunner.run(BlockingDataflowPipelineRunner.java:56)
>>> at com.google.cloud.dataflow.sdk.Pipeline.run(Pipeline.java:181)
>>> at com.example.WordCount.main(WordCount.java:202)
>>> ... 6 more
>>> [INFO] ------------------------------------------------------------
>>> ------------
>>> [INFO] BUILD FAILURE
>>> [INFO] ------------------------------------------------------------
>>> ------------
>>> [INFO] Total time: 06:14 min
>>> [INFO] Finished at: 2017-04-12T15:01:59+05:30
>>> [INFO] Final Memory: 28M/367M
>>> [INFO] ------------------------------------------------------------
>>> ------------
>>> [ERROR] Failed to execute goal 
>>> org.codehaus.mojo:exec-maven-plugin:1.4.0:java
>>> (default-cli) on project first-dataflow: An exception occured while
>>> executing the Java class. null: InvocationTargetException: Job
>>> 2017-04-12_02_26_24-7430037228537323212 failed with status FAILED ->
>>> [Help 1]
>>> [ERROR]
>>> [ERROR] To see the full stack trace of the errors, re-run Maven with the
>>> -e switch.
>>> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
>>> [ERROR]
>>> [ERROR] For more information about the errors and possible solutions,
>>> please read the following articles:
>>> [ERROR] [Help 1] http://cwiki.apache.org/con
>>> fluence/display/MAVEN/MojoExecutionException
>>>
>>> Any help regarding this issue will be highly appreciated as there are
>>> not much documentations and discussions regarding the same.
>>>
>>>
>>> Thanks and Regards
>>> Anil
>>>
>>
>>
>>
>> --
>> Alexandre Crayssac
>> CTO and Co-Founder
>> +33 6 78 00 24 77 <+33%206%2078%2000%2024%2077>
>> alexandre.crays...@polynom.io
>>
>>

Reply via email to