Most likely you cluster is running out of slots to run jobs.

Thanks.


On Mon, Aug 19, 2013 at 2:27 AM, Đức Trung Nguyễn <[email protected]>wrote:

> Hello every body,
>
> I am a student. I tried to run a workflow, which contains two parallel jobs
> (using fork) in a single node (installed Pseudo mode of hadoop) but these
> jobs are always in pending state (in hadoop jobtracker).
>
> Any idea for me ? Thank you very much.
>
> My workflow is bellow:
> <workflow-app name="query1" xmlns="uri:oozie:workflow:0.1"
> >
>     <start to="normalize_data"/>
>     <action name="normalize_data">
>         <pig>
>             <job-tracker>${jobTracker}</job-tracker>
>             <name-node>${nameNode}</name-node>
>             <prepare>
>                 <delete
> path="${nameNode}/user/${wf:user()}/${jobOutput}/data_normalized"/>
>             </prepare>
>             <configuration>
>                 <property>
>                     <name>mapred.job.queue.name</name>
>                     <value>${queueName}</value>
>                 </property>
>                 <property>
>                     <name>mapred.compress.map.output</name>
>                     <value>false</value>
>                 </property>
>             </configuration>
>             <script>normalize_data.pig</script>
>             <param>input=/user/${wf:user()}/${flight_input}</param>
>
> <param>output=/user/${wf:user()}/${jobOutput}/data_normalized</param>
>         </pig>
>         <ok to="forkSelectInboundOutbound" />
>         <error to="fail" />
>     </action>
>     <fork name="forkSelectInboundOutbound">
>         <path start="top_25_inbound"/>
>         <path start="top_25_outbound"/>
>     </fork>
>     <action name="top_25_inbound">
>         <pig>
>             <job-tracker>${jobTracker}</job-tracker>
>             <name-node>${nameNode}</name-node>
>             <prepare>
>                 <delete
> path="${nameNode}/user/${wf:user()}/${jobOutput}/top_25_inbound"/>
>             </prepare>
>             <configuration>
>                 <property>
>                     <name>mapred.job.queue.name</name>
>                     <value>${queueName}</value>
>                 </property>
>                 <property>
>                     <name>mapred.compress.map.output</name>
>                     <value>false</value>
>                 </property>
>             </configuration>
>             <script>select_top_25_inbound.pig</script>
>
> <param>input=/user/${wf:user()}/${jobOutput}/data_normalized/part*</param>
>
> <param>output=/user/${wf:user()}/${jobOutput}/top_25_inbound</param>
>         </pig>
>         <ok to="joinInboundOutbound" />
>         <error to="fail" />
>     </action>
>     <action name="top_25_outbound">
>         <pig>
>             <job-tracker>${jobTracker}</job-tracker>
>             <name-node>${nameNode}</name-node>
>             <prepare>
>                 <delete
> path="${nameNode}/user/${wf:user()}/${jobOutput}/top_25_outbound"/>
>                 <!--mkdir
> path="${nameNode}/user/${wf:user()}/${jobOutput}/top_25_outbound"/-->
>             </prepare>
>             <!--job-xml>job_xml_file</job-xml-->
>             <configuration>
>                 <property>
>                     <name>mapred.job.queue.name</name>
>                     <value>${queueName}</value>
>                 </property>
>                 <property>
>                     <name>mapred.compress.map.output</name>
>                     <value>false</value>
>                 </property>
>             </configuration>
>             <script>select_top_25_outbound.pig</script>
>
> <param>input=/user/${wf:user()}/${jobOutput}/data_normalized/part*</param>
>
> <param>output=/user/${wf:user()}/${jobOutput}/top_25_outbound</param>
>         </pig>
>         <ok to="joinInboundOutbound" />
>         <error to="fail" />
>     </action>
>
>     <join name="joinInboundOutbound" to="end" />
>
>     <action name="top_25_total">
>         <pig>
>             <job-tracker>${jobTracker}</job-tracker>
>             <name-node>${nameNode}</name-node>
>             <prepare>
>                 <delete
> path="${nameNode}/user/${wf:user()}/${jobOutput}/top_25_total"/>
>             </prepare>
>             <!--job-xml>job_xml_file</job-xml-->
>             <configuration>
>                 <property>
>                     <name>mapred.job.queue.name</name>
>                     <value>${queueName}</value>
>                 </property>
>                 <property>
>                     <name>mapred.compress.map.output</name>
>                     <value>false</value>
>                 </property>
>             </configuration>
>             <script>select_top_25_total.pig</script>
>
>
> <param>count_inbound=/user/${wf:user()}/${jobOutput}/top_25_inbound/part*</param>
>
>
> <param>count_outbound=/user/${wf:user()}/${jobOutput}/top_25_outbound/part*</param>
>
> <param>output=/user/${wf:user()}/${jobOutput}/top_25_total</param>
>         </pig>
>         <ok to="get_detail_result" />
>         <error to="fail" />
>     </action>
>     <action name="get_detail_result">
>         <pig>
>             <job-tracker>${jobTracker}</job-tracker>
>             <name-node>${nameNode}</name-node>
>             <prepare>
>                 <delete path="${nameNode}/user/${wf:user()}/${jobOutput}"/>
>             </prepare>
>             <configuration>
>                 <property>
>                     <name>mapred.job.queue.name</name>
>                     <value>${queueName}</value>
>                 </property>
>                 <property>
>                     <name>mapred.compress.map.output</name>
>                     <value>false</value>
>                 </property>
>             </configuration>
>             <script>analyis_different_time_granuraties.pig</script>
>
> <param>input=/user/${wf:user()}/${jobOutput}/top_25_total/part*</param>
>
> <param>output=/user/${wf:user()}/${jobOutput}/detail_result</param>
>         </pig>
>         <ok to="end" />
>         <error to="fail" />
>     </action>
>     <kill name="fail">
>         <message>
>             Pig failed, error
> message[${wf:errorMessage(wf:lastErrorNode())}]
>         </message>
>     </kill>
>     <end name="end"/>
> </workflow-app>
>



-- 
Alejandro

Reply via email to