I believe anything you can do from the web page you can do from the Hadoop Java 
API.

JVS

On Jun 24, 2010, at 11:16 AM, Ryan LeCompte wrote:

Good ideas!

That works great for manually looking at the job tracker UI... but is there a 
way to figure out the job ID of the query programmatically in order to track 
the progress? Or do I need to screen-scape the job tracker web page somehow?

Thanks,
Ryan

On Thu, Jun 24, 2010 at 2:08 PM, Edward Capriolo 
<[email protected]<mailto:[email protected]>> wrote:
On Thu, Jun 24, 2010 at 1:49 PM, John Sichi 
<[email protected]<mailto:[email protected]>> wrote:
> JDBC doesn't have a standard interface for this, but one clumsy way might be 
> to set a unique property on the session, and then scan the job tracker 
> (visiting the job.xml to grab the properties) to find jobs with this property 
> set.  Then you could use the standard Hadoop progress reporting (at least for 
> single-job queries).
>
> I just verified that if I do
>
> set xyzzy=1;
>
> I can find the jobs manually from the job tracker web UI.
>
> JVS
>
> On Jun 24, 2010, at 7:29 AM, Ryan LeCompte wrote:
>
>> Hey guys,
>>
>> We are currently using the JDBC interface to Hive to remotely send Hive 
>> queries.
>>
>> The only problem here is that when the statement is executed, it just sits 
>> and hangs until the Hive query has completed. Is there any way to somehow 
>> submit a query and be able to get a handler on some object that I could ping 
>> to see how far along (percentage, perhaps) the Hive query has gone?
>>
>> Thank you,
>> Ryan
>>
>
>

We do something like this

select /* mynamehere */ ...

This way we know exactly who to blame when the cluster gets saturated
from one look at the JobTracker page.


Reply via email to