On Thu, Jun 24, 2010 at 1:49 PM, John Sichi <[email protected]> wrote:
> JDBC doesn't have a standard interface for this, but one clumsy way might be 
> to set a unique property on the session, and then scan the job tracker 
> (visiting the job.xml to grab the properties) to find jobs with this property 
> set.  Then you could use the standard Hadoop progress reporting (at least for 
> single-job queries).
>
> I just verified that if I do
>
> set xyzzy=1;
>
> I can find the jobs manually from the job tracker web UI.
>
> JVS
>
> On Jun 24, 2010, at 7:29 AM, Ryan LeCompte wrote:
>
>> Hey guys,
>>
>> We are currently using the JDBC interface to Hive to remotely send Hive 
>> queries.
>>
>> The only problem here is that when the statement is executed, it just sits 
>> and hangs until the Hive query has completed. Is there any way to somehow 
>> submit a query and be able to get a handler on some object that I could ping 
>> to see how far along (percentage, perhaps) the Hive query has gone?
>>
>> Thank you,
>> Ryan
>>
>
>

We do something like this

select /* mynamehere */ ...

This way we know exactly who to blame when the cluster gets saturated
from one look at the JobTracker page.

Reply via email to