Good ideas!

That works great for manually looking at the job tracker UI... but is there
a way to figure out the job ID of the query programmatically in order to
track the progress? Or do I need to screen-scape the job tracker web page
somehow?

Thanks,
Ryan

On Thu, Jun 24, 2010 at 2:08 PM, Edward Capriolo <[email protected]>wrote:

> On Thu, Jun 24, 2010 at 1:49 PM, John Sichi <[email protected]> wrote:
> > JDBC doesn't have a standard interface for this, but one clumsy way might
> be to set a unique property on the session, and then scan the job tracker
> (visiting the job.xml to grab the properties) to find jobs with this
> property set.  Then you could use the standard Hadoop progress reporting (at
> least for single-job queries).
> >
> > I just verified that if I do
> >
> > set xyzzy=1;
> >
> > I can find the jobs manually from the job tracker web UI.
> >
> > JVS
> >
> > On Jun 24, 2010, at 7:29 AM, Ryan LeCompte wrote:
> >
> >> Hey guys,
> >>
> >> We are currently using the JDBC interface to Hive to remotely send Hive
> queries.
> >>
> >> The only problem here is that when the statement is executed, it just
> sits and hangs until the Hive query has completed. Is there any way to
> somehow submit a query and be able to get a handler on some object that I
> could ping to see how far along (percentage, perhaps) the Hive query has
> gone?
> >>
> >> Thank you,
> >> Ryan
> >>
> >
> >
>
> We do something like this
>
> select /* mynamehere */ ...
>
> This way we know exactly who to blame when the cluster gets saturated
> from one look at the JobTracker page.
>

Reply via email to