On Thu, Jun 24, 2010 at 1:49 PM, John Sichi <[email protected]> wrote: > JDBC doesn't have a standard interface for this, but one clumsy way might be > to set a unique property on the session, and then scan the job tracker > (visiting the job.xml to grab the properties) to find jobs with this property > set. Then you could use the standard Hadoop progress reporting (at least for > single-job queries). > > I just verified that if I do > > set xyzzy=1; > > I can find the jobs manually from the job tracker web UI. > > JVS > > On Jun 24, 2010, at 7:29 AM, Ryan LeCompte wrote: > >> Hey guys, >> >> We are currently using the JDBC interface to Hive to remotely send Hive >> queries. >> >> The only problem here is that when the statement is executed, it just sits >> and hangs until the Hive query has completed. Is there any way to somehow >> submit a query and be able to get a handler on some object that I could ping >> to see how far along (percentage, perhaps) the Hive query has gone? >> >> Thank you, >> Ryan >> > >
We do something like this select /* mynamehere */ ... This way we know exactly who to blame when the cluster gets saturated from one look at the JobTracker page.
