Sure.  Which part is confusing?  The adjacencies?  Or why you would use it
at all?


On Tue, Jan 28, 2014 at 4:47 PM, Aaron Cody <[email protected]> wrote:

> thanks Billie - do you think you could go into a little more detail about
> the workflow DAG stuff on the wiki? it's a little cryptic (to me anyway)  :)
>
> From: Billie Rinaldi <[email protected]>
> Reply-To: <[email protected]>
> Date: Mon, 20 Jan 2014 07:40:04 -0800
> To: <[email protected]>
> Subject: Re: Jobs view .. how to hook into it....
>
> In Hadoop 1 only, there is a log4j appender on the JobTracker/JobHistory
> that inserts the data into postgres (or whichever db you have configured).
> The code is in contrib/ambari-log4j.
>
> Billie
>
>
> On Fri, Jan 17, 2014 at 1:59 PM, Aaron Cody <[email protected]> wrote:
>
>> hello
>> I'm looking at integrating my own process into the Ambari 'Jobs' view ...
>> and I can see how the web side of things works .. i.e. the view makes REST
>> calls to the server which in turn results in a query to postgres to get the
>> job stats ... but what is not so clear is how those job/task stats get into
>> postgres in the first place....
>> Q: for example, with MapReduce .. is Hadoop/JobTracker somehow inserting
>> the job/task info into postgres directly? Or is there some other mechanism
>> in Ambari that is listening for map reduce jobs/tasks to start/finish?
>>
>> any hints on where to look in the source tree would be greatly appreciated
>> TIA
>>
>
>

Reply via email to