Yes, the Python bindings are still supported.

Can you dump the DebugString of the TaskInfo you're constructing, to
confirm the SlaveID looks ok?

Ben


On Tue, May 28, 2013 at 7:06 AM, David Greenberg <[email protected]>wrote:

> Sorry for the delayed response--I'm having some issues w/ email delivery to
> gmail...
>
> I'm trying to use the Python binding in this application. I am copying from
> offer.slave_id.value to task.slave_id.value using the = operator.
>
> Is the python binding still supported? Either way, due to some new
> concurrency requirements, I'm going to be shifting gears into writing a
> JVM-based Mesos framework now.
>
> Thanks!
>
>
> On Thu, May 23, 2013 at 1:02 PM, Vinod Kone <[email protected]> wrote:
>
> > ---------- Forwarded message ----------
> > From: Vinod Kone <[email protected]>
> > Date: Sun, May 19, 2013 at 6:56 PM
> > Subject: Re: Question about TASK_LOST statuses
> > To: "[email protected]" <[email protected]>
> >
> >
> > On the master's logs, I see this:
> >
> > > - 5600+ instances of "Error validating task XXX: Task uses invalid
> slave:
> > > SOME_UUID"
> > >
> > What do you think the problem is? I am copying the slave_id from the
> offer
> > > into the TaskInfo protobuf.
> > >
> > >
> > This will happen if the slave id in the task doesn't match the slave id
> in
> > the slave. Are you sure you are doing the copying the right slave ids to
> > the right tasks? Looks like there is a mismatch. Maybe some logs/printfs
> on
> > your scheduler, when you launch tasks, can point out the issue.
> >
> >
> >
> > > I'm using the process-based isolation at the moment (I haven't had the
> > time
> > > to set up the cgroups isolation yet).
> > >
> > > I can find and share whatever else is needed so that we can figure out
> > why
> > > these messages are occurring.
> > >
> > > Thanks,
> > > David
> > >
> > >
> > > On Fri, May 17, 2013 at 5:16 PM, Vinod Kone <[email protected]>
> wrote:
> > >
> > > > Hi David,
> > > >
> > > > You are right in that all these status updates are what we call
> > > "terminal"
> > > > status updates and mesos takes specific actions when it
> gets/generates
> > > one
> > > > of these.
> > > >
> > > > TASK_LOST is special in the sense that is not generated by the
> > executor,
> > > > but by the slave/master. You could think of it as an exception in
> > mesos.
> > > > Clearly, these should be rare in a stable mesos system.
> > > >
> > > > What do your logs say about the TASK_LOSTs? Is it always the same
> > issue?
> > > > Are you running w/ cgroups?
> > > >
> > > >
> > > >
> > > > On Fri, May 17, 2013 at 2:04 PM, David Greenberg <
> > [email protected]
> > > > >wrote:
> > > >
> > > > > Hello! Today I began working on a more advanced version of
> > mesos-submit
> > > > > that will handle hot-spares.
> > > > >
> > > > > I was assuming that TASK_{FAILED,FINISHED,LOST,KILLED} were the
> > status
> > > > > updates that meant that I needed to start a new spare process, as
> the
> > > > > monitored task was killed. However, I noticed that I often recieved
> > > > > TASK_LOSTs, and every 5 seconds, my scheduler would think its tasks
> > had
> > > > all
> > > > > died, so it'd restart too many. Nevertheless, the tasks would
> > reappear
> > > > > later on, and I could see them in the web interface of Mesos,
> > > continuing
> > > > to
> > > > > run.
> > > > >
> > > > > What is going on?
> > > > >
> > > > > Thanks!
> > > > > David
> > > > >
> > > >
> > >
> >
>

Reply via email to