Re: Spark 2.3 driver pod stuck in Running state — Kubernetes

2018-06-08 Thread Yinan Li
Yes, it looks like it is because there's not enough resources to run the
executor pods. Have you seen pending executor pods?

On Fri, Jun 8, 2018, 11:49 AM Thodoris Zois  wrote:

> As far as I know from Mesos with Spark, it is a running state and not a
> pending one. What you see is normal, but if I am wrong somebody correct me.
>
>  Spark driver at start operates normally (running state) but when it comes
> to start up executors, then it cannot allocate resources for them and
> hangs..
>
> - Thodoris
>
> On 8 Jun 2018, at 21:24, purna pradeep  wrote:
>
> Hello,
>
> When I run spark-submit on k8s cluster I’m
>
> Seeing driver pod stuck in Running state and when I pulled driver pod logs
> I’m able to see below log
>
> I do understand that this warning might be because of lack of cpu/ Memory
> , but I expect driver pod be in “Pending” state rather than “ Running”
> state though actually it’s not Running
>
> So I had kill the driver pod and resubmit the job
>
> Please suggest here !
>
> 2018-06-08 14:38:01 WARN TaskSchedulerImpl:66 - Initial job has not
> accepted any resources; check your cluster UI to ensure that workers are
> registered and have sufficient resources
>
> 2018-06-08 14:38:16 WARN TaskSchedulerImpl:66 - Initial job has not
> accepted any resources; check your cluster UI to ensure that workers are
> registered and have sufficient resources
>
> 2018-06-08 14:38:31 WARN TaskSchedulerImpl:66 - Initial job has not
> accepted any resources; check your cluster UI to ensure that workers are
> registered and have sufficient resources
>
> 2018-06-08 14:38:46 WARN TaskSchedulerImpl:66 - Initial job has not
> accepted any resources; check your cluster UI to ensure that workers are
> registered and have sufficient resources
>
> 2018-06-08 14:39:01 WARN TaskSchedulerImpl:66 - Initial job has not
> accepted any resources; check your cluster UI to ensure that workers are
> registered and have sufficient resources
>
>


Re: Spark 2.3 driver pod stuck in Running state — Kubernetes

2018-06-08 Thread Thodoris Zois
As far as I know from Mesos with Spark, it is a running state and not a pending 
one. What you see is normal, but if I am wrong somebody correct me.

 Spark driver at start operates normally (running state) but when it comes to 
start up executors, then it cannot allocate resources for them and hangs.. 

- Thodoris

> On 8 Jun 2018, at 21:24, purna pradeep  wrote:
> 
> Hello,
> When I run spark-submit on k8s cluster I’m
> 
> Seeing driver pod stuck in Running state and when I pulled driver pod logs 
> I’m able to see below log
> 
> I do understand that this warning might be because of lack of cpu/ Memory , 
> but I expect driver pod be in “Pending” state rather than “ Running” state 
> though actually it’s not Running 
> 
> So I had kill the driver pod and resubmit the job 
> 
> Please suggest here !
> 
> 2018-06-08 14:38:01 WARN TaskSchedulerImpl:66 - Initial job has not accepted 
> any resources; check your cluster UI to ensure that workers are registered 
> and have sufficient resources
> 
> 2018-06-08 14:38:16 WARN TaskSchedulerImpl:66 - Initial job has not accepted 
> any resources; check your cluster UI to ensure that workers are registered 
> and have sufficient resources
> 
> 2018-06-08 14:38:31 WARN TaskSchedulerImpl:66 - Initial job has not accepted 
> any resources; check your cluster UI to ensure that workers are registered 
> and have sufficient resources
> 
> 2018-06-08 14:38:46 WARN TaskSchedulerImpl:66 - Initial job has not accepted 
> any resources; check your cluster UI to ensure that workers are registered 
> and have sufficient resources
> 
> 2018-06-08 14:39:01 WARN TaskSchedulerImpl:66 - Initial job has not accepted 
> any resources; check your cluster UI to ensure that workers are registered 
> and have sufficient resources


Spark 2.3 driver pod stuck in Running state — Kubernetes

2018-06-08 Thread purna pradeep
Hello,

When I run spark-submit on k8s cluster I’m

Seeing driver pod stuck in Running state and when I pulled driver pod logs
I’m able to see below log

I do understand that this warning might be because of lack of cpu/ Memory ,
but I expect driver pod be in “Pending” state rather than “ Running” state
though actually it’s not Running

So I had kill the driver pod and resubmit the job

Please suggest here !

2018-06-08 14:38:01 WARN TaskSchedulerImpl:66 - Initial job has not
accepted any resources; check your cluster UI to ensure that workers are
registered and have sufficient resources

2018-06-08 14:38:16 WARN TaskSchedulerImpl:66 - Initial job has not
accepted any resources; check your cluster UI to ensure that workers are
registered and have sufficient resources

2018-06-08 14:38:31 WARN TaskSchedulerImpl:66 - Initial job has not
accepted any resources; check your cluster UI to ensure that workers are
registered and have sufficient resources

2018-06-08 14:38:46 WARN TaskSchedulerImpl:66 - Initial job has not
accepted any resources; check your cluster UI to ensure that workers are
registered and have sufficient resources

2018-06-08 14:39:01 WARN TaskSchedulerImpl:66 - Initial job has not
accepted any resources; check your cluster UI to ensure that workers are
registered and have sufficient resources