Github user AzureQ commented on the issue:

    https://github.com/apache/spark/pull/23037
  
    > I see this customization to be specific towards how you build your custom 
Docker image. Unless it is relevant towards testing, we are trying to keep the 
default Docker image as lightweight as possible (as long as it passes our test 
cases). Unless one of the committers sees this as an important thing to include 
in the default image, I believe it to be a customization.
    
    This is to make the default Docker image work properly, not customization 
at all. Without this change, the docker image of pyspark does not work properly 
in "client mode" while spark image and rspark image are working properly.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to