Hi Ankit,
Sure I suppose that elaboration on OS base can be added
BASE_OS="buster"
SPARK_VERSION="3.1.1"
SCALA_VERSION="scala_2.12"
DOCKERFILE="Dockerfile"
DOCKERIMAGETAG="11-jre-slim"
# Building Docker image from provided Dockerfile base 11
cd $SPARK_HOME
/opt/spark/bin/docker-image-tool.sh \
-r spark -t
${SPARK_VERSION}-${SCALA_VERSION}-${DOCKERIMAGETAG}-${BASE_OS} \
-b java_image_tag=${DOCKERIMAGETAG} \
-p ./kubernetes/dockerfiles/spark/${Dockerfile} \
build
and with that we get
REPOSITORY TAG IMAGE ID CREATED
SIZE
spark/spark 3.1.1-scala_2.12-11-jre-slim-buster 6ef051218938 2
minutes ago 635MB
openjdk 8-jre-slim 0d0a85fdf642 4 days
ago 187MB
openjdk 11-jre-slim eb77da2ec13c 4 weeks
ago 221MB
I guess that is descriptive enough with README file as well
HTH
view my Linkedin profile
<https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
*Disclaimer:* Use it at your own risk. Any and all responsibility for any
loss, damage or destruction of data or any other property which may arise
from relying on this email's technical content is explicitly disclaimed.
The author will in no case be liable for any monetary damages arising from
such loss, damage or destruction.
On Sat, 21 Aug 2021 at 15:58, Ankit Gupta <[email protected]> wrote:
> Hey Mich
>
> Adding to what Mich is suggesting, how about having the base OS version in
> the image tag as well, like
>
> <PRODUCT_VERSION>-<SCALA_VERSION>-<JAVA_VERSION>-<BASE_OS>
>
> 3.1.2-scala_2.12-java11-slim
> 3.1.2_sparkpy-scala_2.12-java11-buster
> 3.1.2_sparkR-scala_2.12-java11-slim
>
> Regards.
>
> Ankit Prakash Gupta
> [email protected]
> LinkedIn : https://www.linkedin.com/in/infoankitp/
> Medium: https://medium.com/@info.ankitp
>
> On Mon, Aug 16, 2021 at 5:39 PM Mich Talebzadeh <[email protected]>
> wrote:
>
>> Hi,
>>
>> I propose that for Spark docker images we follow the following convention
>> similar to flink <https://hub.docker.com/_/flink>as shown in the
>> attached file
>>
>> So for Spark we will have
>>
>>
>> <PRODUCT_VERSION>-<SCALA_VERSION>-<JAVA_VERSION>
>>
>> 3.1.2-scala_2.12-java11
>> 3.1.2_sparkpy-scala_2.12-java11
>> 3.1.2_sparkR-scala_2.12-java11
>>
>>
>> If this makes sense please respond, otherwise state your preference
>>
>>
>> HTH
>>
>>
>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>> any loss, damage or destruction of data or any other property which may
>> arise from relying on this email's technical content is explicitly
>> disclaimed. The author will in no case be liable for any monetary damages
>> arising from such loss, damage or destruction.
>>
>>
>>
>> ---------------------------------------------------------------------
>> To unsubscribe e-mail: [email protected]
>
>