I as well think the largest use case of docker images would be on
Kubernetes.

While I have nothing against us adding more varieties I think it’s
important for us to get this started with our current containers, so I’ll
do that but let’s certainly continue exploring improvements after that.

On Mon, Aug 16, 2021 at 11:28 PM Mich Talebzadeh <mich.talebza...@gmail.com>
wrote:

> Thanks for the notes all.
>
> I think we ought to consider what docker general usage is.
>
> Docker image by definition is a self contained general purpose entity
> providing spark service at the common denominator. Some docker images like
> the one for jenkins are far simpler to build as they have less moving parts
> With spark you either deploy your own set-up, get a service provider like
> Google to provide it as a service (and they provide not the most recent
> version, for example Dataproc cluster runs on Spark 3.1.1) or use docker in
> Kubernetes cluster GKE. They provide the cluster and cluster management but
> you deploy your own docker image.
>
> I don't know much about spark-R but I think if we take Spark latest and
> spark-py built on 3.7.3 and Spark-py for data science with the most widely
> used and current packages we should be OK.
>
> This set-up will work as long as the documentation goes into details of
> interpreter and package versions and provides a blueprint to build your own
> custom version of docker with whatever version you prefer.
>
> As I said if we look at flink, they provide flink with scala and java
> version and of course latest
>
> 1.13.2-scala_2.12-java8, 1.13-scala_2.12-java8, scala_2.12-java8,
> 1.13.2-scala_2.12, 1.13-scala_2.12, scala_2.12, 1.13.2-java8, 1.13-java8,
> java8, 1.13.2, 1.13, latest
>
> I personally believe that providing the most popular ones serves the
> purpose for the community and anything above and beyond has to be tailor
> made.
>
> HTH
>
>
>    view my Linkedin profile
> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>
>
>
> *Disclaimer:* Use it at your own risk. Any and all responsibility for any
> loss, damage or destruction of data or any other property which may arise
> from relying on this email's technical content is explicitly disclaimed.
> The author will in no case be liable for any monetary damages arising from
> such loss, damage or destruction.
>
>
>
>
> On Mon, 16 Aug 2021 at 18:46, Maciej <mszymkiew...@gmail.com> wrote:
>
>> I have a few concerns regarding PySpark and SparkR images.
>>
>> First of all, how do we plan to handle interpreter versions? Ideally, we
>> should provide images for all supported variants, but based on the
>> preceding discussion and the proposed naming convention, I assume it is not
>> going to happen. If that's the case, it would be great if we could fix
>> interpreter versions based on some support criteria (lowest supported,
>> lowest non-deprecated, highest supported at the time of release, etc.)
>>
>> Currently, we use the following:
>>
>>    - for R use buster-cran35 Debian repositories which install R 3.6
>>    (provided version already changed in the past and broke image build ‒
>>    SPARK-28606).
>>    - for Python we depend on the system provided python3 packages, which
>>    currently provides Python 3.7.
>>
>> which don't guarantee stability over time and might be hard to
>> synchronize with our support matrix.
>>
>> Secondly, omitting libraries which are required for the full
>> functionality and performance, specifically
>>
>>    - Numpy, Pandas and Arrow for PySpark
>>    - Arrow for SparkR
>>
>> is likely to severely limit usability of the images (out of these, Arrow
>> is probably the hardest to manage, especially when you already depend on
>> system packages to provide R or Python interpreter).
>>
>> On 8/14/21 12:43 AM, Mich Talebzadeh wrote:
>>
>> Hi,
>>
>> We can cater for multiple types (spark, spark-py and spark-r) and spark
>> versions (assuming they are downloaded and available).
>> The challenge is that these docker images built are snapshots. They
>> cannot be amended later and if you change anything by going inside docker,
>> as soon as you are logged out whatever you did is reversed.
>>
>> For example, I want to add tensorflow to my docker image. These are my
>> images
>>
>> REPOSITORY                                TAG           IMAGE ID
>>  CREATED         SIZE
>> eu.gcr.io/axial-glow-224522/spark-py      java8_3.1.1   cfbb0e69f204   5
>> days ago      2.37GB
>> eu.gcr.io/axial-glow-224522/spark         3.1.1         8d1bf8e7e47d   5
>> days ago      805MB
>>
>> using image ID I try to log in as root to the image
>>
>> *docker run -u0 -it cfbb0e69f204 bash*
>>
>> root@b542b0f1483d:/opt/spark/work-dir# pip install keras
>> Collecting keras
>>   Downloading keras-2.6.0-py2.py3-none-any.whl (1.3 MB)
>>      |████████████████████████████████| 1.3 MB 1.1 MB/s
>> Installing collected packages: keras
>> Successfully installed keras-2.6.0
>> WARNING: Running pip as the 'root' user can result in broken permissions
>> and conflicting behaviour with the system package manager. It is
>> recommended to use a virtual environment instead:
>> https://pip.pypa.io/warnings/venv
>> root@b542b0f1483d:/opt/spark/work-dir# pip list
>> Package       Version
>> ------------- -------
>> asn1crypto    0.24.0
>> cryptography  2.6.1
>> cx-Oracle     8.2.1
>> entrypoints   0.3
>> *keras         2.6.0      <--- it is here*
>> keyring       17.1.1
>> keyrings.alt  3.1.1
>> numpy         1.21.1
>> pip           21.2.3
>> py4j          0.10.9
>> pycrypto      2.6.1
>> PyGObject     3.30.4
>> pyspark       3.1.2
>> pyxdg         0.25
>> PyYAML        5.4.1
>> SecretStorage 2.3.1
>> setuptools    57.4.0
>> six           1.12.0
>> wheel         0.32.3
>> root@b542b0f1483d:/opt/spark/work-dir# exit
>>
>> Now I exited from the image and try to log in again
>> (pyspark_venv) hduser@rhes76: /home/hduser/dba/bin/build> docker run -u0
>> -it cfbb0e69f204 bash
>>
>> root@5231ee95aa83:/opt/spark/work-dir# pip list
>> Package       Version
>> ------------- -------
>> asn1crypto    0.24.0
>> cryptography  2.6.1
>> cx-Oracle     8.2.1
>> entrypoints   0.3
>> keyring       17.1.1
>> keyrings.alt  3.1.1
>> numpy         1.21.1
>> pip           21.2.3
>> py4j          0.10.9
>> pycrypto      2.6.1
>> PyGObject     3.30.4
>> pyspark       3.1.2
>> pyxdg         0.25
>> PyYAML        5.4.1
>> SecretStorage 2.3.1
>> setuptools    57.4.0
>> six           1.12.0
>> wheel         0.32.3
>>
>> *Hm that keras is not there*. The docker Image cannot be altered after
>> build! So once the docker image is created that is just a snapshot.
>> However, it will still have tons of useful stuff for most
>> users/organisations. My suggestions is to create for a given type (spark,
>> spark-py etc):
>>
>>
>>    1. One vanilla flavour for everyday use with few useful packages
>>    2. One for medium use with most common packages for ETL/ELT stuff
>>    3. One specialist for ML etc with keras, tensorflow and anything else
>>    needed
>>
>>
>> These images should be maintained as we currently maintain spark releases
>> with accompanying documentation. Any reason why we cannot maintain
>> ourselves?
>>
>> HTH
>>
>>    view my Linkedin profile
>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>
>>
>>
>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>> any loss, damage or destruction of data or any other property which may
>> arise from relying on this email's technical content is explicitly
>> disclaimed. The author will in no case be liable for any monetary damages
>> arising from such loss, damage or destruction.
>>
>>
>>
>>
>> On Fri, 13 Aug 2021 at 17:26, Holden Karau <hol...@pigscanfly.ca> wrote:
>>
>>> So we actually do have a script that does the build already it's more a
>>> matter of publishing the results for easier use. Currently the script
>>> produces three images spark, spark-py, and spark-r. I can certainly see a
>>> solid reason to publish like with a jdk11 & jdk8 suffix as well if there is
>>> interest in the community. If we want to have a say spark-py-pandas for a
>>> Spark container image with everything necessary for the Koalas stuff to
>>> work then I think that could be a great PR from someone to add :)
>>>
>>> On Fri, Aug 13, 2021 at 1:00 AM Mich Talebzadeh <
>>> mich.talebza...@gmail.com> wrote:
>>>
>>>> should read PySpark
>>>>
>>>>
>>>>
>>>>    view my Linkedin profile
>>>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>>>
>>>>
>>>>
>>>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>>>> any loss, damage or destruction of data or any other property which may
>>>> arise from relying on this email's technical content is explicitly
>>>> disclaimed. The author will in no case be liable for any monetary damages
>>>> arising from such loss, damage or destruction.
>>>>
>>>>
>>>>
>>>>
>>>> On Fri, 13 Aug 2021 at 08:51, Mich Talebzadeh <
>>>> mich.talebza...@gmail.com> wrote:
>>>>
>>>>> Agreed.
>>>>>
>>>>> I have already built a few latest for Spark and PYSpark on 3.1.1 with
>>>>> Java 8 as I found out Java 11 does not work with Google BigQuery data
>>>>> warehouse. However, to hack the Dockerfile one finds out the hard way.
>>>>>
>>>>> For example how to add additional Python libraries like tensorflow
>>>>> etc. Loading these libraries through Kubernetes is not practical as
>>>>> unzipping and installing it through --py-files etc will take considerable
>>>>> time so they need to be added to the dockerfile at the built time in
>>>>> directory for Python under Kubernetes
>>>>>
>>>>> /opt/spark/kubernetes/dockerfiles/spark/bindings/python
>>>>>
>>>>> RUN pip install pyyaml numpy cx_Oracle tensorflow ....
>>>>>
>>>>> Also you will need curl to test the ports from inside the docker
>>>>>
>>>>> RUN apt-get update && apt-get install -y curl
>>>>> RUN ["apt-get","install","-y","vim"]
>>>>>
>>>>> As I said I am happy to build these specific dockerfiles plus the
>>>>> complete documentation for it. I have already built one for Google (GCP).
>>>>> The difference between Spark and PySpark version is that in Spark/scala a
>>>>> fat jar file will contain all needed. That is not the case with Python I 
>>>>> am
>>>>> afraid.
>>>>>
>>>>> HTH
>>>>>
>>>>>
>>>>>    view my Linkedin profile
>>>>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>>>>
>>>>>
>>>>>
>>>>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>>>>> any loss, damage or destruction of data or any other property which may
>>>>> arise from relying on this email's technical content is explicitly
>>>>> disclaimed. The author will in no case be liable for any monetary damages
>>>>> arising from such loss, damage or destruction.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Fri, 13 Aug 2021 at 08:13, Bode, Meikel, NMA-CFD <
>>>>> meikel.b...@bertelsmann.de> wrote:
>>>>>
>>>>>> Hi all,
>>>>>>
>>>>>>
>>>>>>
>>>>>> I am Meikel Bode and only an interested reader of dev and user list.
>>>>>> Anyway, I would appreciate to have official docker images available.
>>>>>>
>>>>>> Maybe one could get inspiration from the Jupyter docker stacks and
>>>>>> provide an hierarchy of different images like this:
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> https://jupyter-docker-stacks.readthedocs.io/en/latest/using/selecting.html#image-relationships
>>>>>>
>>>>>>
>>>>>>
>>>>>> Having a core image only supporting Java, an extended supporting
>>>>>> Python and/or R etc.
>>>>>>
>>>>>>
>>>>>>
>>>>>> Looking forward to the discussion.
>>>>>>
>>>>>>
>>>>>>
>>>>>> Best,
>>>>>>
>>>>>> Meikel
>>>>>>
>>>>>>
>>>>>>
>>>>>> *From:* Mich Talebzadeh <mich.talebza...@gmail.com>
>>>>>> *Sent:* Freitag, 13. August 2021 08:45
>>>>>> *Cc:* dev <dev@spark.apache.org>
>>>>>> *Subject:* Re: Time to start publishing Spark Docker Images?
>>>>>>
>>>>>>
>>>>>>
>>>>>> I concur this is a good idea and certainly worth exploring.
>>>>>>
>>>>>>
>>>>>>
>>>>>> In practice, preparing docker images as deployable will throw some
>>>>>> challenges because creating docker for Spark  is not really a singular
>>>>>> modular unit, say  creating docker for Jenkins. It involves different
>>>>>> versions and different images for Spark and PySpark and most likely will
>>>>>> end up as part of Kubernetes deployment.
>>>>>>
>>>>>>
>>>>>>
>>>>>> Individuals and organisations will deploy it as the first cut. Great
>>>>>> but I equally feel that good documentation on how to build a consumable
>>>>>> deployable image will be more valuable.  FRom my own experience the 
>>>>>> current
>>>>>> documentation should be enhanced, for example how to deploy working
>>>>>> directories, additional Python packages, build with different Java
>>>>>> versions  (version 8 or version 11) etc.
>>>>>>
>>>>>>
>>>>>>
>>>>>> HTH
>>>>>>
>>>>>>
>>>>>>
>>>>>>    view my Linkedin profile
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwww.linkedin.com%2Fin%2Fmich-talebzadeh-ph-d-5205b2%2F&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790679755%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=0CkL3HZo9FNVUOnLQ4CYs29Z9HfrwE4xDqLgVmMbr10%3D&reserved=0>
>>>>>>
>>>>>>
>>>>>>
>>>>>> *Disclaimer:* Use it at your own risk. Any and all responsibility
>>>>>> for any loss, damage or destruction of data or any other property which 
>>>>>> may
>>>>>> arise from relying on this email's technical content is explicitly
>>>>>> disclaimed. The author will in no case be liable for any monetary damages
>>>>>> arising from such loss, damage or destruction.
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Fri, 13 Aug 2021 at 01:54, Holden Karau <hol...@pigscanfly.ca>
>>>>>> wrote:
>>>>>>
>>>>>> Awesome, I've filed an INFRA ticket to get the ball rolling.
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Thu, Aug 12, 2021 at 5:48 PM John Zhuge <jzh...@apache.org> wrote:
>>>>>>
>>>>>> +1
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Thu, Aug 12, 2021 at 5:44 PM Hyukjin Kwon <gurwls...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>> +1, I think we generally agreed upon having it. Thanks Holden for
>>>>>> headsup and driving this.
>>>>>>
>>>>>> +@Dongjoon Hyun <dongj...@apache.org> FYI
>>>>>>
>>>>>>
>>>>>>
>>>>>> 2021년 7월 22일 (목) 오후 12:22, Kent Yao <yaooq...@gmail.com>님이 작성:
>>>>>>
>>>>>> +1
>>>>>>
>>>>>>
>>>>>>
>>>>>> Bests,
>>>>>>
>>>>>>
>>>>>>
>>>>>> *Kent Yao*
>>>>>>
>>>>>> @ Data Science Center, Hangzhou Research Institute, NetEase Corp.
>>>>>>
>>>>>> *a spark* *enthusiast*
>>>>>>
>>>>>> *kyuubi
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fyaooqinn%2Fkyuubi&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790679755%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=ZkE%2BAK4%2BUO9JsDzZlAfY5gsATCVm5hidLCp7EGxAWiY%3D&reserved=0>**is
>>>>>> a unified* *multi-tenant* *JDBC interface for large-scale data
>>>>>> processing and analytics,* *built on top of* *Apache Spark
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=http%3A%2F%2Fspark.apache.org%2F&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790689711%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=4YYZ61B6datdx2GsxqnEUOpYuJUn35egYRQSVnUxtF0%3D&reserved=0>*
>>>>>> *.*
>>>>>> *spark-authorizer
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fyaooqinn%2Fspark-authorizer&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790689711%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=P6TMaSh7UeXVyv79RiRqdBpipaIjh2o3DhRs0GGhWF4%3D&reserved=0>**A
>>>>>> Spark SQL extension which provides SQL Standard Authorization for* 
>>>>>> *Apache
>>>>>> Spark
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=http%3A%2F%2Fspark.apache.org%2F&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790689711%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=4YYZ61B6datdx2GsxqnEUOpYuJUn35egYRQSVnUxtF0%3D&reserved=0>*
>>>>>> *.*
>>>>>> *spark-postgres
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fyaooqinn%2Fspark-postgres&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790699667%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=cCM9mLZBaZTF4WYzm22eIf4CU%2FfiiWCD0FUSfXSmaJA%3D&reserved=0>
>>>>>>  **A
>>>>>> library for reading data from and transferring data to Postgres / 
>>>>>> Greenplum
>>>>>> with Spark SQL and DataFrames, 10~100x faster.*
>>>>>> *itatchi
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fyaooqinn%2Fspark-func-extras&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790699667%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=sEhn0HXSzsPSBKhXZlzwQErwwFtcTdTYFqeG9FVpROU%3D&reserved=0>**A
>>>>>> library* *that brings useful functions from various modern database
>>>>>> management systems to* *Apache Spark
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=http%3A%2F%2Fspark.apache.org%2F&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790699667%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=ZIdOvEDv%2FDZWYAB3Bnm4cD1YBVVl3aaHjLiz1HSDsY0%3D&reserved=0>*
>>>>>> *.*
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> On 07/22/2021 11:13,Holden Karau<hol...@pigscanfly.ca>
>>>>>> <hol...@pigscanfly.ca> wrote:
>>>>>>
>>>>>> Hi Folks,
>>>>>>
>>>>>>
>>>>>>
>>>>>> Many other distributed computing (
>>>>>> https://hub.docker.com/r/rayproject/ray
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fhub.docker.com%2Fr%2Frayproject%2Fray&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790709619%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=%2F%2BPp69I10cyEeSTp6POoNZObOpkkzcZfB35vcdkR8P8%3D&reserved=0>
>>>>>> https://hub.docker.com/u/daskdev
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fhub.docker.com%2Fu%2Fdaskdev&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790709619%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=jrQU9WbtFLM1T71SVaZwa0U57F8GcBSFHmXiauQtou0%3D&reserved=0>)
>>>>>> and ASF projects (https://hub.docker.com/u/apache
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fhub.docker.com%2Fu%2Fapache&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790719573%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=yD8NWSYhhL6%2BDb3D%2BfD%2F8ynKAL4Wp8BKDMHV0n7jHHM%3D&reserved=0>)
>>>>>> now publish their images to dockerhub.
>>>>>>
>>>>>>
>>>>>>
>>>>>> We've already got the docker image tooling in place, I think we'd
>>>>>> need to ask the ASF to grant permissions to the PMC to publish containers
>>>>>> and update the release steps but I think this could be useful for folks.
>>>>>>
>>>>>>
>>>>>>
>>>>>> Cheers,
>>>>>>
>>>>>>
>>>>>>
>>>>>> Holden
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>>
>>>>>> Twitter: https://twitter.com/holdenkarau
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Ftwitter.com%2Fholdenkarau&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790719573%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=4qhg1CzKNiiRZkbvzKMp7WL4BoYLzPZ%2FOpFwHu8KNmg%3D&reserved=0>
>>>>>>
>>>>>> Books (Learning Spark, High Performance Spark, etc.):
>>>>>> https://amzn.to/2MaRAG9
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Famzn.to%2F2MaRAG9&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790719573%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=5UCR1Qn0fLovLAdTFnJBnLYF3e2NRnL8wEYPhCfLf2A%3D&reserved=0>
>>>>>>
>>>>>> YouTube Live Streams: https://www.youtube.com/user/holdenkarau
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwww.youtube.com%2Fuser%2Fholdenkarau&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790729540%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=LbsZdvDNTAc804N2dknen%2BoJavleIsh5vwpNaj7xIio%3D&reserved=0>
>>>>>>
>>>>>> ---------------------------------------------------------------------
>>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>
>>>>>> --
>>>>>>
>>>>>> John Zhuge
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>>
>>>>>> Twitter: https://twitter.com/holdenkarau
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Ftwitter.com%2Fholdenkarau&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790729540%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=x6fXgTuoQqVYqu9JPbt0hG2P0zl6l3p%2FrU5bDng85AY%3D&reserved=0>
>>>>>>
>>>>>> Books (Learning Spark, High Performance Spark, etc.):
>>>>>> https://amzn.to/2MaRAG9
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Famzn.to%2F2MaRAG9&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790729540%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=WCHuF%2BcEl0rBZyVOePRQT1AOefwRDlIavu9B0wDmmOk%3D&reserved=0>
>>>>>>
>>>>>> YouTube Live Streams: https://www.youtube.com/user/holdenkarau
>>>>>> <https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwww.youtube.com%2Fuser%2Fholdenkarau&data=04%7C01%7CMeikel.Bode%40bertelsmann.de%7Cd97d97be540246aa975308d95e260c99%7C1ca8bd943c974fc68955bad266b43f0b%7C0%7C0%7C637644339790739490%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=52hSM52z%2FFRahVO%2FcRwJ6eDuDInvhhtt1xQfbhMRazQ%3D&reserved=0>
>>>>>>
>>>>>>
>>>
>>> --
>>> Twitter: https://twitter.com/holdenkarau
>>> Books (Learning Spark, High Performance Spark, etc.):
>>> https://amzn.to/2MaRAG9  <https://amzn.to/2MaRAG9>
>>> YouTube Live Streams: https://www.youtube.com/user/holdenkarau
>>>
>>
>> --
>> Best regards,
>> Maciej Szymkiewicz
>>
>> Web: https://zero323.net
>> Keybase: https://keybase.io/zero323
>> Gigs: https://www.codementor.io/@zero323
>> PGP: A30CEF0C31A501EC
>>
>> --
Twitter: https://twitter.com/holdenkarau
Books (Learning Spark, High Performance Spark, etc.):
https://amzn.to/2MaRAG9  <https://amzn.to/2MaRAG9>
YouTube Live Streams: https://www.youtube.com/user/holdenkarau

Reply via email to