Re: [VOTE] Spark 3.1.3 RC4

2022-02-16 Thread Mridul Muralidharan
+1

Signatures, digests, etc check out fine.
Checked out tag and build/tested with -Pyarn -Pmesos -Pkubernetes

Regards,
Mridul


On Wed, Feb 16, 2022 at 8:32 AM Thomas graves  wrote:

> +1
>
> Tom
>
> On Mon, Feb 14, 2022 at 2:55 PM Holden Karau  wrote:
> >
> > Please vote on releasing the following candidate as Apache Spark version
> 3.1.3.
> >
> > The vote is open until Feb. 18th at 1 PM pacific (9 PM GMT) and passes
> if a majority
> > +1 PMC votes are cast, with a minimum of 3 + 1 votes.
> >
> > [ ] +1 Release this package as Apache Spark 3.1.3
> > [ ] -1 Do not release this package because ...
> >
> > To learn more about Apache Spark, please see http://spark.apache.org/
> >
> > There are currently no open issues targeting 3.1.3 in Spark's JIRA
> https://issues.apache.org/jira/browse
> > (try project = SPARK AND "Target Version/s" = "3.1.3" AND status in
> (Open, Reopened, "In Progress"))
> > at https://s.apache.org/n79dw
> >
> >
> >
> > The tag to be voted on is v3.1.3-rc4 (commit
> > d1f8a503a26bcfb4e466d9accc5fa241a7933667):
> > https://github.com/apache/spark/tree/v3.1.3-rc4
> >
> > The release files, including signatures, digests, etc. can be found at:
> > https://dist.apache.org/repos/dist/dev/spark/v3.1.3-rc4-bin/
> >
> > Signatures used for Spark RCs can be found in this file:
> > https://dist.apache.org/repos/dist/dev/spark/KEYS
> >
> > The staging repository for this release can be found at
> > https://repository.apache.org/content/repositories/orgapachespark-1401
> >
> > The documentation corresponding to this release can be found at:
> > https://dist.apache.org/repos/dist/dev/spark/v3.1.3-rc4-docs/
> >
> > The list of bug fixes going into 3.1.3 can be found at the following URL:
> > https://s.apache.org/x0q9b
> >
> > This release is using the release script from 3.1.3
> > The release docker container was rebuilt since the previous version
> didn't have the necessary components to build the R documentation.
> >
> > FAQ
> >
> >
> > =
> > How can I help test this release?
> > =
> >
> > If you are a Spark user, you can help us test this release by taking
> > an existing Spark workload and running on this release candidate, then
> > reporting any regressions.
> >
> > If you're working in PySpark you can set up a virtual env and install
> > the current RC and see if anything important breaks, in the Java/Scala
> > you can add the staging repository to your projects resolvers and test
> > with the RC (make sure to clean up the artifact cache before/after so
> > you don't end up building with an out of date RC going forward).
> >
> > ===
> > What should happen to JIRA tickets still targeting 3.1.3?
> > ===
> >
> > The current list of open tickets targeted at 3.1.3 can be found at:
> > https://issues.apache.org/jira/projects/SPARK and search for "Target
> > Version/s" = 3.1.3
> >
> > Committers should look at those and triage. Extremely important bug
> > fixes, documentation, and API tweaks that impact compatibility should
> > be worked on immediately. Everything else please retarget to an
> > appropriate release.
> >
> > ==
> > But my bug isn't fixed?
> > ==
> >
> > In order to make timely releases, we will typically not hold the
> > release unless the bug in question is a regression from the previous
> > release. That being said, if there is something that is a regression
> > that has not been correctly targeted please ping me or a committer to
> > help target the issue.
> >
> > Note: I added an extra day to the vote since I know some folks are
> likely busy on the 14th with partner(s).
> >
> >
> > --
> > Twitter: https://twitter.com/holdenkarau
> > Books (Learning Spark, High Performance Spark, etc.):
> https://amzn.to/2MaRAG9
> > YouTube Live Streams: https://www.youtube.com/user/holdenkarau
>
> -
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>


Re: [VOTE] Spark 3.1.3 RC4

2022-02-16 Thread Thomas graves
+1

Tom

On Mon, Feb 14, 2022 at 2:55 PM Holden Karau  wrote:
>
> Please vote on releasing the following candidate as Apache Spark version 
> 3.1.3.
>
> The vote is open until Feb. 18th at 1 PM pacific (9 PM GMT) and passes if a 
> majority
> +1 PMC votes are cast, with a minimum of 3 + 1 votes.
>
> [ ] +1 Release this package as Apache Spark 3.1.3
> [ ] -1 Do not release this package because ...
>
> To learn more about Apache Spark, please see http://spark.apache.org/
>
> There are currently no open issues targeting 3.1.3 in Spark's JIRA 
> https://issues.apache.org/jira/browse
> (try project = SPARK AND "Target Version/s" = "3.1.3" AND status in (Open, 
> Reopened, "In Progress"))
> at https://s.apache.org/n79dw
>
>
>
> The tag to be voted on is v3.1.3-rc4 (commit
> d1f8a503a26bcfb4e466d9accc5fa241a7933667):
> https://github.com/apache/spark/tree/v3.1.3-rc4
>
> The release files, including signatures, digests, etc. can be found at:
> https://dist.apache.org/repos/dist/dev/spark/v3.1.3-rc4-bin/
>
> Signatures used for Spark RCs can be found in this file:
> https://dist.apache.org/repos/dist/dev/spark/KEYS
>
> The staging repository for this release can be found at
> https://repository.apache.org/content/repositories/orgapachespark-1401
>
> The documentation corresponding to this release can be found at:
> https://dist.apache.org/repos/dist/dev/spark/v3.1.3-rc4-docs/
>
> The list of bug fixes going into 3.1.3 can be found at the following URL:
> https://s.apache.org/x0q9b
>
> This release is using the release script from 3.1.3
> The release docker container was rebuilt since the previous version didn't 
> have the necessary components to build the R documentation.
>
> FAQ
>
>
> =
> How can I help test this release?
> =
>
> If you are a Spark user, you can help us test this release by taking
> an existing Spark workload and running on this release candidate, then
> reporting any regressions.
>
> If you're working in PySpark you can set up a virtual env and install
> the current RC and see if anything important breaks, in the Java/Scala
> you can add the staging repository to your projects resolvers and test
> with the RC (make sure to clean up the artifact cache before/after so
> you don't end up building with an out of date RC going forward).
>
> ===
> What should happen to JIRA tickets still targeting 3.1.3?
> ===
>
> The current list of open tickets targeted at 3.1.3 can be found at:
> https://issues.apache.org/jira/projects/SPARK and search for "Target
> Version/s" = 3.1.3
>
> Committers should look at those and triage. Extremely important bug
> fixes, documentation, and API tweaks that impact compatibility should
> be worked on immediately. Everything else please retarget to an
> appropriate release.
>
> ==
> But my bug isn't fixed?
> ==
>
> In order to make timely releases, we will typically not hold the
> release unless the bug in question is a regression from the previous
> release. That being said, if there is something that is a regression
> that has not been correctly targeted please ping me or a committer to
> help target the issue.
>
> Note: I added an extra day to the vote since I know some folks are likely 
> busy on the 14th with partner(s).
>
>
> --
> Twitter: https://twitter.com/holdenkarau
> Books (Learning Spark, High Performance Spark, etc.): https://amzn.to/2MaRAG9
> YouTube Live Streams: https://www.youtube.com/user/holdenkarau

-
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org



Re: Which manufacturers' GPUs support Spark?

2022-02-16 Thread Sean Owen
Spark itself does not use GPUs, and is agnostic to what GPUs exist on a
cluster, scheduled by the resource manager, and used by an application.
In practice, virtually all GPU-related use cases (for deep learning for
example) use CUDA, and this is NVIDIA-specific. Certainly, RAPIDS is from
NVIDIA.

On Wed, Feb 16, 2022 at 7:03 AM 15927907...@163.com <15927907...@163.com>
wrote:

> Hello,
> We have done some Spark GPU accelerated work using the spark-rapids
> component(https://github.com/NVIDIA/spark-rapids). However, we found that
> this component currently only supports Nvidia GPU, and on the official
> Spark website, we did not see the manufacturer's description of the GPU
> supported by spark(
> https://spark.apache.org/docs/3.2.1/configuration.html#custom-resource-scheduling-and-configuration-overview).
> So, Can Spark also support GPUs from other manufacturers? such as AMD.
> Looking forward to your reply.
>
> --
> 15927907...@163.com
>


Re: [VOTE] Spark 3.1.3 RC4

2022-02-16 Thread Gengliang Wang
+1 (non-binding)

On Wed, Feb 16, 2022 at 1:28 PM Wenchen Fan  wrote:

> +1
>
> On Tue, Feb 15, 2022 at 3:59 PM Yuming Wang  wrote:
>
>> +1 (non-binding).
>>
>> On Tue, Feb 15, 2022 at 10:22 AM Ruifeng Zheng 
>> wrote:
>>
>>> +1 (non-binding)
>>>
>>> checked the release script issue Dongjoon mentioned:
>>>
>>> curl -s
>>> https://dist.apache.org/repos/dist/dev/spark/v3.1.3-rc4-bin/spark-3.1.3-bin-hadoop2.7.tgz
>>> | tar tz | grep hadoop-common
>>> spark-3.1.3-bin-hadoop2.7/jars/hadoop-common-2.7.4
>>>
>>>
>>> -- 原始邮件 --
>>> *发件人:* "Sean Owen" ;
>>> *发送时间:* 2022年2月15日(星期二) 上午10:01
>>> *收件人:* "Holden Karau";
>>> *抄送:* "dev";
>>> *主题:* Re: [VOTE] Spark 3.1.3 RC4
>>>
>>> Looks good to me, same results as last RC, +1
>>>
>>> On Mon, Feb 14, 2022 at 2:55 PM Holden Karau 
>>> wrote:
>>>
 Please vote on releasing the following candidate as Apache Spark
 version 3.1.3.

 The vote is open until Feb. 18th at 1 PM pacific (9 PM GMT) and passes
 if a majority
 +1 PMC votes are cast, with a minimum of 3 + 1 votes.

 [ ] +1 Release this package as Apache Spark 3.1.3
 [ ] -1 Do not release this package because ...

 To learn more about Apache Spark, please see http://spark.apache.org/

 There are currently no open issues targeting 3.1.3 in Spark's JIRA
 https://issues.apache.org/jira/browse
 (try project = SPARK AND "Target Version/s" = "3.1.3" AND status in
 (Open, Reopened, "In Progress"))
 at https://s.apache.org/n79dw



 The tag to be voted on is v3.1.3-rc4 (commit
 d1f8a503a26bcfb4e466d9accc5fa241a7933667):
 https://github.com/apache/spark/tree/v3.1.3-rc4

 The release files, including signatures, digests, etc. can be found at:
 https://dist.apache.org/repos/dist/dev/spark/v3.1.3-rc4-bin/

 Signatures used for Spark RCs can be found in this file:
 https://dist.apache.org/repos/dist/dev/spark/KEYS

 The staging repository for this release can be found at
 https://repository.apache.org/content/repositories/orgapachespark-1401

 The documentation corresponding to this release can be found at:
 https://dist.apache.org/repos/dist/dev/spark/v3.1.3-rc4-docs/

 The list of bug fixes going into 3.1.3 can be found at the following
 URL:
 https://s.apache.org/x0q9b

 This release is using the release script from 3.1.3
 The release docker container was rebuilt since the previous version
 didn't have the necessary components to build the R documentation.

 FAQ


 =
 How can I help test this release?
 =

 If you are a Spark user, you can help us test this release by taking
 an existing Spark workload and running on this release candidate, then
 reporting any regressions.

 If you're working in PySpark you can set up a virtual env and install
 the current RC and see if anything important breaks, in the Java/Scala
 you can add the staging repository to your projects resolvers and test
 with the RC (make sure to clean up the artifact cache before/after so
 you don't end up building with an out of date RC going forward).

 ===
 What should happen to JIRA tickets still targeting 3.1.3?
 ===

 The current list of open tickets targeted at 3.1.3 can be found at:
 https://issues.apache.org/jira/projects/SPARK and search for "Target
 Version/s" = 3.1.3

 Committers should look at those and triage. Extremely important bug
 fixes, documentation, and API tweaks that impact compatibility should
 be worked on immediately. Everything else please retarget to an
 appropriate release.

 ==
 But my bug isn't fixed?
 ==

 In order to make timely releases, we will typically not hold the
 release unless the bug in question is a regression from the previous
 release. That being said, if there is something that is a regression
 that has not been correctly targeted please ping me or a committer to
 help target the issue.

 Note: I added an extra day to the vote since I know some folks are
 likely busy on the 14th with partner(s).


 --
 Twitter: https://twitter.com/holdenkarau
 Books (Learning Spark, High Performance Spark, etc.):
 https://amzn.to/2MaRAG9  
 YouTube Live Streams: https://www.youtube.com/user/holdenkarau

>>>