Github user saurfang closed the pull request at:
https://github.com/apache/spark/pull/8375
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user saurfang commented on the pull request:
https://github.com/apache/spark/pull/8375#issuecomment-202711183
Thanks @davies. This looks to be a very reasonable workaround.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
Github user davies commented on the pull request:
https://github.com/apache/spark/pull/8375#issuecomment-202458325
@saurfang This could be done easily with DataFrame APIs:
```
val df = rdd.map(t => Row(gen_key(t), t)).toDF("key", "text")
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/8375#issuecomment-202342748
Build finished. Test FAILed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/8375#issuecomment-202342752
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/8375#issuecomment-202342740
[Test build #54310 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/54310/console)
for PR 8375 at commit
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/8375#issuecomment-202342247
[Test build #54310 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/54310/consoleFull)
for PR 8375 at commit
Github user utkarsh2012 commented on the pull request:
https://github.com/apache/spark/pull/8375#issuecomment-159078940
@saurfang Do you plan to merge this or you are waiting for @silasdavis's
changes?
---
If your project is set up for it, you can reply to this email and have your
Github user saurfang commented on the pull request:
https://github.com/apache/spark/pull/8375#issuecomment-159143064
I myself definitely want to get this merged. That being said, this will
require a spark committer to drive the process. If you feel strongly about this
feature, voice
Github user davies commented on the pull request:
https://github.com/apache/spark/pull/8375#issuecomment-144185163
@silasdavis What's the progress on your side?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user silasdavis commented on the pull request:
https://github.com/apache/spark/pull/8375#issuecomment-134598259
Let's not make the best the enemy of the good. I can base the more general
approach off this work when I have time.
---
If your project is set up for it, you can
Github user n-my commented on the pull request:
https://github.com/apache/spark/pull/8375#issuecomment-134429446
Can't wait for that to be released :+1:
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
GitHub user saurfang opened a pull request:
https://github.com/apache/spark/pull/8375
[SPARK-3533][Core] Add saveAsTextFileByKey() method to RDDs
This adds the functionality of saving a `RDD[(K, V)]` to multiple text
files split by key. It covers Scala/Java/Python API.
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/8375#issuecomment-133760463
Can one of the admins verify this patch?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user silasdavis commented on the pull request:
https://github.com/apache/spark/pull/4895#issuecomment-133078691
I am trying to get this re-opened, please see:
https://issues.apache.org/jira/browse/SPARK-3533
I have a prototype implementation that allows you to write
Github user baconmania commented on the pull request:
https://github.com/apache/spark/pull/4895#issuecomment-123847364
Couldn't find this on Packages... any chance this'll get ported?
---
If your project is set up for it, you can reply to this email and have your
reply appear on
Github user ilganeli commented on the pull request:
https://github.com/apache/spark/pull/4895#issuecomment-109762391
Did not - I can port it over if you need it. Thanks !
Thank you,
Ilya Ganelin
-Original Message-
From: brendancol
Github user brendancol commented on the pull request:
https://github.com/apache/spark/pull/4895#issuecomment-109747790
hey did this ever make it into Packages? Great idea and would love to use
it.
---
If your project is set up for it, you can reply to this email and have your
reply
Github user brendancol commented on the pull request:
https://github.com/apache/spark/pull/4895#issuecomment-109800470
Hey great. yeah just looking for this functionality on the python side.
I'm currently doing n+1 select statements to do a multiple outputs from a
dataframe, looked
Github user ilganeli closed the pull request at:
https://github.com/apache/spark/pull/4895
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user ilganeli commented on the pull request:
https://github.com/apache/spark/pull/4895#issuecomment-96772751
Closing this PR, I'll move it to Packages at some point.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as
Github user ilganeli commented on the pull request:
https://github.com/apache/spark/pull/4895#issuecomment-93486074
All - anything to do here?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user nchammas commented on the pull request:
https://github.com/apache/spark/pull/4895#issuecomment-93489570
@ilganeli - I think this is a useful feature, but given that there isn't a
strong committer sponsor (I am just a random contributor) to take this PR
through to the end,
Github user ilganeli commented on the pull request:
https://github.com/apache/spark/pull/4895#issuecomment-90980662
@davies I've created a JIRA for the Python API here along with some starter
code.
https://issues.apache.org/jira/browse/SPARK-6780
---
If your project is set up
24 matches
Mail list logo