Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13786#discussion_r67771049
--- Diff: R/pkg/inst/tests/testthat/test_sparkSQL.R ---
@@ -1397,6 +1397,26 @@ test_that("group by, agg functions", {
unlink
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13734
LGTM. Thanks for this PR @dongjoon-hyun -- This is very useful to have
going forward.
Merging this to master, branch-2.0
---
If your project is set up for it, you can reply to this email
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13751
LGTM. Merging this to master, branch-2.0
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13768
Thanks for the updates. LGTM. Merging this to master, branch-2.0
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13734
Ok. I think this is a reasonable proposal. I will take one more final pass
on this PR today
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13782
Thanks - LGTM. Merging this to master and branch-2.0 -- (We can reverify
this is #13734 I guess ?)
---
If your project is set up for it, you can reply to this email and have your
reply appear on
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13295
@dongjoon-hyun @felixcheung -- I think @mhnatiuk is busy. If you have time
it will be cool to submit another version of this PR as I think this is a
useful function for R users.
---
If your
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13768#discussion_r67748564
--- Diff: R/pkg/R/functions.R ---
@@ -1179,6 +1179,27 @@ setMethod("soundex",
column(jc)
})
+#
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13109
@felixcheung Is this PR still relevant ?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13023
@mengxr @yanboliang Is this PR still active ? Just checking if this is
something we should track for the 2.0 release
---
If your project is set up for it, you can reply to this email and have
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13751
Changes look pretty good to me. Thanks -- I just had a couple of minor
comments. Also I think we should look at #13592 to make sure there are no other
inconsistencies in how we describe
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13751#discussion_r67745787
--- Diff: docs/sparkr.md ---
@@ -158,20 +152,19 @@ write.df(people, path="people.parquet",
source="parquet", mode="overwrite&q
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13782
cc @dongjoon-hyun
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13751#discussion_r67744654
--- Diff: R/pkg/R/schema.R ---
@@ -29,11 +29,8 @@
#' @export
#' @examples
#'\dontrun{
-#' sc <- spark
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13751#discussion_r67744683
--- Diff: R/pkg/R/schema.R ---
@@ -90,13 +87,10 @@ print.structType <- function(x, ...) {
#' @export
#' @examples
#'\d
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13752
Thanks - merging this to master and branch-2.0 after jenkins passes
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13734
- Yeah we can make documentation changes after the RC as the doc updates
are pushed separately to the Spark website. However I prefer to get R doc
changes in before the RC as these are the ones
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13763
LGTM. Merging this to master and branch-2.0
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13768#discussion_r67740472
--- Diff: R/pkg/R/functions.R ---
@@ -1179,6 +1179,27 @@ setMethod("soundex",
column(jc)
})
+#' spa
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13753
Thanks @felixcheung @liancheng - LGTM. Merging this to master, branch-2.0
Are there any other catalog functions that have changed in Spark 2.0 that
we also expose in SparkR ? Or are
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13752
LGTM. Minor comment about docs
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13752#discussion_r67739171
--- Diff: R/pkg/R/context.R ---
@@ -252,17 +252,19 @@ setCheckpointDir <- function(sc, dirName) {
#' }
#'
#' @r
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13660
Yeah we can remove the duplication by having separate rd files or by just
removing documentation for the overlapping arguments (I think in this case `x`
and `func` are the same for `dapply` and
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13774
Thanks @dongjoon-hyun and @felixcheung -- Merging this to master and
branch-2.0
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13760
Thanks @NarineK -- cc @sun-rui for review
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13635
LGTM. Merging this to master and branch-2.0
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13734
Yeah since the notes appear in a separate section I think its better to be
more explicit -- so `read.df since 1.6.0` will be good for these cases
---
If your project is set up for it, you can
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13635
Thanks - Could you also bring this up to date with master branch ?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13584
@jkbradley Is this important for 2.0 ?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13722
Thanks @felixcheung - LGTM. Merging this to master and branch-2.0
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13635
@felixcheung Thanks for the update. The change looks pretty good to me. I
think there are 2-3 follow up JIRAs I opened from the review that can have
separate PRs. There was only one comment in
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67585149
--- Diff: R/pkg/inst/profile/shell.R ---
@@ -18,17 +18,17 @@
.First <- function() {
home <- Sys.getenv("SPARK_HOME")
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67585088
--- Diff: R/pkg/inst/tests/testthat/test_context.R ---
@@ -156,7 +160,8 @@ test_that("sparkJars sparkPackages as comma-separated
st
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67585014
--- Diff: R/pkg/inst/tests/testthat/test_context.R ---
@@ -47,31 +47,33 @@ test_that("Check masked functions", {
test_that("repe
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13721
Thanks @dongjoon-hyun - LGTM. Will merge once Jenkins passes
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67584715
--- Diff: R/pkg/inst/tests/testthat/test_context.R ---
@@ -156,7 +160,8 @@ test_that("sparkJars sparkPackages as comma-separated
st
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13721#discussion_r67582965
--- Diff: R/pkg/R/DataFrame.R ---
@@ -2908,3 +2908,39 @@ setMethod("write.jdbc",
write <- callJMethod(write,
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13721#discussion_r67582902
--- Diff: R/pkg/R/DataFrame.R ---
@@ -2908,3 +2908,39 @@ setMethod("write.jdbc",
write <- callJMethod(write,
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67580826
--- Diff: R/pkg/R/SQLContext.R ---
@@ -615,11 +619,12 @@ clearCache <- function() {
#' @method dropTempTable default
dropTempTable
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13722
LGTM. I just had a couple of points about the docs
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13721#discussion_r67457504
--- Diff: R/pkg/inst/tests/testthat/test_sparkSQL.R ---
@@ -2264,6 +2264,14 @@ test_that("createDataFrame sqlContext parameter
backward compatib
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13721#discussion_r67457476
--- Diff: R/pkg/inst/tests/testthat/test_sparkSQL.R ---
@@ -2264,6 +2264,14 @@ test_that("createDataFrame sqlContext parameter
backward compatib
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13721#discussion_r67457426
--- Diff: R/pkg/R/DataFrame.R ---
@@ -2884,3 +2884,39 @@ setMethod("write.jdbc",
write <- callJMethod(write,
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13722#discussion_r67457292
--- Diff: R/pkg/R/DataFrame.R ---
@@ -455,6 +455,17 @@ setMethod("createOrReplaceTempView",
invisible(callJMe
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13722#discussion_r67457183
--- Diff: R/pkg/R/DataFrame.R ---
@@ -455,6 +455,17 @@ setMethod("createOrReplaceTempView",
invisible(callJMe
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13684
LGTM. Merging into master, branch-2.0
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13714#discussion_r67438648
--- Diff: examples/src/main/r/data-manipulation.R ---
@@ -75,8 +75,8 @@ destDF <- select(flightsDF, "dest", "cancelled")
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13684#discussion_r67412216
--- Diff: R/pkg/R/DataFrame.R ---
@@ -1949,14 +1950,24 @@ setMethod("where",
#' path <- "path/to/file.json"
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13714
Merging this to master and branch-2.0
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13714
LGTM. Thanks for the fix @dongjoon-hyun
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67387830
--- Diff: R/pkg/R/sparkR.R ---
@@ -270,27 +291,97 @@ sparkRSQL.init <- function(jsc = NULL) {
#'}
sparkRHive.init <- function
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67387375
--- Diff: R/pkg/NAMESPACE ---
@@ -6,10 +6,15 @@ importFrom(methods, setGeneric, setMethod, setOldClass)
#useDynLib(SparkR, stringHashCode
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67387255
--- Diff: R/pkg/R/sparkR.R ---
@@ -270,27 +291,97 @@ sparkRSQL.init <- function(jsc = NULL) {
#'}
sparkRHive.init <- function
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67387218
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/api/r/SQLUtils.scala
---
@@ -18,27 +18,56 @@
package org.apache.spark.sql.api.r
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67383511
--- Diff: R/pkg/R/sparkR.R ---
@@ -270,27 +291,97 @@ sparkRSQL.init <- function(jsc = NULL) {
#'}
sparkRHive.init <- function
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/12836
Merging this to master and branch-2.0
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/12836#discussion_r67266001
--- Diff: R/pkg/R/DataFrame.R ---
@@ -1266,6 +1266,83 @@ setMethod("dapplyCollect",
ldf
})
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13635
Thanks @felixcheung for the PR. Other than the naming issues, I think the
code changes look pretty good to me. I think there are some more docs,
programming guide changes we'll need to make
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67263062
--- Diff: R/pkg/R/sparkR.R ---
@@ -270,27 +291,97 @@ sparkRSQL.init <- function(jsc = NULL) {
#'}
sparkRHive.init <- function
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67262684
--- Diff: R/pkg/R/sparkR.R ---
@@ -270,27 +291,97 @@ sparkRSQL.init <- function(jsc = NULL) {
#'}
sparkRHive.init <- function
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67262366
--- Diff: R/pkg/R/sparkR.R ---
@@ -31,20 +31,27 @@ connExists <- function(env) {
#' Stop the Spark context.
#'
#' Also term
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/12836
@NarineK Thanks again for the updates to this PR and thanks @sun-rui for
reviewing. The code changes LGTM -- the refactoring of worker.R is especially
useful for readability.
I just had a
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/12836#discussion_r67261006
--- Diff: R/pkg/R/DataFrame.R ---
@@ -1266,6 +1266,83 @@ setMethod("dapplyCollect",
ldf
})
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/12836#discussion_r67260862
--- Diff: R/pkg/R/DataFrame.R ---
@@ -1266,6 +1266,83 @@ setMethod("dapplyCollect",
ldf
})
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13684
Thanks @dongjoon-hyun -- Also would be good if @sun-rui took a look
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13684#discussion_r67234728
--- Diff: R/pkg/R/generics.R ---
@@ -462,12 +462,9 @@ setGeneric("describe", function(x, col, ...) {
standardGeneric("describe"
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13394
Yeah I think the approach used by @vectorijk is fine. We could have the
title as `Model Predictions` instead of `predict` (this is what R uses when you
do `?predict`)
---
If your project is set
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13684#discussion_r67233674
--- Diff: R/pkg/R/DataFrame.R ---
@@ -1869,14 +1871,23 @@ setMethod("where",
#' path <- "path/to/file.json"
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13684#discussion_r67208482
--- Diff: R/pkg/R/DataFrame.R ---
@@ -1859,7 +1859,7 @@ setMethod("where",
#' @param colnames A character vector of column names.
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13684#discussion_r67208357
--- Diff: R/pkg/R/DataFrame.R ---
@@ -1869,6 +1869,7 @@ setMethod("where",
#' path <- "path/to/file.json"
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13684#discussion_r67205381
--- Diff: R/pkg/R/DataFrame.R ---
@@ -1859,7 +1859,7 @@ setMethod("where",
#' @param colnames A character vector of column names.
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67204831
--- Diff: R/pkg/R/sparkR.R ---
@@ -31,20 +31,27 @@ connExists <- function(env) {
#' Stop the Spark context.
#'
#' Also term
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13636
LGTM. Thanks all. I will merge this after Jenkins passes
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13635#discussion_r67202631
--- Diff: R/pkg/NAMESPACE ---
@@ -6,10 +6,15 @@ importFrom(methods, setGeneric, setMethod, setOldClass)
#useDynLib(SparkR, stringHashCode
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13636#discussion_r67025422
--- Diff: R/pkg/inst/tests/testthat/test_context.R ---
@@ -19,21 +19,25 @@ context("test functions in sparkR.R")
test_that(&qu
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13660
Thanks @vectorijk - I left some comments inline.
cc @felixcheung
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13660#discussion_r67018458
--- Diff: docs/sparkr.md ---
@@ -262,6 +262,67 @@ head(df)
{% endhighlight %}
+### Applying User-defined Function
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13660#discussion_r67018219
--- Diff: docs/sparkr.md ---
@@ -262,6 +262,67 @@ head(df)
{% endhighlight %}
+### Applying User-defined Function
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13660#discussion_r67018163
--- Diff: docs/sparkr.md ---
@@ -262,6 +262,67 @@ head(df)
{% endhighlight %}
+### Applying User-defined Function
+
--- End
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13636
Thanks @felixcheung - @JoshRosen / @liancheng can you also test this PR
with R 3.3.0 before we merge ?
---
If your project is set up for it, you can reply to this email and have your
reply appear
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13636#discussion_r67009490
--- Diff: R/pkg/inst/tests/testthat/test_context.R ---
@@ -19,21 +19,25 @@ context("test functions in sparkR.R")
test_that(&qu
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13636
@liancheng we can use this PR to also address
https://issues.apache.org/jira/browse/SPARK-15931
---
If your project is set up for it, you can reply to this email and have your
reply appear on
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13644
Merging this to master and branch-2.0
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13644
Otherwise code change LGTM
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13644
Thanks @liancheng - there is a style error because line length exceeded 100
chars.
Also we'll need to note this breaking change in the programming guide
(http://spark.apache.org
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13635
Thanks @felixcheung - I'll take a look at this today.
cc @rxin
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13636
LGTM. Just to confirm your local tests pass with R version > 3.2 ?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your proj
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/12836
@rxin I think in this case we need access to grouping expression and
DataFrame from within the RelationalGroupedDataset class. One solution could be
to move the function `flatMapGroupsInR` to the
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/12836#discussion_r66713462
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/RelationalGroupedDataset.scala ---
@@ -381,6 +385,50 @@ class RelationalGroupedDataset protected[sql
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13610
Thanks @NarineK - Change looks pretty good to me. Minor comment: Can you
update the title to have `[SPARKR][SQL][SPARK-###]` at the beginning ?
---
If your project is set up for it, you can
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/12836#discussion_r66673292
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/object.scala
---
@@ -286,6 +290,9 @@ case class FlatMapGroupsInR
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/12836#discussion_r66671272
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/object.scala
---
@@ -286,6 +290,9 @@ case class FlatMapGroupsInR
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13508
Thanks @wangmiao1981 - LGTM. Merging this to master and branch-2.0
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/12836
Thanks @liancheng for clarification and @NarineK for implementing the
override. I just had one minor comment.
@sun-rui Can you take one final look ? Since we have not still cut RC1, we
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/12836#discussion_r9908
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/object.scala
---
@@ -286,6 +290,9 @@ case class FlatMapGroupsInR
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13394#discussion_r66363113
--- Diff: R/pkg/R/mllib.R ---
@@ -197,11 +197,10 @@ print.summary.GeneralizedLinearRegressionModel <-
function(x, ...) {
invisibl
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/13394#discussion_r66299402
--- Diff: R/pkg/R/mllib.R ---
@@ -197,11 +197,10 @@ print.summary.GeneralizedLinearRegressionModel <-
function(x, ...) {
invisibl
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/12836
I think I found the commit which causes this problem -
https://github.com/apache/spark/commit/6dde27404cb3d921d75dd6afca4b383f9df5976a
added toString to include arrays and the output we get is
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/13476
Merging this to master and branch-2.0
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user shivaram commented on the issue:
https://github.com/apache/spark/pull/12836
I don't know what could cause this - Do we have the beginning of the string
? My guess is `MapPartitions` or one of the nodes in the plan is calling
`toString` on a byte Array that contains s
801 - 900 of 2530 matches
Mail list logo