HyukjinKwon commented on a change in pull request #28350:
URL: https://github.com/apache/spark/pull/28350#discussion_r415489665
##########
File path: R/pkg/R/DataFrame.R
##########
@@ -1669,29 +1667,43 @@ setMethod("dapplyCollect",
#' @aliases gapply,SparkDataFrame-method
#' @rdname gapply
#' @name gapply
+#' @details
+#' \code{func} is a function of two arguments. The first, usually named
\code{key}
+#' (though this is not enforced) corresponds to the grouping key, will be an
+#' unnamed \code{list} of \code{length(cols)} length-one objects corresponding
+#' to the grouping columns' values for the current group.
+#'
+#' The second, herein \code{x}, will be a local \code{\link{data.frame}} with
the
+#' columns of the input not in \code{cols} for the rows corresponding to
\code{key}.
+#'
+#' The output of \code{func} must be a \code{data.frame} matching
\code{schema} --
+#' in particular this means the names of the output \code{data.frame} are
irrelevant
Review comment:
Can be .. any good clean and consistent way is fine as long as we can
clarify the differences - `gapplyCollect` is a short cut of `gapply` and
`collect`; `gapplyCollect` can omit the return schema, and directly collects to
the driver side. Except this, all are same.
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]