Skip to content
Closed
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -101,7 +101,8 @@ class GroupedDataset[K, V] private[sql](
*
* This function does not support partial aggregation, and as a result requires shuffling all
* the data in the [[Dataset]]. If an application intends to perform an aggregation over each
* key, it is best to use the reduce function or an [[Aggregator]].
* key, it is best to use the reduce function or an
* [[org.apache.spark.sql.expressions#Aggregator Aggregator]].
*
* Internally, the implementation will spill to disk if any given group is too large to fit into
* memory. However, users must take care to avoid materializing the whole iterator for a group
Expand All @@ -128,7 +129,8 @@ class GroupedDataset[K, V] private[sql](
*
* This function does not support partial aggregation, and as a result requires shuffling all
* the data in the [[Dataset]]. If an application intends to perform an aggregation over each
* key, it is best to use the reduce function or an [[Aggregator]].
* key, it is best to use the reduce function or an
* [[org.apache.spark.sql.expressions#Aggregator Aggregator]].
*
* Internally, the implementation will spill to disk if any given group is too large to fit into
* memory. However, users must take care to avoid materializing the whole iterator for a group
Expand All @@ -148,7 +150,8 @@ class GroupedDataset[K, V] private[sql](
*
* This function does not support partial aggregation, and as a result requires shuffling all
* the data in the [[Dataset]]. If an application intends to perform an aggregation over each
* key, it is best to use the reduce function or an [[Aggregator]].
* key, it is best to use the reduce function or an
* [[org.apache.spark.sql.expressions#Aggregator Aggregator]].
*
* Internally, the implementation will spill to disk if any given group is too large to fit into
* memory. However, users must take care to avoid materializing the whole iterator for a group
Expand All @@ -169,7 +172,8 @@ class GroupedDataset[K, V] private[sql](
*
* This function does not support partial aggregation, and as a result requires shuffling all
* the data in the [[Dataset]]. If an application intends to perform an aggregation over each
* key, it is best to use the reduce function or an [[Aggregator]].
* key, it is best to use the reduce function or an
* [[org.apache.spark.sql.expressions#Aggregator Aggregator]].
*
* Internally, the implementation will spill to disk if any given group is too large to fit into
* memory. However, users must take care to avoid materializing the whole iterator for a group
Expand Down