Skip to content

Commit

Permalink
[SPARK-13274] Fix Aggregator Links on GroupedDataset Scala API
Browse files Browse the repository at this point in the history
Update Aggregator links to point to #org.apache.spark.sql.expressions.Aggregator

Author: raela <raela@databricks.com>

Closes #11158 from raelawang/master.

(cherry picked from commit 719973b)
Signed-off-by: Reynold Xin <rxin@databricks.com>
  • Loading branch information
raelawang authored and rxin committed Feb 11, 2016
1 parent 93f1d91 commit b57fac5
Showing 1 changed file with 8 additions and 4 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -101,7 +101,8 @@ class GroupedDataset[K, V] private[sql](
*
* This function does not support partial aggregation, and as a result requires shuffling all
* the data in the [[Dataset]]. If an application intends to perform an aggregation over each
* key, it is best to use the reduce function or an [[Aggregator]].
* key, it is best to use the reduce function or an
* [[org.apache.spark.sql.expressions#Aggregator Aggregator]].
*
* Internally, the implementation will spill to disk if any given group is too large to fit into
* memory. However, users must take care to avoid materializing the whole iterator for a group
Expand Down Expand Up @@ -129,7 +130,8 @@ class GroupedDataset[K, V] private[sql](
*
* This function does not support partial aggregation, and as a result requires shuffling all
* the data in the [[Dataset]]. If an application intends to perform an aggregation over each
* key, it is best to use the reduce function or an [[Aggregator]].
* key, it is best to use the reduce function or an
* [[org.apache.spark.sql.expressions#Aggregator Aggregator]].
*
* Internally, the implementation will spill to disk if any given group is too large to fit into
* memory. However, users must take care to avoid materializing the whole iterator for a group
Expand All @@ -149,7 +151,8 @@ class GroupedDataset[K, V] private[sql](
*
* This function does not support partial aggregation, and as a result requires shuffling all
* the data in the [[Dataset]]. If an application intends to perform an aggregation over each
* key, it is best to use the reduce function or an [[Aggregator]].
* key, it is best to use the reduce function or an
* [[org.apache.spark.sql.expressions#Aggregator Aggregator]].
*
* Internally, the implementation will spill to disk if any given group is too large to fit into
* memory. However, users must take care to avoid materializing the whole iterator for a group
Expand All @@ -170,7 +173,8 @@ class GroupedDataset[K, V] private[sql](
*
* This function does not support partial aggregation, and as a result requires shuffling all
* the data in the [[Dataset]]. If an application intends to perform an aggregation over each
* key, it is best to use the reduce function or an [[Aggregator]].
* key, it is best to use the reduce function or an
* [[org.apache.spark.sql.expressions#Aggregator Aggregator]].
*
* Internally, the implementation will spill to disk if any given group is too large to fit into
* memory. However, users must take care to avoid materializing the whole iterator for a group
Expand Down

0 comments on commit b57fac5

Please sign in to comment.