From a4a2d822c5c3a3172ba9636af56524e5c2ca98da Mon Sep 17 00:00:00 2001 From: Chris Snow Date: Wed, 11 Nov 2015 22:43:20 +0000 Subject: [PATCH] [SPARK-11658] simplify documentation for PySpark combineByKey --- python/pyspark/rdd.py | 1 - 1 file changed, 1 deletion(-) diff --git a/python/pyspark/rdd.py b/python/pyspark/rdd.py index 56e892243c79c..4b4d59647b2bc 100644 --- a/python/pyspark/rdd.py +++ b/python/pyspark/rdd.py @@ -1760,7 +1760,6 @@ def combineByKey(self, createCombiner, mergeValue, mergeCombiners, In addition, users can control the partitioning of the output RDD. >>> x = sc.parallelize([("a", 1), ("b", 1), ("a", 1)]) - >>> def f(x): return x >>> def add(a, b): return a + str(b) >>> sorted(x.combineByKey(str, add, add).collect()) [('a', '11'), ('b', '1')]