Skip to content

Commit

Permalink
spark 1.2
Browse files Browse the repository at this point in the history
  • Loading branch information
sorenmacbeth committed Dec 21, 2014
1 parent dece563 commit a47458a
Show file tree
Hide file tree
Showing 3 changed files with 7 additions and 10 deletions.
1 change: 0 additions & 1 deletion dev-resources/log4j.properties
Original file line number Diff line number Diff line change
Expand Up @@ -13,4 +13,3 @@ log4j.logger.org.apache.spark.repl.SparkILoop$SparkILoopInterpreter=INFO
# uncomment for tracing of fns
# log4j.logger.flambo=trace
# log4j.logger.serializable.fn=trace
log4j.logger.flambo=trace
12 changes: 5 additions & 7 deletions project.clj
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,6 @@
:mailing-list {:name "flambo user mailing list"
:archive "https://groups.google.com/d/forum/flambo-user"
:post "flambo-user@googlegroups.com"}
:repositories [["spark-staging" {:url "https://repository.apache.org/content/repositories/orgapachespark-1055/"
:checksum :ignore}]]
:dependencies [[org.clojure/clojure "1.6.0"]
[org.clojure/tools.logging "0.2.6"]
[yieldbot/serializable-fn "0.0.6"
Expand All @@ -27,11 +25,11 @@
flambo.example.tfidf]}
:provided
{:dependencies
[[org.apache.spark/spark-core_2.10 "1.1.1"]
[org.apache.spark/spark-streaming_2.10 "1.1.1"]
[org.apache.spark/spark-streaming-kafka_2.10 "1.1.1"]
[org.apache.spark/spark-streaming-flume_2.10 "1.1.1"]
[org.apache.spark/spark-sql_2.10 "1.1.1"]]}
[[org.apache.spark/spark-core_2.10 "1.2.0"]
[org.apache.spark/spark-streaming_2.10 "1.2.0"]
[org.apache.spark/spark-streaming-kafka_2.10 "1.2.0"]
[org.apache.spark/spark-streaming-flume_2.10 "1.2.0"]
[org.apache.spark/spark-sql_2.10 "1.2.0"]]}
:clojure-1.7
{:dependencies [[org.clojure/clojure "1.7.0-alpha4"]]}
:uberjar
Expand Down
4 changes: 2 additions & 2 deletions test/flambo/api_test.clj
Original file line number Diff line number Diff line change
Expand Up @@ -159,9 +159,9 @@
"sample returns a fraction of the RDD, with/without replacement,
using a given random number generator seed"
(-> (f/parallelize c [0 1 2 3 4 5 6 7 8 9])
(f/sample false 0.1 2)
(f/sample false 0.2 666)
f/collect
vec) => [6])
vec) => [1])

(fact
"combine-by-key returns an RDD by combining the elements for each key using a custom
Expand Down

0 comments on commit a47458a

Please sign in to comment.