Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Method not found when creating time series add #53

Closed
geoHeil opened this issue Oct 13, 2018 · 3 comments
Closed

Method not found when creating time series add #53

geoHeil opened this issue Oct 13, 2018 · 3 comments

Comments

@geoHeil
Copy link

geoHeil commented Oct 13, 2018

val tsRdd = TimeSeriesRDD.fromDF(dataFrame = df)(isSorted = true, timeUnit = MILLISECONDS)

throws

scala> val tsRdd = TimeSeriesRDD.fromDF(dataFrame = cellFeed)(isSorted = true, timeUnit = MILLISECONDS)
java.lang.NoSuchMethodError: org.apache.spark.sql.catalyst.plans.physical.ClusteredDistribution$.apply$default$2()Lscala/Option;
  at com.twosigma.flint.timeseries.TimeSeriesStore$.isClustered(TimeSeriesStore.scala:149)
  at com.twosigma.flint.timeseries.TimeSeriesStore$.apply(TimeSeriesStore.scala:64)
  at com.twosigma.flint.timeseries.TimeSeriesRDD$.fromDFWithPartInfo(TimeSeriesRDD.scala:509)
  at com.twosigma.flint.timeseries.TimeSeriesRDD$.fromDF(TimeSeriesRDD.scala:304)
  ... 52 elided

on spark 2.2 when trying to create the initial RDD.

Minimal reproducible sample:

import spark.implicits._
  import com.twosigma.flint.timeseries.TimeSeriesRDD
  import scala.concurrent.duration._
  val df = Seq((1, 1, 1L), (2, 3, 1L), (1, 4, 2L), (2, 2, 2L)).toDF("id", "value", "time")
  val tsRdd = TimeSeriesRDD.fromDF(dataFrame = df)(isSorted = true, timeUnit = MILLISECONDS)

on spark 2.2 via HDP 2.6.4

@geoHeil
Copy link
Author

geoHeil commented Oct 14, 2018

Regular spark 2.3.2 works fine, regular spark 2.2.1 also fails with above exception.

@geoHeil
Copy link
Author

geoHeil commented Oct 14, 2018

When creating a custom build like https://github.com/geoHeil/flint/tree/flint-spark-2.2 the basic example works again with spark 2.2

However, there are 2 unit test failures

  • CSV reading (some timestamps did not match), maybe due to different locale
 - should correctly convert SQL TimestampType with default format *** FAILED *** (237 milliseconds)
[info]   1199232000000000000 did not equal 1199228400000000000 (CSVSpec.scala:100)

NOTE: csv test failure is also is on regular master branch

  • partition preserving something failed because a list was empty and head was called

@icexelloss
Copy link
Member

icexelloss commented Oct 14, 2018 via email

@geoHeil geoHeil closed this as completed Oct 20, 2018
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants