aironman@MacBook-Pro-Retina-de-Alonso ~> spark-shell --packages org.diana-hep:spark-root_2.11:0.1.15 Ivy Default Cache set to: /Users/aironman/.ivy2/cache The jars for the packages stored in: /Users/aironman/.ivy2/jars :: loading settings :: url = jar:file:/usr/local/Cellar/apache-spark/2.4.1/libexec/jars/ivy-2.4.0.jar!/org/apache/ivy/core/settings/ivysettings.xml org.diana-hep#spark-root_2.11 added as a dependency :: resolving dependencies :: org.apache.spark#spark-submit-parent-dbc23b11-0a90-47f2-8619-67431f399658;1.0 confs: [default] found org.diana-hep#spark-root_2.11;0.1.15 in central found org.diana-hep#root4j;0.1.6 in central found org.apache.bcel#bcel;5.2 in central found jakarta-regexp#jakarta-regexp;1.4 in spark-list found org.tukaani#xz;1.2 in local-m2-cache found org.apache.logging.log4j#log4j;2.8 in central downloading https://repo1.maven.org/maven2/org/diana-hep/spark-root_2.11/0.1.15/spark-root_2.11-0.1.15.jar ... [SUCCESSFUL ] org.diana-hep#spark-root_2.11;0.1.15!spark-root_2.11.jar (105ms) :: resolution report :: resolve 2169ms :: artifacts dl 112ms :: modules in use: jakarta-regexp#jakarta-regexp;1.4 from spark-list in [default] org.apache.bcel#bcel;5.2 from central in [default] org.apache.logging.log4j#log4j;2.8 from central in [default] org.diana-hep#root4j;0.1.6 from central in [default] org.diana-hep#spark-root_2.11;0.1.15 from central in [default] org.tukaani#xz;1.2 from local-m2-cache in [default] --------------------------------------------------------------------- | | modules || artifacts | | conf | number| search|dwnlded|evicted|| number|dwnlded| --------------------------------------------------------------------- | default | 6 | 1 | 1 | 0 || 5 | 1 | --------------------------------------------------------------------- :: retrieving :: org.apache.spark#spark-submit-parent-dbc23b11-0a90-47f2-8619-67431f399658 confs: [default] 1 artifacts copied, 4 already retrieved (770kB/10ms) 19/05/07 13:53:31 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties Setting default log level to "WARN". To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel). Spark context Web UI available at http://192.168.1.36:4040 Spark context available as 'sc' (master = local[*], app id = local-1557230018873). Spark session available as 'spark'. Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 2.4.1 /_/ Using Scala version 2.11.12 (Java HotSpot(TM) 64-Bit Server VM, Java 1.8.0_172) Type in expressions to have them evaluated. Type :help for more information. scala> import org.dianahep.sparkroot.experimental._ import org.dianahep.sparkroot.experimental._ scala> val dfGamma = spark.read.root("/Users/aironman/Downloads/complete_set_of_ATLAS_open_data_samples_July_2016/Data/DataEgamma.root") java.lang.AbstractMethodError: org.dianahep.sparkroot.experimental.DefaultSource.supportDataType(Lorg/apache/spark/sql/types/DataType;Z)Z at org.apache.spark.sql.execution.datasources.DataSourceUtils$$anonfun$verifySchema$1.apply(DataSourceUtils.scala:48) at org.apache.spark.sql.execution.datasources.DataSourceUtils$$anonfun$verifySchema$1.apply(DataSourceUtils.scala:47) at scala.collection.Iterator$class.foreach(Iterator.scala:891) at scala.collection.AbstractIterator.foreach(Iterator.scala:1334) at scala.collection.IterableLike$class.foreach(IterableLike.scala:72) at org.apache.spark.sql.types.StructType.foreach(StructType.scala:99) at org.apache.spark.sql.execution.datasources.DataSourceUtils$.verifySchema(DataSourceUtils.scala:47) at org.apache.spark.sql.execution.datasources.DataSourceUtils$.verifyReadSchema(DataSourceUtils.scala:39) at org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:400) at org.apache.spark.sql.DataFrameReader.loadV1Source(DataFrameReader.scala:223) at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:211) at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:178) at org.dianahep.sparkroot.experimental.package$RootDataFrameReader.root(package.scala:44) ... 51 elided scala> val dfMuons = spark.read.root("/Users/aironman/Downloads/complete_set_of_ATLAS_open_data_samples_July_2016/Data") java.lang.AbstractMethodError: org.dianahep.sparkroot.experimental.DefaultSource.supportDataType(Lorg/apache/spark/sql/types/DataType;Z)Z at org.apache.spark.sql.execution.datasources.DataSourceUtils$$anonfun$verifySchema$1.apply(DataSourceUtils.scala:48) at org.apache.spark.sql.execution.datasources.DataSourceUtils$$anonfun$verifySchema$1.apply(DataSourceUtils.scala:47) at scala.collection.Iterator$class.foreach(Iterator.scala:891) at scala.collection.AbstractIterator.foreach(Iterator.scala:1334) at scala.collection.IterableLike$class.foreach(IterableLike.scala:72) at org.apache.spark.sql.types.StructType.foreach(StructType.scala:99) at org.apache.spark.sql.execution.datasources.DataSourceUtils$.verifySchema(DataSourceUtils.scala:47) at org.apache.spark.sql.execution.datasources.DataSourceUtils$.verifyReadSchema(DataSourceUtils.scala:39) at org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:400) at org.apache.spark.sql.DataFrameReader.loadV1Source(DataFrameReader.scala:223) at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:211) at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:178) at org.dianahep.sparkroot.experimental.package$RootDataFrameReader.root(package.scala:44) ... 51 elided