scala> import org.apache.spark.sql.types._
import org.apache.spark.sql.types._
scala> val inputs = Seq("TR", "FA").toDF("n")
inputs: org.apache.spark.sql.DataFrame = [n: string]
scala> inputs.write.parquet("test.parquet")
24/02/13 07:58:44 INFO src/lib.rs: Comet native library initialized
scala> val df = spark.read.parquet("test.parquet")
df: org.apache.spark.sql.DataFrame = [n: string]
scala> val df2 = df.withColumn("converted", col("n").cast(DataTypes.BooleanType))
df2: org.apache.spark.sql.DataFrame = [n: string, converted: boolean]
scala> df2.show
+---+---------+
| n|converted|
+---+---------+
| FA| false|
| TR| true|
+---+---------+
scala> spark.conf.set("spark.comet.enabled", false)
scala> df2.show
+---+---------+
| n|converted|
+---+---------+
| FA| null|
| TR| null|
+---+---------+