Skip to content

Commit

Permalink
Tweak the test.
Browse files Browse the repository at this point in the history
  • Loading branch information
viirya committed Sep 9, 2017
1 parent 2cf6bbb commit fbf8a88
Show file tree
Hide file tree
Showing 2 changed files with 14 additions and 10 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ import java.util.Calendar

import org.apache.spark.SparkFunSuite
import org.apache.spark.sql.catalyst.InternalRow
import org.apache.spark.sql.catalyst.analysis.UnresolvedException
import org.apache.spark.sql.catalyst.errors.TreeNodeException
import org.apache.spark.sql.catalyst.util.{DateTimeTestUtils, DateTimeUtils, GenericArrayData, PermissiveMode}
import org.apache.spark.sql.types._
import org.apache.spark.unsafe.types.UTF8String
Expand Down Expand Up @@ -626,7 +626,7 @@ class JsonExpressionsSuite extends SparkFunSuite with ExpressionEvalHelper {
val mapType2 = MapType(IntegerType, CalendarIntervalType)
val schema2 = StructType(StructField("a", mapType2) :: Nil)
val struct2 = Literal.create(null, schema2)
intercept[UnresolvedException[_]] {
intercept[TreeNodeException[_]] {
checkEvaluation(
StructsToJson(Map.empty, struct2, gmtId),
null
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -195,6 +195,16 @@ class JsonFunctionsSuite extends QueryTest with SharedSQLContext {
Row("""{"_1":"26/08/2015 18:00"}""") :: Nil)
}

test("to_json - key types of map don't matter") {
// interval type is invalid for converting to JSON. However, the keys of a map are treated
// as strings, so its type doesn't matter.
val df = Seq(Tuple1(Tuple1("interval -3 month 7 hours"))).toDF("a")
.select(struct(map($"a._1".cast(CalendarIntervalType), lit("a")).as("col1")).as("c"))
checkAnswer(
df.select(to_json($"c")),
Row("""{"col1":{"interval -3 months 7 hours":"a"}}""") :: Nil)
}

test("to_json unsupported type") {
val baseDf = Seq(Tuple1(Tuple1("interval -3 month 7 hours"))).toDF("a")
val df = baseDf.select(struct($"a._1".cast(CalendarIntervalType).as("a")).as("c"))
Expand All @@ -205,17 +215,11 @@ class JsonFunctionsSuite extends QueryTest with SharedSQLContext {
assert(e.getMessage.contains(
"Unable to convert column a of type calendarinterval to JSON."))

// interval type is invalid for converting to JSON. However, the keys of a map are treated
// as strings, so its type doesn't matter.
// interval type is invalid for converting to JSON. We can't use it as value type of a map.
val df2 = baseDf
.select(struct(map($"a._1".cast(CalendarIntervalType), lit("a")).as("col1")).as("c"))
val df3 = baseDf
.select(struct(map(lit("a"), $"a._1".cast(CalendarIntervalType)).as("col1")).as("c"))
checkAnswer(
df2.select(to_json($"c")),
Row("""{"col1":{"interval -3 months 7 hours":"a"}}""") :: Nil)
val e2 = intercept[AnalysisException] {
df3.select(to_json($"c")).collect()
df2.select(to_json($"c")).collect()
}
assert(e2.getMessage.contains("Unable to convert column col1 of type calendarinterval to JSON"))
}
Expand Down

0 comments on commit fbf8a88

Please sign in to comment.