From b17b9a32f4325a86ed05fe92b25606c01ad10eac Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Du=C5=A1an=20Ti=C5=A1ma?= Date: Mon, 8 Sep 2025 19:33:06 +0200 Subject: [PATCH] enable scripting by default --- .../org/apache/spark/sql/internal/SQLConf.scala | 2 +- .../parser/SqlScriptingParserSuite.scala | 12 ------------ .../ExecuteImmediateEndToEndSuite.scala | 17 +++++++---------- .../sql/scripting/SqlScriptingE2eSuite.scala | 1 - .../scripting/SqlScriptingExecutionSuite.scala | 1 - .../SqlScriptingInterpreterSuite.scala | 7 +------ 6 files changed, 9 insertions(+), 31 deletions(-) diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala index 12fd80c5a6269..5a99814c8cc85 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala @@ -4142,7 +4142,7 @@ object SQLConf { "flow and error handling.") .version("4.0.0") .booleanConf - .createWithDefault(false) + .createWithDefault(true) val CONCAT_BINARY_AS_STRING = buildConf("spark.sql.function.concatBinaryAsString") .doc("When this option is set to false and all inputs are binary, `functions.concat` returns " + diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/SqlScriptingParserSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/SqlScriptingParserSuite.scala index abcea96f0831f..9902374ce8e9b 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/SqlScriptingParserSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/SqlScriptingParserSuite.scala @@ -23,22 +23,10 @@ import org.apache.spark.sql.catalyst.plans.SQLHelper import org.apache.spark.sql.catalyst.plans.logical.{CompoundBody, CreateVariable, ExceptionHandler, ForStatement, IfElseStatement, IterateStatement, LeaveStatement, LoopStatement, Project, RepeatStatement, SearchedCaseStatement, SetVariable, SimpleCaseStatement, SingleStatement, WhileStatement} import org.apache.spark.sql.errors.DataTypeErrors.toSQLId import org.apache.spark.sql.exceptions.SqlScriptingException -import org.apache.spark.sql.internal.SQLConf class SqlScriptingParserSuite extends SparkFunSuite with SQLHelper { import CatalystSqlParser._ - // Tests setup - protected override def beforeAll(): Unit = { - super.beforeAll() - conf.setConfString(SQLConf.SQL_SCRIPTING_ENABLED.key, "true") - } - - protected override def afterAll(): Unit = { - conf.unsetConf(SQLConf.SQL_SCRIPTING_ENABLED.key) - super.afterAll() - } - // Tests test("single select") { val sqlScriptText = "SELECT 1;" diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/ExecuteImmediateEndToEndSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/ExecuteImmediateEndToEndSuite.scala index d8c24c1ac3970..c252047b3abe5 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/ExecuteImmediateEndToEndSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/ExecuteImmediateEndToEndSuite.scala @@ -17,7 +17,6 @@ package org.apache.spark.sql.execution import org.apache.spark.sql.{AnalysisException, QueryTest} -import org.apache.spark.sql.internal.SQLConf import org.apache.spark.sql.test.SharedSparkSession class ExecuteImmediateEndToEndSuite extends QueryTest with SharedSparkSession { @@ -39,14 +38,12 @@ class ExecuteImmediateEndToEndSuite extends QueryTest with SharedSparkSession { } test("SQL Scripting not supported inside EXECUTE IMMEDIATE") { - withSQLConf(SQLConf.SQL_SCRIPTING_ENABLED.key -> "true") { - val executeImmediateText = "EXECUTE IMMEDIATE 'BEGIN SELECT 1; END'" - checkError( - exception = intercept[AnalysisException ] { - spark.sql(executeImmediateText) - }, - condition = "SQL_SCRIPT_IN_EXECUTE_IMMEDIATE", - parameters = Map("sqlString" -> "BEGIN SELECT 1; END")) - } + val executeImmediateText = "EXECUTE IMMEDIATE 'BEGIN SELECT 1; END'" + checkError( + exception = intercept[AnalysisException ] { + spark.sql(executeImmediateText) + }, + condition = "SQL_SCRIPT_IN_EXECUTE_IMMEDIATE", + parameters = Map("sqlString" -> "BEGIN SELECT 1; END")) } } diff --git a/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingE2eSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingE2eSuite.scala index 67305019e70d6..e24407912eb02 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingE2eSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingE2eSuite.scala @@ -58,7 +58,6 @@ class SqlScriptingE2eSuite extends QueryTest with SharedSparkSession { override protected def sparkConf: SparkConf = { super.sparkConf .set(SQLConf.ANSI_ENABLED.key, "true") - .set(SQLConf.SQL_SCRIPTING_ENABLED.key, "true") } // Tests diff --git a/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingExecutionSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingExecutionSuite.scala index 0be9e5f0f578b..7e6de2b990ffd 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingExecutionSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingExecutionSuite.scala @@ -40,7 +40,6 @@ class SqlScriptingExecutionSuite extends QueryTest with SharedSparkSession { override protected def sparkConf: SparkConf = { super.sparkConf .set(SQLConf.ANSI_ENABLED.key, "true") - .set(SQLConf.SQL_SCRIPTING_ENABLED.key, "true") } // Helpers diff --git a/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingInterpreterSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingInterpreterSuite.scala index 0eafb1fbce2cc..684a5a72e6d83 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingInterpreterSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingInterpreterSuite.scala @@ -17,7 +17,7 @@ package org.apache.spark.sql.scripting -import org.apache.spark.{SparkConf, SparkException, SparkNumberFormatException} +import org.apache.spark.{SparkException, SparkNumberFormatException} import org.apache.spark.sql.{AnalysisException, QueryTest, Row} import org.apache.spark.sql.catalyst.QueryPlanningTracker import org.apache.spark.sql.catalyst.expressions.Expression @@ -38,11 +38,6 @@ class SqlScriptingInterpreterSuite with SharedSparkSession with SqlScriptingTestUtils { - // Tests setup - override protected def sparkConf: SparkConf = { - super.sparkConf.set(SQLConf.SQL_SCRIPTING_ENABLED.key, "true") - } - // Helpers private def runSqlScript( sqlText: String,