diff --git a/regression-test/suites/correctness_p0/table_valued_function/test_hdfs_tvf.groovy b/regression-test/suites/correctness_p0/table_valued_function/test_hdfs_tvf.groovy index 65996cca81ffc18..d89dcadabf115c4 100644 --- a/regression-test/suites/correctness_p0/table_valued_function/test_hdfs_tvf.groovy +++ b/regression-test/suites/correctness_p0/table_valued_function/test_hdfs_tvf.groovy @@ -193,7 +193,7 @@ suite("test_hdfs_tvf") { assertTrue(result2[0][0] == 5, "Insert should update 12 rows") qt_insert """ select * from test_hdfs_tvf order by id; """ } finally { - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" } } } diff --git a/regression-test/suites/external_catalog_p0/hive/test_hive_orc.groovy b/regression-test/suites/external_catalog_p0/hive/test_hive_orc.groovy index 7e1f8b78fde8bcf..90398e293a5cfbd 100644 --- a/regression-test/suites/external_catalog_p0/hive/test_hive_orc.groovy +++ b/regression-test/suites/external_catalog_p0/hive/test_hive_orc.groovy @@ -66,29 +66,6 @@ suite("test_hive_orc", "all_types") { qt_only_partition_col """select count(p1_col), count(p2_col) from orc_all_types;""" } - def set_be_config = { flag -> - String[][] backends = sql """ show backends; """ - assertTrue(backends.size() > 0) - for (String[] backend in backends) { - StringBuilder setConfigCommand = new StringBuilder(); - setConfigCommand.append("curl -X POST http://") - setConfigCommand.append(backend[2]) - setConfigCommand.append(":") - setConfigCommand.append(backend[5]) - setConfigCommand.append("/api/update_config?") - String command1 = setConfigCommand.toString() + "enable_new_load_scan_node=$flag" - logger.info(command1) - String command2 = setConfigCommand.toString() + "enable_new_file_scanner=$flag" - logger.info(command2) - def process1 = command1.execute() - int code = process1.waitFor() - assertEquals(code, 0) - def process2 = command2.execute() - code = process1.waitFor() - assertEquals(code, 0) - } - } - String enabled = context.config.otherConfigs.get("enableHiveTest") if (enabled != null && enabled.equalsIgnoreCase("true")) { try { @@ -96,7 +73,6 @@ suite("test_hive_orc", "all_types") { String catalog_name = "hive_test_orc" sql """admin set frontend config ("enable_multi_catalog" = "true")""" sql """admin set frontend config ("enable_new_load_scan_node" = "true");""" - set_be_config.call('true') sql """drop catalog if exists ${catalog_name}""" sql """ create catalog if not exists ${catalog_name} properties ( @@ -114,8 +90,7 @@ suite("test_hive_orc", "all_types") { only_partition_col() } finally { - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""" - set_be_config.call('false') + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" } } } diff --git a/regression-test/suites/external_catalog_p0/hive/test_hive_other.groovy b/regression-test/suites/external_catalog_p0/hive/test_hive_other.groovy index 82b6bb1cfc8480c..0c207c3754d8b84 100644 --- a/regression-test/suites/external_catalog_p0/hive/test_hive_other.groovy +++ b/regression-test/suites/external_catalog_p0/hive/test_hive_other.groovy @@ -50,35 +50,10 @@ suite("test_hive_other", "p0") { } - def set_be_config = { -> - String[][] backends = sql """ show backends; """ - assertTrue(backends.size() > 0) - for (String[] backend in backends) { - // No need to set this config anymore, but leave this code sample here - // StringBuilder setConfigCommand = new StringBuilder(); - // setConfigCommand.append("curl -X POST http://") - // setConfigCommand.append(backend[2]) - // setConfigCommand.append(":") - // setConfigCommand.append(backend[5]) - // setConfigCommand.append("/api/update_config?") - // String command1 = setConfigCommand.toString() + "enable_new_load_scan_node=true" - // logger.info(command1) - // String command2 = setConfigCommand.toString() + "enable_new_file_scanner=true" - // logger.info(command2) - // def process1 = command1.execute() - // int code = process1.waitFor() - // assertEquals(code, 0) - // def process2 = command2.execute() - // code = process1.waitFor() - // assertEquals(code, 0) - } - } - String enabled = context.config.otherConfigs.get("enableHiveTest") if (enabled != null && enabled.equalsIgnoreCase("true")) { String hms_port = context.config.otherConfigs.get("hms_port") String catalog_name = "hive_test_other" - set_be_config.call() sql """admin set frontend config ("enable_multi_catalog" = "true")""" sql """drop catalog if exists ${catalog_name}""" diff --git a/regression-test/suites/external_catalog_p0/hive/test_hive_parquet.groovy b/regression-test/suites/external_catalog_p0/hive/test_hive_parquet.groovy index 8cb89baec5b22c2..bb2ad552a69e38f 100644 --- a/regression-test/suites/external_catalog_p0/hive/test_hive_parquet.groovy +++ b/regression-test/suites/external_catalog_p0/hive/test_hive_parquet.groovy @@ -139,30 +139,6 @@ suite("test_hive_parquet", "p0") { """ } - - def set_be_config = { flag -> - String[][] backends = sql """ show backends; """ - assertTrue(backends.size() > 0) - for (String[] backend in backends) { - StringBuilder setConfigCommand = new StringBuilder(); - setConfigCommand.append("curl -X POST http://") - setConfigCommand.append(backend[2]) - setConfigCommand.append(":") - setConfigCommand.append(backend[5]) - setConfigCommand.append("/api/update_config?") - String command1 = setConfigCommand.toString() + "enable_new_load_scan_node=$flag" - logger.info(command1) - String command2 = setConfigCommand.toString() + "enable_new_file_scanner=$flag" - logger.info(command2) - def process1 = command1.execute() - int code = process1.waitFor() - assertEquals(code, 0) - def process2 = command2.execute() - code = process1.waitFor() - assertEquals(code, 0) - } - } - String enabled = context.config.otherConfigs.get("enableHiveTest") if (enabled != null && enabled.equalsIgnoreCase("true")) { try { @@ -170,7 +146,6 @@ suite("test_hive_parquet", "p0") { String catalog_name = "hive_test_parquet" sql """admin set frontend config ("enable_multi_catalog" = "true")""" sql """admin set frontend config ("enable_new_load_scan_node" = "true");""" - set_be_config.call('true') sql """drop catalog if exists ${catalog_name}""" sql """ create catalog if not exists ${catalog_name} properties ( @@ -201,8 +176,7 @@ suite("test_hive_parquet", "p0") { q19() q20() } finally { - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""" - set_be_config.call('false') + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" } } } diff --git a/regression-test/suites/load_p0/broker_load/test_array_load.groovy b/regression-test/suites/load_p0/broker_load/test_array_load.groovy index 7f5d10995215f59..bbdae4659fa9f55 100644 --- a/regression-test/suites/load_p0/broker_load/test_array_load.groovy +++ b/regression-test/suites/load_p0/broker_load/test_array_load.groovy @@ -204,7 +204,7 @@ suite("test_array_load", "load_p0") { for ( i in 0..1 ) { // should be deleted after new_load_scan is ready if (i == 1) { - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" } else { sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" } @@ -280,7 +280,7 @@ suite("test_array_load", "load_p0") { } } } finally { - try_sql("""ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""") + try_sql("""ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""") } diff --git a/regression-test/suites/load_p0/broker_load/test_broker_load.groovy b/regression-test/suites/load_p0/broker_load/test_broker_load.groovy index e1f16676dd1efd2..e67397e821a5e5c 100644 --- a/regression-test/suites/load_p0/broker_load/test_broker_load.groovy +++ b/regression-test/suites/load_p0/broker_load/test_broker_load.groovy @@ -192,28 +192,8 @@ suite("test_broker_load", "p0") { logger.info("Submit load with lable: $uuid, table: $table, path: $path") } - def set_be_config = { flag-> - String[][] backends = sql """ show backends; """ - assertTrue(backends.size() > 0) - for (String[] backend in backends) { - // No need to set this config anymore, but leave this code sample here - // StringBuilder setConfigCommand = new StringBuilder(); - // setConfigCommand.append("curl -X POST http://") - // setConfigCommand.append(backend[2]) - // setConfigCommand.append(":") - // setConfigCommand.append(backend[5]) - // setConfigCommand.append("/api/update_config?") - // String command1 = setConfigCommand.toString() + "enable_new_load_scan_node=$flag" - // logger.info(command1) - // def process1 = command1.execute() - // int code = process1.waitFor() - // assertEquals(code, 0) - } - } - if (enabled != null && enabled.equalsIgnoreCase("true")) { def uuids = [] - set_be_config.call('true') try { def i = 0 for (String table in tables) { @@ -258,7 +238,6 @@ suite("test_broker_load", "p0") { order_qt_parquet_s3_case9 """ select * from parquet_s3_case9""" } finally { - set_be_config.call('false') for (String table in tables) { sql new File("""${context.file.parent}/ddl/${table}_drop.sql""").text } diff --git a/regression-test/suites/load_p0/stream_load/load_json_column_exclude_schema_without_jsonpath.groovy b/regression-test/suites/load_p0/stream_load/load_json_column_exclude_schema_without_jsonpath.groovy index 760af3344e7aead..a6ce6a16f262e7e 100644 --- a/regression-test/suites/load_p0/stream_load/load_json_column_exclude_schema_without_jsonpath.groovy +++ b/regression-test/suites/load_p0/stream_load/load_json_column_exclude_schema_without_jsonpath.groovy @@ -52,7 +52,7 @@ suite("test_load_json_column_exclude_schema_without_jsonpath", "p0") { def load_array_data = {new_json_reader_flag, table_name, strip_flag, read_flag, format_flag, exprs, json_paths, json_root, where_expr, fuzzy_flag, column_sep, file_name -> // should be deleted after new_load_scan is ready - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "${new_json_reader_flag}");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" // load the json data streamLoad { diff --git a/regression-test/suites/load_p0/stream_load/load_json_null_to_nullable.groovy b/regression-test/suites/load_p0/stream_load/load_json_null_to_nullable.groovy index f934c038a2aceaf..8e7caba45fb1daf 100644 --- a/regression-test/suites/load_p0/stream_load/load_json_null_to_nullable.groovy +++ b/regression-test/suites/load_p0/stream_load/load_json_null_to_nullable.groovy @@ -43,7 +43,7 @@ suite("test_load_json_null_to_nullable", "p0") { def load_array_data = {new_json_reader_flag, table_name, strip_flag, read_flag, format_flag, exprs, json_paths, json_root, where_expr, fuzzy_flag, column_sep, file_name -> // should be deleted after new_load_scan is ready - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "${new_json_reader_flag}");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" // load the json data streamLoad { @@ -79,7 +79,7 @@ suite("test_load_json_null_to_nullable", "p0") { } // should be deleted after new_load_scan is ready - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" } def check_data_correct = {table_name -> diff --git a/regression-test/suites/load_p0/stream_load/load_json_with_jsonpath.groovy b/regression-test/suites/load_p0/stream_load/load_json_with_jsonpath.groovy index 02ffd808e2ff7bf..1254356d41673ad 100644 --- a/regression-test/suites/load_p0/stream_load/load_json_with_jsonpath.groovy +++ b/regression-test/suites/load_p0/stream_load/load_json_with_jsonpath.groovy @@ -44,7 +44,7 @@ suite("test_load_json_with_jsonpath", "p0") { json_root, where_expr, fuzzy_flag, column_sep, file_name -> // should be deleted after new_load_scan is ready - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "${new_json_reader_flag}");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" // load the json data streamLoad { @@ -80,7 +80,7 @@ suite("test_load_json_with_jsonpath", "p0") { } // should be deleted after new_load_scan is ready - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" } def check_data_correct = {table_name -> diff --git a/regression-test/suites/load_p0/stream_load/test_hdfs_json_load.groovy b/regression-test/suites/load_p0/stream_load/test_hdfs_json_load.groovy index 5c79cf1b976f7cc..e1a09dff8d2fdca 100644 --- a/regression-test/suites/load_p0/stream_load/test_hdfs_json_load.groovy +++ b/regression-test/suites/load_p0/stream_load/test_hdfs_json_load.groovy @@ -45,7 +45,7 @@ suite("test_hdfs_json_load", "p0") { def load_from_hdfs1 = {new_json_reader_flag, strip_flag, fuzzy_flag, testTablex, label, fileName, fsPath, hdfsUser, exprs, jsonpaths, json_root, columns_parameter, where -> // should be delete after new_load_scan is ready - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "${new_json_reader_flag}");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" def hdfsFilePath = "${fsPath}/user/doris/preinstalled_data/json_format_test/${fileName}" def result1= sql """ @@ -78,7 +78,7 @@ suite("test_hdfs_json_load", "p0") { assertTrue(result1[0][0] == 0, "Query OK, 0 rows affected") // should be delete after new_load_scan is ready - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" } def check_load_result = {checklabel, testTablex -> diff --git a/regression-test/suites/load_p0/stream_load/test_json_load.groovy b/regression-test/suites/load_p0/stream_load/test_json_load.groovy index e066467e3e849d4..8b49883907770df 100644 --- a/regression-test/suites/load_p0/stream_load/test_json_load.groovy +++ b/regression-test/suites/load_p0/stream_load/test_json_load.groovy @@ -117,7 +117,7 @@ suite("test_json_load", "p0") { def load_json_data = {new_json_reader_flag, label, strip_flag, read_flag, format_flag, exprs, json_paths, json_root, where_expr, fuzzy_flag, file_name, ignore_failure=false -> // should be delete after new_load_scan is ready - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "${new_json_reader_flag}");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" // load the json data streamLoad { @@ -152,7 +152,7 @@ suite("test_json_load", "p0") { } // should be deleted after new_load_scan is ready - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" } def load_from_hdfs1 = {testTablex, label, hdfsFilePath, format, brokerName, hdfsUser, hdfsPasswd -> @@ -530,7 +530,7 @@ suite("test_json_load", "p0") { sql "DROP TABLE IF EXISTS ${testTable}" create_test_table3.call(testTable) // should be delete after new_load_scan is ready - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" // load the json data streamLoad { table "${testTable}" @@ -558,7 +558,7 @@ suite("test_json_load", "p0") { } } // should be deleted after new_load_scan is ready - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" sql "sync" qt_select13 "select * from ${testTable} order by id" @@ -594,7 +594,7 @@ suite("test_json_load", "p0") { } } // should be deleted after new_load_scan is ready - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" sql "sync" qt_select13 "select * from ${testTable} order by id" diff --git a/regression-test/suites/load_p0/stream_load/test_txt_special_delimiter.groovy b/regression-test/suites/load_p0/stream_load/test_txt_special_delimiter.groovy index fff343078b051c5..9d39eb2681e49ba 100644 --- a/regression-test/suites/load_p0/stream_load/test_txt_special_delimiter.groovy +++ b/regression-test/suites/load_p0/stream_load/test_txt_special_delimiter.groovy @@ -33,7 +33,7 @@ suite("test_txt_special_delimiter", "p0") { for ( i in 0..1 ) { // should be deleted after new_load_scan is ready if (i == 1) { - sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "false");""" + sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" } else { sql """ADMIN SET FRONTEND CONFIG ("enable_new_load_scan_node" = "true");""" } diff --git a/regression-test/suites/tpch_sf1_p0/multi_catalog_query/hive_catalog_orc.groovy b/regression-test/suites/tpch_sf1_p0/multi_catalog_query/hive_catalog_orc.groovy index 13b12e002761ea0..e2c2cf8b1ec9ba7 100644 --- a/regression-test/suites/tpch_sf1_p0/multi_catalog_query/hive_catalog_orc.groovy +++ b/regression-test/suites/tpch_sf1_p0/multi_catalog_query/hive_catalog_orc.groovy @@ -797,35 +797,10 @@ order by """ } - def set_be_config = { -> - String[][] backends = sql """ show backends; """ - assertTrue(backends.size() > 0) - for (String[] backend in backends) { - // No need to set this config anymore, but leave this code sample here - // StringBuilder setConfigCommand = new StringBuilder(); - // setConfigCommand.append("curl -X POST http://") - // setConfigCommand.append(backend[2]) - // setConfigCommand.append(":") - // setConfigCommand.append(backend[5]) - // setConfigCommand.append("/api/update_config?") - // String command1 = setConfigCommand.toString() + "enable_new_load_scan_node=true" - // logger.info(command1) - // String command2 = setConfigCommand.toString() + "enable_new_file_scanner=true" - // logger.info(command2) - // def process1 = command1.execute() - // int code = process1.waitFor() - // assertEquals(code, 0) - // def process2 = command2.execute() - // code = process1.waitFor() - // assertEquals(code, 0) - } - } - String enabled = context.config.otherConfigs.get("enableHiveTest") if (enabled != null && enabled.equalsIgnoreCase("true")) { String hms_port = context.config.otherConfigs.get("hms_port") String catalog_name = "test_catalog_hive_orc" - set_be_config.call() sql """admin set frontend config ("enable_multi_catalog" = "true")""" sql """drop catalog if exists ${catalog_name}""" diff --git a/regression-test/suites/tpch_sf1_p0/multi_catalog_query/hive_catalog_parquet.groovy b/regression-test/suites/tpch_sf1_p0/multi_catalog_query/hive_catalog_parquet.groovy index ce36a181a46a29e..01530b738af3fc5 100644 --- a/regression-test/suites/tpch_sf1_p0/multi_catalog_query/hive_catalog_parquet.groovy +++ b/regression-test/suites/tpch_sf1_p0/multi_catalog_query/hive_catalog_parquet.groovy @@ -797,35 +797,10 @@ order by """ } - def set_be_config = { -> - String[][] backends = sql """ show backends; """ - assertTrue(backends.size() > 0) - for (String[] backend in backends) { - // No need to set this config anymore, but leave this code sample here - // StringBuilder setConfigCommand = new StringBuilder(); - // setConfigCommand.append("curl -X POST http://") - // setConfigCommand.append(backend[2]) - // setConfigCommand.append(":") - // setConfigCommand.append(backend[5]) - // setConfigCommand.append("/api/update_config?") - // String command1 = setConfigCommand.toString() + "enable_new_load_scan_node=true" - // logger.info(command1) - // String command2 = setConfigCommand.toString() + "enable_new_file_scanner=true" - // logger.info(command2) - // def process1 = command1.execute() - // int code = process1.waitFor() - // assertEquals(code, 0) - // def process2 = command2.execute() - // code = process1.waitFor() - // assertEquals(code, 0) - } - } - String enabled = context.config.otherConfigs.get("enableHiveTest") if (enabled != null && enabled.equalsIgnoreCase("true")) { String hms_port = context.config.otherConfigs.get("hms_port") String catalog_name = "test_catalog_hive_parquet" - set_be_config.call() sql """admin set frontend config ("enable_multi_catalog" = "true")""" sql """drop catalog if exists ${catalog_name}"""