[LOG-PATH]: /opt/module/dolphinscheduler/worker-server/logs/20220615/5713974970944_66-716-10287.log, [HOST]: Host{address='10.80.29.38:1234', ip='10.80.29.38', port=1234} [INFO] 2022-06-15 08:00:14.919 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[69] - spark task params {"localParams":[],"rawScript":"","resourceList":[],"programType":"SCALA","mainClass":"org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer","mainJar":{"id":2053,"resourceName":"/obei_jar/hudi-utilities-bundle_2.12-0.11.0.jar","res":"hudi-utilities-bundle_2.12-0.11.0.jar"},"deployMode":"cluster","mainArgs":"--table-type COPY_ON_WRITE \\\n--props /dolphinscheduler/obeiadmin/resources/Logistics_screen/data_ingestion/mysql-cnb_order-jdbc-source.properties \\\n--source-class org.apache.hudi.utilities.sources.JdbcSource \\\n--source-ordering-field ts \\\n--target-base-path /user/hive/warehouse/bwdhmosaas.db/ods_cnb_order \\\n--target-table ods_cnb_order \\\n--transformer-class org.apache.hudi.utilities.transform.SqlQueryBasedTransformer","sparkVersion":"SPARK2","driverCores":1,"driverMemory":"2G","numExecutors":2,"executorMemory":"4G","executorCores":2} [INFO] 2022-06-15 08:00:14.921 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[130] - spark task command: ${SPARK_HOME2}/bin/spark-submit --master yarn --deploy-mode cluster --class org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer --driver-cores 1 --driver-memory 2G --num-executors 2 --executor-cores 2 --executor-memory 4G --queue spark obei_jar/hudi-utilities-bundle_2.12-0.11.0.jar --table-type COPY_ON_WRITE \ --props /dolphinscheduler/obeiadmin/resources/Logistics_screen/data_ingestion/mysql-cnb_order-jdbc-source.properties \ --source-class org.apache.hudi.utilities.sources.JdbcSource \ --source-ordering-field ts \ --target-base-path /user/hive/warehouse/bwdhmosaas.db/ods_cnb_order \ --target-table ods_cnb_order \ --transformer-class org.apache.hudi.utilities.transform.SqlQueryBasedTransformer [INFO] 2022-06-15 08:00:14.921 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[85] - tenantCode user:obeiadmin, task dir:716_10287 [INFO] 2022-06-15 08:00:14.922 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[90] - create command file:/home/obeiadmin/exec/process/5691395579072/5713974970944_66/716/10287/716_10287.command [INFO] 2022-06-15 08:00:14.922 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[116] - command : #!/bin/sh BASEDIR=$(cd `dirname $0`; pwd) cd $BASEDIR export SPARK_HOME1=/opt/module/spark-3.1.2 export SPARK_HOME2=/opt/module/hudi-spark export HADOOP_HOME=/opt/module/hadoop-3.2.2 export JAVA_HOME=/opt/module/jdk1.8.0_301 export BWDHMOSAAS_PATH=/user/hive/warehouse/bwdhmosaas.db ${SPARK_HOME2}/bin/spark-submit --master yarn --deploy-mode cluster --class org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer --driver-cores 1 --driver-memory 2G --num-executors 2 --executor-cores 2 --executor-memory 4G --queue spark obei_jar/hudi-utilities-bundle_2.12-0.11.0.jar --table-type COPY_ON_WRITE \ --props /dolphinscheduler/obeiadmin/resources/Logistics_screen/data_ingestion/mysql-cnb_order-jdbc-source.properties \ --source-class org.apache.hudi.utilities.sources.JdbcSource \ --source-ordering-field ts \ --target-base-path /user/hive/warehouse/bwdhmosaas.db/ods_cnb_order \ --target-table ods_cnb_order \ --transformer-class org.apache.hudi.utilities.transform.SqlQueryBasedTransformer [INFO] 2022-06-15 08:00:14.926 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[290] - task run command: sudo -u obeiadmin sh /home/obeiadmin/exec/process/5691395579072/5713974970944_66/716/10287/716_10287.command [INFO] 2022-06-15 08:00:14.926 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[181] - process start, process id is: 116509 [INFO] 2022-06-15 08:00:17.927 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:17 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable [INFO] 2022-06-15 08:00:18.932 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:18 INFO DefaultNoHARMFailoverProxyProvider: Connecting to ResourceManager at PVVMON0308/10.80.29.39:8032 22/06/15 08:00:18 INFO Client: Requesting a new application from cluster with 3 NodeManagers [INFO] 2022-06-15 08:00:19.933 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:18 INFO Configuration: resource-types.xml not found 22/06/15 08:00:18 INFO ResourceUtils: Unable to find 'resource-types.xml'. 22/06/15 08:00:19 INFO Client: Verifying our application has not requested more than the maximum memory capability of the cluster (57344 MB per container) 22/06/15 08:00:19 INFO Client: Will allocate AM container, with 2432 MB memory including 384 MB overhead 22/06/15 08:00:19 INFO Client: Setting up container launch context for our AM 22/06/15 08:00:19 INFO Client: Setting up the launch environment for our AM container 22/06/15 08:00:19 INFO Client: Preparing resources for our AM container 22/06/15 08:00:19 WARN Client: Neither spark.yarn.jars nor spark.yarn.archive is set, falling back to uploading libraries under SPARK_HOME. [INFO] 2022-06-15 08:00:22.934 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:22 INFO Client: Uploading resource file:/tmp/spark-599c91bf-6678-4004-9a23-579bcb58b7fa/__spark_libs__8144923296880632219.zip -> hdfs://PVVMON0296:8020/user/obeiadmin/.sparkStaging/application_1653911505034_0117/__spark_libs__8144923296880632219.zip [INFO] 2022-06-15 08:00:24.935 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:24 INFO Client: Uploading resource file:/home/obeiadmin/exec/process/5691395579072/5713974970944_66/716/10287/obei_jar/hudi-utilities-bundle_2.12-0.11.0.jar -> hdfs://PVVMON0296:8020/user/obeiadmin/.sparkStaging/application_1653911505034_0117/hudi-utilities-bundle_2.12-0.11.0.jar [INFO] 2022-06-15 08:00:25.936 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:25 INFO Client: Uploading resource file:/tmp/spark-599c91bf-6678-4004-9a23-579bcb58b7fa/__spark_conf__765929196316604171.zip -> hdfs://PVVMON0296:8020/user/obeiadmin/.sparkStaging/application_1653911505034_0117/__spark_conf__.zip 22/06/15 08:00:25 INFO SecurityManager: Changing view acls to: obeiadmin 22/06/15 08:00:25 INFO SecurityManager: Changing modify acls to: obeiadmin 22/06/15 08:00:25 INFO SecurityManager: Changing view acls groups to: 22/06/15 08:00:25 INFO SecurityManager: Changing modify acls groups to: 22/06/15 08:00:25 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(obeiadmin); groups with view permissions: Set(); users with modify permissions: Set(obeiadmin); groups with modify permissions: Set() 22/06/15 08:00:25 INFO Client: Submitting application application_1653911505034_0117 to ResourceManager 22/06/15 08:00:25 INFO YarnClientImpl: Submitted application application_1653911505034_0117 [INFO] 2022-06-15 08:00:26.937 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:26 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) 22/06/15 08:00:26 INFO Client: client token: N/A diagnostics: AM container is launched, waiting for AM container to Register with RM ApplicationMaster host: N/A ApplicationMaster RPC port: -1 queue: spark start time: 1655251225462 final status: UNDEFINED tracking URL: http://PVVMON0308:8088/proxy/application_1653911505034_0117/ user: obeiadmin [INFO] 2022-06-15 08:00:27.938 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:27 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:00:28.939 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:28 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:00:29.940 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:29 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:00:30.941 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:30 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:00:31.942 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:31 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:00:32.943 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:32 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:00:33.944 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:33 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:00:34.945 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:34 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:00:35.946 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:35 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:00:36.947 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:36 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:00:37.948 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:37 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) 22/06/15 08:00:37 INFO Client: client token: N/A diagnostics: N/A ApplicationMaster host: PVVMON0295 ApplicationMaster RPC port: 37260 queue: spark start time: 1655251225462 final status: UNDEFINED tracking URL: http://PVVMON0308:8088/proxy/application_1653911505034_0117/ user: obeiadmin [INFO] 2022-06-15 08:00:38.949 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:38 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:39.950 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:39 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:40.951 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:40 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:41.952 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:41 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:42.953 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:42 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:43.954 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:43 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:44.955 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:44 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:45.956 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:45 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:46.957 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:46 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:47.958 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:47 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:48.959 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:48 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:49.960 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:49 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:50.961 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:50 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:51.962 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:51 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:52.963 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:52 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:53.964 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:53 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:54.965 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:54 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:55.966 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:55 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:56.967 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:56 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:57.968 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:57 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:58.969 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:58 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:00:59.970 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:00:59 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:00.972 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:00 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:01.976 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:01 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:02.980 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:02 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:03.984 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:03 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:04.987 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:04 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:05.989 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:05 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:06.990 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:06 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:07.991 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:07 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:08.992 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:08 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:09.993 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:09 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:10.994 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:10 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:11.995 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:11 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:12.996 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:12 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:13.997 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:13 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:14.998 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:14 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:15.999 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:15 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:17.000 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:16 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:18.002 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:17 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:19.003 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:18 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:20.005 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:19 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:21.006 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:20 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:22.007 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:21 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:23.008 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:22 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:24.009 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:23 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:25.010 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:24 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:26.011 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:25 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:27.012 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:26 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:28.013 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:27 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:29.014 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:28 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:30.015 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:29 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:31.016 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:30 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:32.017 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:31 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:33.018 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:32 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:34.019 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:33 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:35.020 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:34 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:36.021 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:35 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:37.022 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:36 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:38.023 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:37 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:39.024 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:38 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:40.025 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:39 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:41.026 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:40 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) 22/06/15 08:01:40 INFO Client: client token: N/A diagnostics: [星期三 六月 15 08:01:40 +0800 2022] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = ; Partition Resource = ; Queue's Absolute capacity = 60.000004 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 80.0 % ; Queue's capacity (absolute resource) = ; Queue's used capacity (absolute resource) = ; Queue's max capacity (absolute resource) = ; ApplicationMaster host: N/A ApplicationMaster RPC port: -1 queue: spark start time: 1655251225462 final status: UNDEFINED tracking URL: http://PVVMON0308:8088/proxy/application_1653911505034_0117/ user: obeiadmin [INFO] 2022-06-15 08:01:42.027 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:41 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:01:43.028 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:42 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:01:44.029 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:43 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:01:45.030 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:44 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:01:46.031 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:45 INFO Client: Application report for application_1653911505034_0117 (state: ACCEPTED) [INFO] 2022-06-15 08:01:47.032 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:46 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) 22/06/15 08:01:46 INFO Client: client token: N/A diagnostics: N/A ApplicationMaster host: PVVMON0296 ApplicationMaster RPC port: 43091 queue: spark start time: 1655251225462 final status: UNDEFINED tracking URL: http://PVVMON0308:8088/proxy/application_1653911505034_0117/ user: obeiadmin [INFO] 2022-06-15 08:01:48.033 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:47 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:49.034 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:48 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:50.035 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:49 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:51.036 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:50 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:52.037 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:51 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:53.038 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:52 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:54.039 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:53 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:55.040 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:54 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:56.041 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:55 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:57.042 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:56 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:58.043 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:57 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:01:59.044 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:58 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:00.046 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:01:59 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:01.047 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:00 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:02.048 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:01 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:03.049 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:02 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:04.050 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:03 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:05.051 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:04 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:06.052 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:05 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:07.053 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:06 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:08.054 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:07 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:09.055 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:08 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:10.056 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:09 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:11.057 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:10 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:12.058 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:11 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:13.059 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:12 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:14.060 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:13 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:15.061 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:14 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:16.062 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:15 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:17.063 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:16 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:18.064 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:17 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:19.065 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:18 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:20.066 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:19 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:21.067 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:20 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:22.068 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:21 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:23.069 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:22 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:24.070 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:23 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:25.071 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:24 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:26.072 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:25 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:27.073 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:26 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:28.074 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:27 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:29.075 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:28 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:30.076 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:29 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:31.077 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:30 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:32.078 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:31 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:33.079 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:32 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:34.080 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:33 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:35.081 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:34 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:36.082 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:35 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:37.083 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:36 INFO Client: Application report for application_1653911505034_0117 (state: RUNNING) [INFO] 2022-06-15 08:02:38.083 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[63] - -> 22/06/15 08:02:37 INFO Client: Application report for application_1653911505034_0117 (state: FINISHED) 22/06/15 08:02:37 INFO Client: client token: N/A diagnostics: User class threw exception: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 52.0 failed 4 times, most recent failure: Lost task 0.3 in stage 52.0 (TID 47) (PVVMON0295 executor 1): java.lang.NoSuchMethodError: org.apache.hadoop.hdfs.client.HdfsDataInputStream.getReadStatistics()Lorg/apache/hadoop/hdfs/DFSInputStream$ReadStatistics; at org.apache.hudi.org.apache.hadoop.hbase.io.FSDataInputStreamWrapper.updateInputStreamStatistics(FSDataInputStreamWrapper.java:249) at org.apache.hudi.org.apache.hadoop.hbase.io.FSDataInputStreamWrapper.close(FSDataInputStreamWrapper.java:296) at org.apache.hudi.org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.closeStreams(HFileBlock.java:1825) at org.apache.hudi.org.apache.hadoop.hbase.io.hfile.HFilePreadReader.close(HFilePreadReader.java:107) at org.apache.hudi.org.apache.hadoop.hbase.io.hfile.HFileReaderImpl.close(HFileReaderImpl.java:1421) at org.apache.hudi.io.storage.HoodieHFileReader.close(HoodieHFileReader.java:218) at org.apache.hudi.table.action.commit.HoodieMergeHelper.runMerge(HoodieMergeHelper.java:152) at org.apache.hudi.table.HoodieSparkCopyOnWriteTable.handleUpdateInternal(HoodieSparkCopyOnWriteTable.java:230) at org.apache.hudi.table.HoodieSparkCopyOnWriteTable.handleUpdate(HoodieSparkCopyOnWriteTable.java:221) at org.apache.hudi.table.action.compact.HoodieCompactor.compact(HoodieCompactor.java:227) at org.apache.hudi.table.action.compact.HoodieCompactor.lambda$compact$57154431$1(HoodieCompactor.java:138) at org.apache.spark.api.java.JavaPairRDD$.$anonfun$toScalaFunction$1(JavaPairRDD.scala:1070) at scala.collection.Iterator$$anon$10.next(Iterator.scala:461) at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492) at org.apache.spark.storage.memory.MemoryStore.putIterator(MemoryStore.scala:223) at org.apache.spark.storage.memory.MemoryStore.putIteratorAsBytes(MemoryStore.scala:352) at org.apache.spark.storage.BlockManager.$anonfun$doPutIterator$1(BlockManager.scala:1498) at org.apache.spark.storage.BlockManager.org$apache$spark$storage$BlockManager$$doPut(BlockManager.scala:1408) at org.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:1472) at org.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:1295) at org.apache.spark.rdd.RDD.getOrCompute(RDD.scala:384) at org.apache.spark.rdd.RDD.iterator(RDD.scala:335) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) at org.apache.spark.scheduler.Task.run(Task.scala:131) at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:506) at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1462) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:509) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Driver stacktrace: at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2454) at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2403) at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2402) at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62) at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49) at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2402) at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1160) at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1160) at scala.Option.foreach(Option.scala:407) at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1160) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2642) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2584) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2573) at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:938) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2214) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2235) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2254) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2279) at org.apache.spark.rdd.RDD.$anonfun$collect$1(RDD.scala:1030) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) at org.apache.spark.rdd.RDD.withScope(RDD.scala:414) at org.apache.spark.rdd.RDD.collect(RDD.scala:1029) at org.apache.spark.api.java.JavaRDDLike.collect(JavaRDDLike.scala:362) at org.apache.spark.api.java.JavaRDDLike.collect$(JavaRDDLike.scala:361) at org.apache.spark.api.java.AbstractJavaRDDLike.collect(JavaRDDLike.scala:45) at org.apache.hudi.data.HoodieJavaRDD.collectAsList(HoodieJavaRDD.java:157) at org.apache.hudi.table.action.compact.RunCompactionActionExecutor.execute(RunCompactionActionExecutor.java:92) at org.apache.hudi.table.HoodieSparkMergeOnReadTable.compact(HoodieSparkMergeOnReadTable.java:142) at org.apache.hudi.client.SparkRDDWriteClient.compact(SparkRDDWriteClient.java:345) at org.apache.hudi.client.BaseHoodieWriteClient.lambda$runAnyPendingCompactions$1(BaseHoodieWriteClient.java:588) at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1384) at java.util.stream.ReferencePipeline$Head.forEach(ReferencePipeline.java:580) at org.apache.hudi.client.BaseHoodieWriteClient.runAnyPendingCompactions(BaseHoodieWriteClient.java:586) at org.apache.hudi.client.BaseHoodieWriteClient.runAnyPendingCompactions(BaseHoodieWriteClient.java:637) at org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.compactIfNecessary(HoodieBackedTableMetadataWriter.java:997) at org.apache.hudi.metadata.SparkHoodieBackedTableMetadataWriter.commit(SparkHoodieBackedTableMetadataWriter.java:140) at org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.processAndCommit(HoodieBackedTableMetadataWriter.java:803) at org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.update(HoodieBackedTableMetadataWriter.java:870) at org.apache.hudi.client.BaseHoodieWriteClient.lambda$writeTableMetadata$0(BaseHoodieWriteClient.java:338) at org.apache.hudi.common.util.Option.ifPresent(Option.java:97) at org.apache.hudi.client.BaseHoodieWriteClient.writeTableMetadata(BaseHoodieWriteClient.java:338) at org.apache.hudi.client.BaseHoodieWriteClient.commit(BaseHoodieWriteClient.java:269) at org.apache.hudi.client.BaseHoodieWriteClient.commitStats(BaseHoodieWriteClient.java:234) at org.apache.hudi.client.SparkRDDWriteClient.commit(SparkRDDWriteClient.java:122) at org.apache.hudi.utilities.deltastreamer.DeltaSync.writeToSink(DeltaSync.java:622) at org.apache.hudi.utilities.deltastreamer.DeltaSync.syncOnce(DeltaSync.java:331) at org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer.lambda$sync$2(HoodieDeltaStreamer.java:200) at org.apache.hudi.common.util.Option.ifPresent(Option.java:97) at org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer.sync(HoodieDeltaStreamer.java:198) at org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer.main(HoodieDeltaStreamer.java:549) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:737) Caused by: java.lang.NoSuchMethodError: org.apache.hadoop.hdfs.client.HdfsDataInputStream.getReadStatistics()Lorg/apache/hadoop/hdfs/DFSInputStream$ReadStatistics; at org.apache.hudi.org.apache.hadoop.hbase.io.FSDataInputStreamWrapper.updateInputStreamStatistics(FSDataInputStreamWrapper.java:249) at org.apache.hudi.org.apache.hadoop.hbase.io.FSDataInputStreamWrapper.close(FSDataInputStreamWrapper.java:296) at org.apache.hudi.org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.closeStreams(HFileBlock.java:1825) at org.apache.hudi.org.apache.hadoop.hbase.io.hfile.HFilePreadReader.close(HFilePreadReader.java:107) at org.apache.hudi.org.apache.hadoop.hbase.io.hfile.HFileReaderImpl.close(HFileReaderImpl.java:1421) at org.apache.hudi.io.storage.HoodieHFileReader.close(HoodieHFileReader.java:218) at org.apache.hudi.table.action.commit.HoodieMergeHelper.runMerge(HoodieMergeHelper.java:152) at org.apache.hudi.table.HoodieSparkCopyOnWriteTable.handleUpdateInternal(HoodieSparkCopyOnWriteTable.java:230) at org.apache.hudi.table.HoodieSparkCopyOnWriteTable.handleUpdate(HoodieSparkCopyOnWriteTable.java:221) at org.apache.hudi.table.action.compact.HoodieCompactor.compact(HoodieCompactor.java:227) at org.apache.hudi.table.action.compact.HoodieCompactor.lambda$compact$57154431$1(HoodieCompactor.java:138) at org.apache.spark.api.java.JavaPairRDD$.$anonfun$toScalaFunction$1(JavaPairRDD.scala:1070) at scala.collection.Iterator$$anon$10.next(Iterator.scala:461) at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492) at org.apache.spark.storage.memory.MemoryStore.putIterator(MemoryStore.scala:223) at org.apache.spark.storage.memory.MemoryStore.putIteratorAsBytes(MemoryStore.scala:352) at org.apache.spark.storage.BlockManager.$anonfun$doPutIterator$1(BlockManager.scala:1498) at org.apache.spark.storage.BlockManager.org$apache$spark$storage$BlockManager$$doPut(BlockManager.scala:1408) at org.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:1472) at org.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:1295) at org.apache.spark.rdd.RDD.getOrCompute(RDD.scala:384) at org.apache.spark.rdd.RDD.iterator(RDD.scala:335) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) at org.apache.spark.scheduler.Task.run(Task.scala:131) at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:506) at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1462) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:509) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) ApplicationMaster host: PVVMON0296 ApplicationMaster RPC port: 43091 queue: spark start time: 1655251225462 final status: FAILED tracking URL: http://PVVMON0308:8088/proxy/application_1653911505034_0117/ user: obeiadmin 22/06/15 08:02:37 ERROR Client: Application diagnostics message: User class threw exception: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 52.0 failed 4 times, most recent failure: Lost task 0.3 in stage 52.0 (TID 47) (PVVMON0295 executor 1): java.lang.NoSuchMethodError: org.apache.hadoop.hdfs.client.HdfsDataInputStream.getReadStatistics()Lorg/apache/hadoop/hdfs/DFSInputStream$ReadStatistics; at org.apache.hudi.org.apache.hadoop.hbase.io.FSDataInputStreamWrapper.updateInputStreamStatistics(FSDataInputStreamWrapper.java:249) at org.apache.hudi.org.apache.hadoop.hbase.io.FSDataInputStreamWrapper.close(FSDataInputStreamWrapper.java:296) at org.apache.hudi.org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.closeStreams(HFileBlock.java:1825) at org.apache.hudi.org.apache.hadoop.hbase.io.hfile.HFilePreadReader.close(HFilePreadReader.java:107) at org.apache.hudi.org.apache.hadoop.hbase.io.hfile.HFileReaderImpl.close(HFileReaderImpl.java:1421) at org.apache.hudi.io.storage.HoodieHFileReader.close(HoodieHFileReader.java:218) at org.apache.hudi.table.action.commit.HoodieMergeHelper.runMerge(HoodieMergeHelper.java:152) at org.apache.hudi.table.HoodieSparkCopyOnWriteTable.handleUpdateInternal(HoodieSparkCopyOnWriteTable.java:230) at org.apache.hudi.table.HoodieSparkCopyOnWriteTable.handleUpdate(HoodieSparkCopyOnWriteTable.java:221) at org.apache.hudi.table.action.compact.HoodieCompactor.compact(HoodieCompactor.java:227) at org.apache.hudi.table.action.compact.HoodieCompactor.lambda$compact$57154431$1(HoodieCompactor.java:138) at org.apache.spark.api.java.JavaPairRDD$.$anonfun$toScalaFunction$1(JavaPairRDD.scala:1070) at scala.collection.Iterator$$anon$10.next(Iterator.scala:461) at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492) at org.apache.spark.storage.memory.MemoryStore.putIterator(MemoryStore.scala:223) at org.apache.spark.storage.memory.MemoryStore.putIteratorAsBytes(MemoryStore.scala:352) at org.apache.spark.storage.BlockManager.$anonfun$doPutIterator$1(BlockManager.scala:1498) at org.apache.spark.storage.BlockManager.org$apache$spark$storage$BlockManager$$doPut(BlockManager.scala:1408) at org.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:1472) at org.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:1295) at org.apache.spark.rdd.RDD.getOrCompute(RDD.scala:384) at org.apache.spark.rdd.RDD.iterator(RDD.scala:335) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) at org.apache.spark.scheduler.Task.run(Task.scala:131) at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:506) at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1462) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:509) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Driver stacktrace: at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2454) at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2403) at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2402) at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62) at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49) at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2402) at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1160) at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1160) at scala.Option.foreach(Option.scala:407) at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1160) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2642) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2584) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2573) at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:938) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2214) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2235) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2254) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2279) at org.apache.spark.rdd.RDD.$anonfun$collect$1(RDD.scala:1030) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) at org.apache.spark.rdd.RDD.withScope(RDD.scala:414) at org.apache.spark.rdd.RDD.collect(RDD.scala:1029) at org.apache.spark.api.java.JavaRDDLike.collect(JavaRDDLike.scala:362) at org.apache.spark.api.java.JavaRDDLike.collect$(JavaRDDLike.scala:361) at org.apache.spark.api.java.AbstractJavaRDDLike.collect(JavaRDDLike.scala:45) at org.apache.hudi.data.HoodieJavaRDD.collectAsList(HoodieJavaRDD.java:157) at org.apache.hudi.table.action.compact.RunCompactionActionExecutor.execute(RunCompactionActionExecutor.java:92) at org.apache.hudi.table.HoodieSparkMergeOnReadTable.compact(HoodieSparkMergeOnReadTable.java:142) at org.apache.hudi.client.SparkRDDWriteClient.compact(SparkRDDWriteClient.java:345) at org.apache.hudi.client.BaseHoodieWriteClient.lambda$runAnyPendingCompactions$1(BaseHoodieWriteClient.java:588) at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1384) at java.util.stream.ReferencePipeline$Head.forEach(ReferencePipeline.java:580) at org.apache.hudi.client.BaseHoodieWriteClient.runAnyPendingCompactions(BaseHoodieWriteClient.java:586) at org.apache.hudi.client.BaseHoodieWriteClient.runAnyPendingCompactions(BaseHoodieWriteClient.java:637) at org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.compactIfNecessary(HoodieBackedTableMetadataWriter.java:997) at org.apache.hudi.metadata.SparkHoodieBackedTableMetadataWriter.commit(SparkHoodieBackedTableMetadataWriter.java:140) at org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.processAndCommit(HoodieBackedTableMetadataWriter.java:803) at org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.update(HoodieBackedTableMetadataWriter.java:870) at org.apache.hudi.client.BaseHoodieWriteClient.lambda$writeTableMetadata$0(BaseHoodieWriteClient.java:338) at org.apache.hudi.common.util.Option.ifPresent(Option.java:97) at org.apache.hudi.client.BaseHoodieWriteClient.writeTableMetadata(BaseHoodieWriteClient.java:338) at org.apache.hudi.client.BaseHoodieWriteClient.commit(BaseHoodieWriteClient.java:269) at org.apache.hudi.client.BaseHoodieWriteClient.commitStats(BaseHoodieWriteClient.java:234) at org.apache.hudi.client.SparkRDDWriteClient.commit(SparkRDDWriteClient.java:122) at org.apache.hudi.utilities.deltastreamer.DeltaSync.writeToSink(DeltaSync.java:622) at org.apache.hudi.utilities.deltastreamer.DeltaSync.syncOnce(DeltaSync.java:331) at org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer.lambda$sync$2(HoodieDeltaStreamer.java:200) at org.apache.hudi.common.util.Option.ifPresent(Option.java:97) at org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer.sync(HoodieDeltaStreamer.java:198) at org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer.main(HoodieDeltaStreamer.java:549) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:737) Caused by: java.lang.NoSuchMethodError: org.apache.hadoop.hdfs.client.HdfsDataInputStream.getReadStatistics()Lorg/apache/hadoop/hdfs/DFSInputStream$ReadStatistics; at org.apache.hudi.org.apache.hadoop.hbase.io.FSDataInputStreamWrapper.updateInputStreamStatistics(FSDataInputStreamWrapper.java:249) at org.apache.hudi.org.apache.hadoop.hbase.io.FSDataInputStreamWrapper.close(FSDataInputStreamWrapper.java:296) at org.apache.hudi.org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.closeStreams(HFileBlock.java:1825) at org.apache.hudi.org.apache.hadoop.hbase.io.hfile.HFilePreadReader.close(HFilePreadReader.java:107) at org.apache.hudi.org.apache.hadoop.hbase.io.hfile.HFileReaderImpl.close(HFileReaderImpl.java:1421) at org.apache.hudi.io.storage.HoodieHFileReader.close(HoodieHFileReader.java:218) at org.apache.hudi.table.action.commit.HoodieMergeHelper.runMerge(HoodieMergeHelper.java:152) at org.apache.hudi.table.HoodieSparkCopyOnWriteTable.handleUpdateInternal(HoodieSparkCopyOnWriteTable.java:230) at org.apache.hudi.table.HoodieSparkCopyOnWriteTable.handleUpdate(HoodieSparkCopyOnWriteTable.java:221) at org.apache.hudi.table.action.compact.HoodieCompactor.compact(HoodieCompactor.java:227) at org.apache.hudi.table.action.compact.HoodieCompactor.lambda$compact$57154431$1(HoodieCompactor.java:138) at org.apache.spark.api.java.JavaPairRDD$.$anonfun$toScalaFunction$1(JavaPairRDD.scala:1070) at scala.collection.Iterator$$anon$10.next(Iterator.scala:461) at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492) at org.apache.spark.storage.memory.MemoryStore.putIterator(MemoryStore.scala:223) at org.apache.spark.storage.memory.MemoryStore.putIteratorAsBytes(MemoryStore.scala:352) at org.apache.spark.storage.BlockManager.$anonfun$doPutIterator$1(BlockManager.scala:1498) at org.apache.spark.storage.BlockManager.org$apache$spark$storage$BlockManager$$doPut(BlockManager.scala:1408) at org.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:1472) at org.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:1295) at org.apache.spark.rdd.RDD.getOrCompute(RDD.scala:384) at org.apache.spark.rdd.RDD.iterator(RDD.scala:335) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) at org.apache.spark.scheduler.Task.run(Task.scala:131) at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:506) at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1462) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:509) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Exception in thread "main" org.apache.spark.SparkException: Application application_1653911505034_0117 finished with failed status at org.apache.spark.deploy.yarn.Client.run(Client.scala:1283) at org.apache.spark.deploy.yarn.YarnClusterApplication.start(Client.scala:1677) at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:955) at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180) at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203) at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90) at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1043) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1052) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 22/06/15 08:02:37 INFO ShutdownHookManager: Shutdown hook called 22/06/15 08:02:37 INFO ShutdownHookManager: Deleting directory /tmp/spark-599c91bf-6678-4004-9a23-579bcb58b7fa 22/06/15 08:02:37 INFO ShutdownHookManager: Deleting directory /tmp/spark-04477586-7547-469e-a5ed-b9fa1e76c62f [INFO] 2022-06-15 08:02:38.132 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[375] - find app id: application_1653911505034_0117 [INFO] 2022-06-15 08:02:38.134 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[205] - process has exited, execute path:/home/obeiadmin/exec/process/5691395579072/5713974970944_66/716/10287, processId:116509 ,exitStatusCode:1 ,processWaitForStatus:true ,processExitValue:1 [INFO] 2022-06-15 08:02:39.087 +0800 [taskAppId=TASK-20220615-5713974970944_66-716-10287] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.spark.SparkTask:[57] - FINALIZE_SESSION