Skip to content

Commit

Permalink
[FLINK-18246][python][e2e] Disable PyFlink e2e tests when running on …
Browse files Browse the repository at this point in the history
…jdk11

This closes #12598.
  • Loading branch information
WeiZhong94 authored and dianfu committed Jun 12, 2020
1 parent cf2aba7 commit bfcd800
Showing 1 changed file with 47 additions and 43 deletions.
90 changes: 47 additions & 43 deletions flink-end-to-end-tests/test-scripts/test_pyflink.sh
Expand Up @@ -20,7 +20,6 @@
set -Eeuo pipefail
CURRENT_DIR=`cd "$(dirname "$0")" && pwd -P`
source "${CURRENT_DIR}"/common.sh
source "${CURRENT_DIR}"/common_yarn_docker.sh

cp -r "${FLINK_DIR}/conf" "${TEST_DATA_DIR}/conf"

Expand Down Expand Up @@ -162,45 +161,50 @@ wait_job_terminal_state "$JOB_ID" "FINISHED"

stop_cluster

# test submitting on yarn
start_hadoop_cluster_and_prepare_flink

# copy test files
docker cp "${FLINK_PYTHON_DIR}/dev/lint-python.sh" master:/tmp/
docker cp "${FLINK_PYTHON_TEST_DIR}/target/PythonUdfSqlJobExample.jar" master:/tmp/
docker cp "${FLINK_PYTHON_TEST_DIR}/python/add_one.py" master:/tmp/
docker cp "${REQUIREMENTS_PATH}" master:/tmp/
docker cp "${FLINK_PYTHON_TEST_DIR}/python/python_job.py" master:/tmp/
PYFLINK_PACKAGE_FILE=$(basename "${FLINK_PYTHON_DIR}"/dist/apache-flink-*.tar.gz)
docker cp "${FLINK_PYTHON_DIR}/dist/${PYFLINK_PACKAGE_FILE}" master:/tmp/

# prepare environment
docker exec master bash -c "
/tmp/lint-python.sh -s miniconda
source /tmp/.conda/bin/activate
pip install /tmp/${PYFLINK_PACKAGE_FILE}
conda install -y -q zip=3.0
rm -rf /tmp/.conda/pkgs
cd /tmp
zip -q -r /tmp/venv.zip .conda
echo \"taskmanager.memory.task.off-heap.size: 100m\" >> \"/home/hadoop-user/$FLINK_DIRNAME/conf/flink-conf.yaml\"
"

docker exec master bash -c "export HADOOP_CLASSPATH=\`hadoop classpath\` && \
export PYFLINK_CLIENT_EXECUTABLE=/tmp/.conda/bin/python && \
/home/hadoop-user/$FLINK_DIRNAME/bin/flink run -m yarn-cluster -ytm 1500 -yjm 1000 \
-pyfs /tmp/add_one.py \
-pyreq /tmp/requirements.txt \
-pyarch /tmp/venv.zip \
-pyexec venv.zip/.conda/bin/python \
/tmp/PythonUdfSqlJobExample.jar"

docker exec master bash -c "export HADOOP_CLASSPATH=\`hadoop classpath\` && \
export PYFLINK_CLIENT_EXECUTABLE=/tmp/.conda/bin/python && \
/home/hadoop-user/$FLINK_DIRNAME/bin/flink run -m yarn-cluster -ytm 1500 -yjm 1000 \
-pyfs /tmp/add_one.py \
-pyreq /tmp/requirements.txt \
-pyarch /tmp/venv.zip \
-pyexec venv.zip/.conda/bin/python \
-py /tmp/python_job.py \
pipeline.jars file:/tmp/PythonUdfSqlJobExample.jar"
# These tests are known to fail on JDK11. See FLINK-13719
if [[ ${PROFILE} != *"jdk11"* ]]; then
cd "${CURRENT_DIR}/../"
source "${CURRENT_DIR}"/common_yarn_docker.sh
# test submitting on yarn
start_hadoop_cluster_and_prepare_flink

# copy test files
docker cp "${FLINK_PYTHON_DIR}/dev/lint-python.sh" master:/tmp/
docker cp "${FLINK_PYTHON_TEST_DIR}/target/PythonUdfSqlJobExample.jar" master:/tmp/
docker cp "${FLINK_PYTHON_TEST_DIR}/python/add_one.py" master:/tmp/
docker cp "${REQUIREMENTS_PATH}" master:/tmp/
docker cp "${FLINK_PYTHON_TEST_DIR}/python/python_job.py" master:/tmp/
PYFLINK_PACKAGE_FILE=$(basename "${FLINK_PYTHON_DIR}"/dist/apache-flink-*.tar.gz)
docker cp "${FLINK_PYTHON_DIR}/dist/${PYFLINK_PACKAGE_FILE}" master:/tmp/

# prepare environment
docker exec master bash -c "
/tmp/lint-python.sh -s miniconda
source /tmp/.conda/bin/activate
pip install /tmp/${PYFLINK_PACKAGE_FILE}
conda install -y -q zip=3.0
rm -rf /tmp/.conda/pkgs
cd /tmp
zip -q -r /tmp/venv.zip .conda
echo \"taskmanager.memory.task.off-heap.size: 100m\" >> \"/home/hadoop-user/$FLINK_DIRNAME/conf/flink-conf.yaml\"
"

docker exec master bash -c "export HADOOP_CLASSPATH=\`hadoop classpath\` && \
export PYFLINK_CLIENT_EXECUTABLE=/tmp/.conda/bin/python && \
/home/hadoop-user/$FLINK_DIRNAME/bin/flink run -m yarn-cluster -ytm 1500 -yjm 1000 \
-pyfs /tmp/add_one.py \
-pyreq /tmp/requirements.txt \
-pyarch /tmp/venv.zip \
-pyexec venv.zip/.conda/bin/python \
/tmp/PythonUdfSqlJobExample.jar"

docker exec master bash -c "export HADOOP_CLASSPATH=\`hadoop classpath\` && \
export PYFLINK_CLIENT_EXECUTABLE=/tmp/.conda/bin/python && \
/home/hadoop-user/$FLINK_DIRNAME/bin/flink run -m yarn-cluster -ytm 1500 -yjm 1000 \
-pyfs /tmp/add_one.py \
-pyreq /tmp/requirements.txt \
-pyarch /tmp/venv.zip \
-pyexec venv.zip/.conda/bin/python \
-py /tmp/python_job.py \
pipeline.jars file:/tmp/PythonUdfSqlJobExample.jar"
fi

0 comments on commit bfcd800

Please sign in to comment.