This repository was archived by the owner on Feb 27, 2025. It is now read-only.

Description
Ran the test with the same file 8.5GB with old connector https://github.com/Azure/azure-sqldb-spark and this one sql-spark-connector. Used the very same parameters bulkcopy , batchsize 150000, tablelock true for both. With the old connector (spark 2.4.0, scala code) job finished in 16m and with this connector, job took around 38m to finish with spark 3.3.1 and pyspark code. What could be the reason?