Skip to content

[SPARK-42115][SQL] Push down limit through Python UDFs#39842

Closed
kelvinjian-db wants to merge 2 commits intoapache:masterfrom
kelvinjian-db:SPARK-42115-limit-through-python-udfs
Closed

[SPARK-42115][SQL] Push down limit through Python UDFs#39842
kelvinjian-db wants to merge 2 commits intoapache:masterfrom
kelvinjian-db:SPARK-42115-limit-through-python-udfs

Conversation

@kelvinjian-db
Copy link
Contributor

What changes were proposed in this pull request?

This PR adds cases in LimitPushDown to push limits through Python UDFs. In order to allow for this, we need to call LimitPushDown in SparkOptimizer after the "Extract Python UDFs" batch. We also add PushProjectionThroughLimit afterwards in order to plan CollectLimit.

Why are the changes needed?

Right now, LimitPushdown does not push limits through Python UDFs, which means that expensive Python UDFs can be run on potentially large amounts of input. This PR adds this capability, while ensuring that a GlobalLimit - LocalLimit pattern stays at the top in order to trigger the CollectLimit code path.

Does this PR introduce any user-facing change?

No.

How was this patch tested?

Added a UT.

@cloud-fan
Copy link
Contributor

This is actually a regression caused by #37941

@cloud-fan
Copy link
Contributor

The failed test is no longer valid: test_scalar_iter_udf_close_early (pyspark.sql.tests.pandas.test_pandas_udf_scalar.ScalarPandasUDFTests)

We can remove or ignore it.

@github-actions github-actions bot added the CORE label Feb 1, 2023
@cloud-fan
Copy link
Contributor

thanks, merging to master/3.4!

@cloud-fan cloud-fan closed this in 0fe361e Feb 2, 2023
cloud-fan pushed a commit that referenced this pull request Feb 2, 2023
### What changes were proposed in this pull request?

This PR adds cases in LimitPushDown to push limits through Python UDFs. In order to allow for this, we need to call LimitPushDown in SparkOptimizer after the "Extract Python UDFs" batch. We also add PushProjectionThroughLimit afterwards in order to plan CollectLimit.

### Why are the changes needed?

Right now, LimitPushdown does not push limits through Python UDFs, which means that expensive Python UDFs can be run on potentially large amounts of input. This PR adds this capability, while ensuring that a GlobalLimit - LocalLimit pattern stays at the top in order to trigger the CollectLimit code path.

### Does this PR introduce _any_ user-facing change?

No.

### How was this patch tested?

Added a UT.

Closes #39842 from kelvinjian-db/SPARK-42115-limit-through-python-udfs.

Authored-by: Kelvin Jiang <kelvin.jiang@databricks.com>
Signed-off-by: Wenchen Fan <wenchen@databricks.com>
(cherry picked from commit 0fe361e)
Signed-off-by: Wenchen Fan <wenchen@databricks.com>
snmvaughan pushed a commit to snmvaughan/spark that referenced this pull request Jun 20, 2023
### What changes were proposed in this pull request?

This PR adds cases in LimitPushDown to push limits through Python UDFs. In order to allow for this, we need to call LimitPushDown in SparkOptimizer after the "Extract Python UDFs" batch. We also add PushProjectionThroughLimit afterwards in order to plan CollectLimit.

### Why are the changes needed?

Right now, LimitPushdown does not push limits through Python UDFs, which means that expensive Python UDFs can be run on potentially large amounts of input. This PR adds this capability, while ensuring that a GlobalLimit - LocalLimit pattern stays at the top in order to trigger the CollectLimit code path.

### Does this PR introduce _any_ user-facing change?

No.

### How was this patch tested?

Added a UT.

Closes apache#39842 from kelvinjian-db/SPARK-42115-limit-through-python-udfs.

Authored-by: Kelvin Jiang <kelvin.jiang@databricks.com>
Signed-off-by: Wenchen Fan <wenchen@databricks.com>
(cherry picked from commit 0fe361e)
Signed-off-by: Wenchen Fan <wenchen@databricks.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants