Skip to content

Set spark.sql.execution.topKSortFallbackThreshold to a reasonable value #1018

Closed
@yaooqinn

Description

@yaooqinn

1. Describe the feature

In apache/spark#33904, I faced a performance issue with the topK scenario where K is very large which can be avoided by setting spark.sql.execution.topKSortFallbackThreshold less than the K here.

For Kyuubi users who run SQL only, improper topK SQLs will hold engines for quite a long time and they don't know why.

With a proper value for spark.sql.execution.topKSortFallbackThreshold is more suitable for Kyuubi's cases

I'd suggest this value be set to 10000

2. Motivation

3. Describe the solution

4. Additional context

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions