Skip to content

[SPARK-48763][CONNECT][BUILD] Move connect server and common to builtin module #47157

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed

Conversation

HyukjinKwon
Copy link
Member

@HyukjinKwon HyukjinKwon commented Jul 1, 2024

What changes were proposed in this pull request?

This PR proposes to move the connect server to builtin module.

From:

connector/connect/server
connector/connect/common

To:

connect/server
connect/common

Why are the changes needed?

So the end users do not have to specify --packages when they start the Spark Connect server. Spark Connect client remains as a separate module. This was also pointed out in #39928 (comment).

Does this PR introduce any user-facing change?

Yes, users don't have to specify --packages anymore.

How was this patch tested?

CI in this PR should verify them.
Also manually tested several basic commands such as:

  • Maven build

  • SBT build

  • Running basic Scala client commands

    cd connector/connect
    bin/spark-connect
    bin/spark-connect-scala-client
  • Running basic PySpark client commands

    bin/pyspark --remote local
  • Connecting to the server launched by ./sbin/start-connect-server.sh

    ./sbin/start-connect-server.sh
     bin/pyspark --remote "sc://localhost"

Was this patch authored or co-authored using generative AI tooling?

No.

@HyukjinKwon HyukjinKwon marked this pull request as draft July 1, 2024 02:00
@HyukjinKwon HyukjinKwon force-pushed the move-connect-server-builtin branch from a61206e to fb4e543 Compare July 1, 2024 02:02
@HyukjinKwon HyukjinKwon marked this pull request as ready for review July 1, 2024 03:39
@HyukjinKwon HyukjinKwon changed the title [SPARK-48763][CONNECT][BUILD] Move connect server to builtin module [SPARK-48763][CONNECT][BUILD] Move connect server and common to builtin module Jul 1, 2024
@LuciferYang
Copy link
Contributor

connect = Module(
name="connect",
dependencies=[hive, avro, protobuf],
source_file_regexes=[
"connector/connect",
],
build_profile_flags=["-Pconnect"],
sbt_test_goals=[
"connect/test",
"connect-client-jvm/test",
],
)

This place may need to be corrected.

@HyukjinKwon HyukjinKwon force-pushed the move-connect-server-builtin branch from 326739a to 366d76a Compare July 1, 2024 05:57
@zhengruifeng
Copy link
Contributor

qq: what about the client module?

@HyukjinKwon HyukjinKwon force-pushed the move-connect-server-builtin branch from a4c8b27 to 44f87df Compare July 1, 2024 07:16
@HyukjinKwon
Copy link
Member Author

qq: what about the client module?

That stays as a non-builtin library for now because they cannot coexist with Spark Classic jars

@@ -62,7 +62,6 @@ object CheckConnectJvmClientCompatibility {
private val clientJar = {
val path = Paths.get(
sparkHome,
"connector",
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

should not delete this line

@LuciferYang
Copy link
Contributor

@HyukjinKwon I have a question. If we move the dependencies defined in the assembly's connect profile to the default dependency list, perhaps we won't need to make such extensive changes? Or is it a rule that the contents under the connector directory should not be packaged by default in the Spark distribution?

@HyukjinKwon
Copy link
Member Author

Yeah we may unpack the shaded dependencies but I was thinking that could be done separately. Or if you meant something else, we could try.

Copy link
Member

@dongjoon-hyun dongjoon-hyun left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Given the size of this change (2253 files) and the diff of spark-deps-hadoop-3-hive-2.3, I believe we need to be more careful about this change. Maybe, a discussion thread in dev@spark and possibly a vote to get a community bless, @HyukjinKwon .

@HyukjinKwon
Copy link
Member Author

Sure, will do.

@HyukjinKwon
Copy link
Member Author

@@ -74,6 +74,11 @@
<artifactId>spark-repl_${scala.binary.version}</artifactId>
<version>${project.version}</version>
</dependency>
<dependency>
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

So the end users do not have to specify --packages when they start the Spark Connect server.

What I mean is, if we are just trying to address this issue, it would be sufficient to just modify this file, and it appears that there is no need to move the code of the connect server module to a new directory.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I will reply in the dev mailing list

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Maybe We can add the following exclusions to spark-connect_${scala.binary.version, which will prevent any changes to spark-deps-hadoop-3-hive-2.3, since the current spark-connect module is still a shaded jar and does not require the addition of extra third-party dependencies to the project:

<exclusions>
        <exclusion>
          <groupId>org.apache.spark</groupId>
          <artifactId>spark-connect-common_${scala.binary.version}</artifactId>
        </exclusion>
        <exclusion>
          <groupId>io.grpc</groupId>
          <artifactId>*</artifactId>
        </exclusion>
        <exclusion>
          <groupId>com.google.code.gson</groupId>
          <artifactId>gson</artifactId>
        </exclusion>
        <exclusion>
          <groupId>com.google.guava</groupId>
          <artifactId>failureaccess</artifactId>
        </exclusion>
      </exclusions>

Copy link
Member

@pan3793 pan3793 Jul 2, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

FYI: we are already good to unify the Guava version in the whole Spark project, see #42493

@HyukjinKwon HyukjinKwon force-pushed the move-connect-server-builtin branch from 91f7f2c to 5b3b80a Compare July 2, 2024 04:05
<groupId>org.apache.spark</groupId>
<artifactId>spark-avro_${scala.binary.version}</artifactId>
<version>${project.version}</version>
<scope>provided</scope>
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Since the scope is configured as provided, these two jars should not be included in the packaging result.

@pan3793
Copy link
Member

pan3793 commented Jul 2, 2024

leaving the connect client in a separate folder looks weird. I suppose there is no contract to leave all optional modules under connector? e.g. resource-managers/kubernetes/{docker,integration-tests}, hadoop-cloud. What about moving the whole connect folder to the top level?

@HyukjinKwon
Copy link
Member Author

I actually plan to put the connect client at the top level too but it should be together with SPIP so it is there for now.

@HyukjinKwon HyukjinKwon force-pushed the move-connect-server-builtin branch from 3887184 to 5ace07f Compare July 10, 2024 07:38
@HyukjinKwon
Copy link
Member Author

Vote passed at https://lists.apache.org/thread/7pyfy90nbs8ltqyorpgzpl3cp24xrx8s. Thanks all!

@HyukjinKwon
Copy link
Member Author

Merged to master.

HyukjinKwon pushed a commit that referenced this pull request Jul 11, 2024
…rate

### What changes were proposed in this pull request?
The pr is followuping #47157, to make `dev/lint-scala` error message more accurate.

### Why are the changes needed?
After move from: `connector/connect/server` `connector/connect/common` to: `connect/server``connect/common`
Our error message in `dev/lint-scala` should be updated synchronously.

eg:
<img width="709" alt="image" src="https://github.com/apache/spark/assets/15246973/d749e371-7621-4063-b512-279d0690d573">
<img width="772" alt="image" src="https://github.com/apache/spark/assets/15246973/44b80571-bdb6-40cb-9571-8b34d009b5f8">

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
Pass GA.
Manually test.

### Was this patch authored or co-authored using generative AI tooling?
No.

Closes #47291 from panbingkun/SPARK-48763_FOLLOWUP.

Authored-by: panbingkun <panbingkun@baidu.com>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
biruktesf-db pushed a commit to biruktesf-db/spark that referenced this pull request Jul 11, 2024
…in module

### What changes were proposed in this pull request?

This PR proposes to move the connect server to builtin module.

From:

```
connector/connect/server
connector/connect/common
```

To:

```
connect/server
connect/common
```

### Why are the changes needed?

So the end users do not have to specify `--packages` when they start the Spark Connect server. Spark Connect client remains as a separate module. This was also pointed out in apache#39928 (comment).

### Does this PR introduce _any_ user-facing change?

Yes, users don't have to specify `--packages` anymore.

### How was this patch tested?

CI in this PR should verify them.
Also manually tested several basic commands such as:

- Maven build
- SBT build
- Running basic Scala client commands
   ```bash
   cd connector/connect
   bin/spark-connect
   bin/spark-connect-scala-client
   ```
- Running basic PySpark client commands

   ```bash
   bin/pyspark --remote local
   ```
- Connecting to the server launched by `./sbin/start-connect-server.sh`

   ```bash
   ./sbin/start-connect-server.sh
    bin/pyspark --remote "sc://localhost"
   ```

### Was this patch authored or co-authored using generative AI tooling?

No.

Closes apache#47157 from HyukjinKwon/move-connect-server-builtin.

Authored-by: Hyukjin Kwon <gurwls223@apache.org>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
biruktesf-db pushed a commit to biruktesf-db/spark that referenced this pull request Jul 11, 2024
…rate

### What changes were proposed in this pull request?
The pr is followuping apache#47157, to make `dev/lint-scala` error message more accurate.

### Why are the changes needed?
After move from: `connector/connect/server` `connector/connect/common` to: `connect/server``connect/common`
Our error message in `dev/lint-scala` should be updated synchronously.

eg:
<img width="709" alt="image" src="https://github.com/apache/spark/assets/15246973/d749e371-7621-4063-b512-279d0690d573">
<img width="772" alt="image" src="https://github.com/apache/spark/assets/15246973/44b80571-bdb6-40cb-9571-8b34d009b5f8">

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
Pass GA.
Manually test.

### Was this patch authored or co-authored using generative AI tooling?
No.

Closes apache#47291 from panbingkun/SPARK-48763_FOLLOWUP.

Authored-by: panbingkun <panbingkun@baidu.com>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
@bjornjorgensen
Copy link
Contributor

I looks like ./dev/make-distribution.sh --name custom-spark --pip --tgz -Pkubernetes does not work after this.
buildx failed with: ERROR: failed to solve: process "/bin/sh -c ./dev/make-distribution.sh --name custom-spark --pip --tgz -Pkubernetes" did not complete successfully: exit code: 1
but without --pip ./dev/make-distribution.sh --name custom-spark --tgz -Pkubernetes that works.

@HyukjinKwon
Copy link
Member Author

Lemme take a look

@HyukjinKwon
Copy link
Member Author

Hm, it works to me. Can you share the exact command and full error message?

@bjornjorgensen
Copy link
Contributor

oh.. my fault there was something else that did not work as intended on my build system.
@HyukjinKwon thanks for double checking.

jingz-db pushed a commit to jingz-db/spark that referenced this pull request Jul 22, 2024
…in module

### What changes were proposed in this pull request?

This PR proposes to move the connect server to builtin module.

From:

```
connector/connect/server
connector/connect/common
```

To:

```
connect/server
connect/common
```

### Why are the changes needed?

So the end users do not have to specify `--packages` when they start the Spark Connect server. Spark Connect client remains as a separate module. This was also pointed out in apache#39928 (comment).

### Does this PR introduce _any_ user-facing change?

Yes, users don't have to specify `--packages` anymore.

### How was this patch tested?

CI in this PR should verify them.
Also manually tested several basic commands such as:

- Maven build
- SBT build
- Running basic Scala client commands
   ```bash
   cd connector/connect
   bin/spark-connect
   bin/spark-connect-scala-client
   ```
- Running basic PySpark client commands

   ```bash
   bin/pyspark --remote local
   ```
- Connecting to the server launched by `./sbin/start-connect-server.sh`

   ```bash
   ./sbin/start-connect-server.sh
    bin/pyspark --remote "sc://localhost"
   ```

### Was this patch authored or co-authored using generative AI tooling?

No.

Closes apache#47157 from HyukjinKwon/move-connect-server-builtin.

Authored-by: Hyukjin Kwon <gurwls223@apache.org>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
jingz-db pushed a commit to jingz-db/spark that referenced this pull request Jul 22, 2024
…rate

### What changes were proposed in this pull request?
The pr is followuping apache#47157, to make `dev/lint-scala` error message more accurate.

### Why are the changes needed?
After move from: `connector/connect/server` `connector/connect/common` to: `connect/server``connect/common`
Our error message in `dev/lint-scala` should be updated synchronously.

eg:
<img width="709" alt="image" src="https://github.com/apache/spark/assets/15246973/d749e371-7621-4063-b512-279d0690d573">
<img width="772" alt="image" src="https://github.com/apache/spark/assets/15246973/44b80571-bdb6-40cb-9571-8b34d009b5f8">

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
Pass GA.
Manually test.

### Was this patch authored or co-authored using generative AI tooling?
No.

Closes apache#47291 from panbingkun/SPARK-48763_FOLLOWUP.

Authored-by: panbingkun <panbingkun@baidu.com>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
HyukjinKwon added a commit that referenced this pull request Aug 2, 2024
…rver into sql directory

### What changes were proposed in this pull request?

This PR is a followup of #47157 that moves `connect` into `sql/connect`.

### Why are the changes needed?

The reasons are as follow:
- There was a bit of question about moving `connect` as a standalone top level (#47157 (comment)).
- Technically all Spark Connect related code have to placed under `sql` just like Hive thrift server.
  - Spark Connect server is 99% SQL dedicated code for now
  - Spark Connect server already is using a lot of `spark.sql` configurations, e.g., `spark.sql.connect.serverStacktrace.enabled`
  - Spark Connect common is only for SQL module. If other components have to be implemented, that common has to be placed within that directory.

### Does this PR introduce _any_ user-facing change?

No.

### How was this patch tested?

CI in this PR should verify it.

### Was this patch authored or co-authored using generative AI tooling?

No.

Closes #47579 from HyukjinKwon/SPARK-48763-followup.

Authored-by: Hyukjin Kwon <gurwls223@apache.org>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
dongjoon-hyun pushed a commit that referenced this pull request Aug 3, 2024
…rate

### What changes were proposed in this pull request?
The pr is followuping #47157, to make `dev/lint-scala` error message more accurate.

### Why are the changes needed?
After move from: `connector/connect/server` `connector/connect/common` to: `sql/connect/server` `sql/connect/common`
Our error message in `dev/lint-scala` should be updated synchronously.

eg:
<img width="709" alt="image" src="https://github.com/apache/spark/assets/15246973/d749e371-7621-4063-b512-279d0690d573">
<img width="1406" alt="image" src="https://github.com/user-attachments/assets/ab681963-37f7-4f48-9458-61f591477365">

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
Pass GA.
Manually test.

### Was this patch authored or co-authored using generative AI tooling?
No.

Closes #47585 from panbingkun/SPARK-48763_FOLLOWUP_2.

Authored-by: panbingkun <panbingkun@baidu.com>
Signed-off-by: Dongjoon Hyun <dhyun@apple.com>
fusheng9399 pushed a commit to fusheng9399/spark that referenced this pull request Aug 6, 2024
…rver into sql directory

### What changes were proposed in this pull request?

This PR is a followup of apache#47157 that moves `connect` into `sql/connect`.

### Why are the changes needed?

The reasons are as follow:
- There was a bit of question about moving `connect` as a standalone top level (apache#47157 (comment)).
- Technically all Spark Connect related code have to placed under `sql` just like Hive thrift server.
  - Spark Connect server is 99% SQL dedicated code for now
  - Spark Connect server already is using a lot of `spark.sql` configurations, e.g., `spark.sql.connect.serverStacktrace.enabled`
  - Spark Connect common is only for SQL module. If other components have to be implemented, that common has to be placed within that directory.

### Does this PR introduce _any_ user-facing change?

No.

### How was this patch tested?

CI in this PR should verify it.

### Was this patch authored or co-authored using generative AI tooling?

No.

Closes apache#47579 from HyukjinKwon/SPARK-48763-followup.

Authored-by: Hyukjin Kwon <gurwls223@apache.org>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
fusheng9399 pushed a commit to fusheng9399/spark that referenced this pull request Aug 6, 2024
…rate

### What changes were proposed in this pull request?
The pr is followuping apache#47157, to make `dev/lint-scala` error message more accurate.

### Why are the changes needed?
After move from: `connector/connect/server` `connector/connect/common` to: `sql/connect/server` `sql/connect/common`
Our error message in `dev/lint-scala` should be updated synchronously.

eg:
<img width="709" alt="image" src="https://github.com/apache/spark/assets/15246973/d749e371-7621-4063-b512-279d0690d573">
<img width="1406" alt="image" src="https://github.com/user-attachments/assets/ab681963-37f7-4f48-9458-61f591477365">

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
Pass GA.
Manually test.

### Was this patch authored or co-authored using generative AI tooling?
No.

Closes apache#47585 from panbingkun/SPARK-48763_FOLLOWUP_2.

Authored-by: panbingkun <panbingkun@baidu.com>
Signed-off-by: Dongjoon Hyun <dhyun@apple.com>
attilapiros pushed a commit to attilapiros/spark that referenced this pull request Oct 4, 2024
…in module

### What changes were proposed in this pull request?

This PR proposes to move the connect server to builtin module.

From:

```
connector/connect/server
connector/connect/common
```

To:

```
connect/server
connect/common
```

### Why are the changes needed?

So the end users do not have to specify `--packages` when they start the Spark Connect server. Spark Connect client remains as a separate module. This was also pointed out in apache#39928 (comment).

### Does this PR introduce _any_ user-facing change?

Yes, users don't have to specify `--packages` anymore.

### How was this patch tested?

CI in this PR should verify them.
Also manually tested several basic commands such as:

- Maven build
- SBT build
- Running basic Scala client commands
   ```bash
   cd connector/connect
   bin/spark-connect
   bin/spark-connect-scala-client
   ```
- Running basic PySpark client commands

   ```bash
   bin/pyspark --remote local
   ```
- Connecting to the server launched by `./sbin/start-connect-server.sh`

   ```bash
   ./sbin/start-connect-server.sh
    bin/pyspark --remote "sc://localhost"
   ```

### Was this patch authored or co-authored using generative AI tooling?

No.

Closes apache#47157 from HyukjinKwon/move-connect-server-builtin.

Authored-by: Hyukjin Kwon <gurwls223@apache.org>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
attilapiros pushed a commit to attilapiros/spark that referenced this pull request Oct 4, 2024
…rate

### What changes were proposed in this pull request?
The pr is followuping apache#47157, to make `dev/lint-scala` error message more accurate.

### Why are the changes needed?
After move from: `connector/connect/server` `connector/connect/common` to: `connect/server``connect/common`
Our error message in `dev/lint-scala` should be updated synchronously.

eg:
<img width="709" alt="image" src="https://github.com/apache/spark/assets/15246973/d749e371-7621-4063-b512-279d0690d573">
<img width="772" alt="image" src="https://github.com/apache/spark/assets/15246973/44b80571-bdb6-40cb-9571-8b34d009b5f8">

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
Pass GA.
Manually test.

### Was this patch authored or co-authored using generative AI tooling?
No.

Closes apache#47291 from panbingkun/SPARK-48763_FOLLOWUP.

Authored-by: panbingkun <panbingkun@baidu.com>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
attilapiros pushed a commit to attilapiros/spark that referenced this pull request Oct 4, 2024
…rver into sql directory

### What changes were proposed in this pull request?

This PR is a followup of apache#47157 that moves `connect` into `sql/connect`.

### Why are the changes needed?

The reasons are as follow:
- There was a bit of question about moving `connect` as a standalone top level (apache#47157 (comment)).
- Technically all Spark Connect related code have to placed under `sql` just like Hive thrift server.
  - Spark Connect server is 99% SQL dedicated code for now
  - Spark Connect server already is using a lot of `spark.sql` configurations, e.g., `spark.sql.connect.serverStacktrace.enabled`
  - Spark Connect common is only for SQL module. If other components have to be implemented, that common has to be placed within that directory.

### Does this PR introduce _any_ user-facing change?

No.

### How was this patch tested?

CI in this PR should verify it.

### Was this patch authored or co-authored using generative AI tooling?

No.

Closes apache#47579 from HyukjinKwon/SPARK-48763-followup.

Authored-by: Hyukjin Kwon <gurwls223@apache.org>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
attilapiros pushed a commit to attilapiros/spark that referenced this pull request Oct 4, 2024
…rate

### What changes were proposed in this pull request?
The pr is followuping apache#47157, to make `dev/lint-scala` error message more accurate.

### Why are the changes needed?
After move from: `connector/connect/server` `connector/connect/common` to: `sql/connect/server` `sql/connect/common`
Our error message in `dev/lint-scala` should be updated synchronously.

eg:
<img width="709" alt="image" src="https://github.com/apache/spark/assets/15246973/d749e371-7621-4063-b512-279d0690d573">
<img width="1406" alt="image" src="https://github.com/user-attachments/assets/ab681963-37f7-4f48-9458-61f591477365">

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
Pass GA.
Manually test.

### Was this patch authored or co-authored using generative AI tooling?
No.

Closes apache#47585 from panbingkun/SPARK-48763_FOLLOWUP_2.

Authored-by: panbingkun <panbingkun@baidu.com>
Signed-off-by: Dongjoon Hyun <dhyun@apple.com>
himadripal pushed a commit to himadripal/spark that referenced this pull request Oct 19, 2024
…in module

### What changes were proposed in this pull request?

This PR proposes to move the connect server to builtin module.

From:

```
connector/connect/server
connector/connect/common
```

To:

```
connect/server
connect/common
```

### Why are the changes needed?

So the end users do not have to specify `--packages` when they start the Spark Connect server. Spark Connect client remains as a separate module. This was also pointed out in apache#39928 (comment).

### Does this PR introduce _any_ user-facing change?

Yes, users don't have to specify `--packages` anymore.

### How was this patch tested?

CI in this PR should verify them.
Also manually tested several basic commands such as:

- Maven build
- SBT build
- Running basic Scala client commands
   ```bash
   cd connector/connect
   bin/spark-connect
   bin/spark-connect-scala-client
   ```
- Running basic PySpark client commands

   ```bash
   bin/pyspark --remote local
   ```
- Connecting to the server launched by `./sbin/start-connect-server.sh`

   ```bash
   ./sbin/start-connect-server.sh
    bin/pyspark --remote "sc://localhost"
   ```

### Was this patch authored or co-authored using generative AI tooling?

No.

Closes apache#47157 from HyukjinKwon/move-connect-server-builtin.

Authored-by: Hyukjin Kwon <gurwls223@apache.org>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
himadripal pushed a commit to himadripal/spark that referenced this pull request Oct 19, 2024
…rate

### What changes were proposed in this pull request?
The pr is followuping apache#47157, to make `dev/lint-scala` error message more accurate.

### Why are the changes needed?
After move from: `connector/connect/server` `connector/connect/common` to: `connect/server``connect/common`
Our error message in `dev/lint-scala` should be updated synchronously.

eg:
<img width="709" alt="image" src="https://github.com/apache/spark/assets/15246973/d749e371-7621-4063-b512-279d0690d573">
<img width="772" alt="image" src="https://github.com/apache/spark/assets/15246973/44b80571-bdb6-40cb-9571-8b34d009b5f8">

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
Pass GA.
Manually test.

### Was this patch authored or co-authored using generative AI tooling?
No.

Closes apache#47291 from panbingkun/SPARK-48763_FOLLOWUP.

Authored-by: panbingkun <panbingkun@baidu.com>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
himadripal pushed a commit to himadripal/spark that referenced this pull request Oct 19, 2024
…rver into sql directory

### What changes were proposed in this pull request?

This PR is a followup of apache#47157 that moves `connect` into `sql/connect`.

### Why are the changes needed?

The reasons are as follow:
- There was a bit of question about moving `connect` as a standalone top level (apache#47157 (comment)).
- Technically all Spark Connect related code have to placed under `sql` just like Hive thrift server.
  - Spark Connect server is 99% SQL dedicated code for now
  - Spark Connect server already is using a lot of `spark.sql` configurations, e.g., `spark.sql.connect.serverStacktrace.enabled`
  - Spark Connect common is only for SQL module. If other components have to be implemented, that common has to be placed within that directory.

### Does this PR introduce _any_ user-facing change?

No.

### How was this patch tested?

CI in this PR should verify it.

### Was this patch authored or co-authored using generative AI tooling?

No.

Closes apache#47579 from HyukjinKwon/SPARK-48763-followup.

Authored-by: Hyukjin Kwon <gurwls223@apache.org>
Signed-off-by: Hyukjin Kwon <gurwls223@apache.org>
himadripal pushed a commit to himadripal/spark that referenced this pull request Oct 19, 2024
…rate

### What changes were proposed in this pull request?
The pr is followuping apache#47157, to make `dev/lint-scala` error message more accurate.

### Why are the changes needed?
After move from: `connector/connect/server` `connector/connect/common` to: `sql/connect/server` `sql/connect/common`
Our error message in `dev/lint-scala` should be updated synchronously.

eg:
<img width="709" alt="image" src="https://github.com/apache/spark/assets/15246973/d749e371-7621-4063-b512-279d0690d573">
<img width="1406" alt="image" src="https://github.com/user-attachments/assets/ab681963-37f7-4f48-9458-61f591477365">

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
Pass GA.
Manually test.

### Was this patch authored or co-authored using generative AI tooling?
No.

Closes apache#47585 from panbingkun/SPARK-48763_FOLLOWUP_2.

Authored-by: panbingkun <panbingkun@baidu.com>
Signed-off-by: Dongjoon Hyun <dhyun@apple.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

6 participants