-
Notifications
You must be signed in to change notification settings - Fork 33
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support Python 3.11 #30
Comments
Python 3.11 has been released as scheduled on October 25, 2022 and this is the first attempt to see how far Airflow (mostly dependencies) are from being ready to officially support 3.11. So far we had to exclude the following dependencies: - [ ] Pyarrow dependency: apache/arrow#14499 - [ ] Google Provider: #27292 and googleapis/python-bigquery#1386 - [ ] Databricks Provider: databricks/databricks-sql-python#59 - [ ] Papermill Provider: nteract/papermill#700 - [ ] Azure Provider: Azure/azure-uamqp-python#334 and Azure/azure-sdk-for-python#27066 - [ ] Apache Beam Provider: apache/beam#23848 - [ ] Snowflake Provider: snowflakedb/snowflake-connector-python#1294 - [ ] JDBC Provider: jpype-project/jpype#1087 - [ ] Hive Provider: cloudera/python-sasl#30 We might decide to release Airflow in 3.11 with those providers disabled in case they are lagging behind eventually, but for the moment we want to work with all the projects in concert to be able to release all providers (Google Provider requires quite a lot of work and likely Google Team stepping up and community helping with migration to latest Goofle cloud libraries)
Python 3.11 has been released as scheduled on October 25, 2022 and this is the first attempt to see how far Airflow (mostly dependencies) are from being ready to officially support 3.11. So far we had to exclude the following dependencies: - [ ] Pyarrow dependency: apache/arrow#14499 - [ ] Google Provider: #27292 and googleapis/python-bigquery#1386 - [ ] Databricks Provider: databricks/databricks-sql-python#59 - [ ] Papermill Provider: nteract/papermill#700 - [ ] Azure Provider: Azure/azure-uamqp-python#334 and Azure/azure-sdk-for-python#27066 - [ ] Apache Beam Provider: apache/beam#23848 - [ ] Snowflake Provider: snowflakedb/snowflake-connector-python#1294 - [ ] JDBC Provider: jpype-project/jpype#1087 - [ ] Hive Provider: cloudera/python-sasl#30 We might decide to release Airflow in 3.11 with those providers disabled in case they are lagging behind eventually, but for the moment we want to work with all the projects in concert to be able to release all providers (Google Provider requires quite a lot of work and likely Google Team stepping up and community helping with migration to latest Goofle cloud libraries)
Python 3.11 has been released as scheduled on October 25, 2022 and this is the first attempt to see how far Airflow (mostly dependencies) are from being ready to officially support 3.11. So far we had to exclude the following dependencies: - [ ] Pyarrow dependency: apache/arrow#14499 - [ ] Google Provider: #27292 and googleapis/python-bigquery#1386 - [ ] Databricks Provider: databricks/databricks-sql-python#59 - [ ] Papermill Provider: nteract/papermill#700 - [ ] Azure Provider: Azure/azure-uamqp-python#334 and Azure/azure-sdk-for-python#27066 - [ ] Apache Beam Provider: apache/beam#23848 - [ ] Snowflake Provider: snowflakedb/snowflake-connector-python#1294 - [ ] JDBC Provider: jpype-project/jpype#1087 - [ ] Hive Provider: cloudera/python-sasl#30 We might decide to release Airflow in 3.11 with those providers disabled in case they are lagging behind eventually, but for the moment we want to work with all the projects in concert to be able to release all providers (Google Provider requires quite a lot of work and likely Google Team stepping up and community helping with migration to latest Goofle cloud libraries)
Python 3.11 has been released as scheduled on October 25, 2022 and this is the first attempt to see how far Airflow (mostly dependencies) are from being ready to officially support 3.11. So far we had to exclude the following dependencies: - [ ] Pyarrow dependency: apache/arrow#14499 - [ ] Google Provider: #27292 and googleapis/python-bigquery#1386 - [ ] Databricks Provider: databricks/databricks-sql-python#59 - [ ] Papermill Provider: nteract/papermill#700 - [ ] Azure Provider: Azure/azure-uamqp-python#334 and Azure/azure-sdk-for-python#27066 - [ ] Apache Beam Provider: apache/beam#23848 - [ ] Snowflake Provider: snowflakedb/snowflake-connector-python#1294 - [ ] JDBC Provider: jpype-project/jpype#1087 - [ ] Hive Provider: cloudera/python-sasl#30 We might decide to release Airflow in 3.11 with those providers disabled in case they are lagging behind eventually, but for the moment we want to work with all the projects in concert to be able to release all providers (Google Provider requires quite a lot of work and likely Google Team stepping up and community helping with migration to latest Goofle cloud libraries)
Python 3.11 has been released as scheduled on October 25, 2022 and this is the first attempt to see how far Airflow (mostly dependencies) are from being ready to officially support 3.11. So far we had to exclude the following dependencies: - [ ] Pyarrow dependency: apache/arrow#14499 - [ ] Google Provider: #27292 and googleapis/python-bigquery#1386 - [ ] Databricks Provider: databricks/databricks-sql-python#59 - [ ] Papermill Provider: nteract/papermill#700 - [ ] Azure Provider: Azure/azure-uamqp-python#334 and Azure/azure-sdk-for-python#27066 - [ ] Apache Beam Provider: apache/beam#23848 - [ ] Snowflake Provider: snowflakedb/snowflake-connector-python#1294 - [ ] JDBC Provider: jpype-project/jpype#1087 - [ ] Hive Provider: cloudera/python-sasl#30 We might decide to release Airflow in 3.11 with those providers disabled in case they are lagging behind eventually, but for the moment we want to work with all the projects in concert to be able to release all providers (Google Provider requires quite a lot of work and likely Google Team stepping up and community helping with migration to latest Goofle cloud libraries)
Python 3.11 has been released as scheduled on October 25, 2022 and this is the first attempt to see how far Airflow (mostly dependencies) are from being ready to officially support 3.11. So far we had to exclude the following dependencies: - [ ] Pyarrow dependency: apache/arrow#14499 - [ ] Google Provider: apache#27292 and googleapis/python-bigquery#1386 - [ ] Databricks Provider: databricks/databricks-sql-python#59 - [ ] Papermill Provider: nteract/papermill#700 - [ ] Azure Provider: Azure/azure-uamqp-python#334 and Azure/azure-sdk-for-python#27066 - [ ] Apache Beam Provider: apache/beam#23848 - [ ] Snowflake Provider: snowflakedb/snowflake-connector-python#1294 - [ ] JDBC Provider: jpype-project/jpype#1087 - [ ] Hive Provider: cloudera/python-sasl#30 We might decide to release Airflow in 3.11 with those providers disabled in case they are lagging behind eventually, but for the moment we want to work with all the projects in concert to be able to release all providers (Google Provider requires quite a lot of work and likely Google Team stepping up and community helping with migration to latest Goofle cloud libraries)
bumping this up. I would review the MR myself but unfortunately it's a bit beyond my depth. |
I hope someone from cloudera (@attilajeges ? ) might actually review and merge the PR #31 and release python-sasl. I am forced to remove hive provider because of this problem from our Python 3.11 releases. |
Bump ... running into the same problem with "longintrepr.h" ... Would appreciate this fix. |
Ping, Please look into PR 31 |
I am installing localstack[runtime] and it failed because of this error, not sure whether there is a workaround? |
Just a bit of a warning here for Python-sasl maintainers. We are very close to have Python 3.11 support in Apache Airflow - we are just about to merge the Google Provider upgrading ~20 client libraries which was the biggest blocker, and ApacheBeam released 2.47.0 version for Python 3.11. Similarly as it happened alredy with yandex provider - we are going to suspend the hive provider from our releases if Papermill will be holding us back. We have a process for that described in our process: https://github.com/apache/airflow/blob/main/PROVIDERS.rst#suspending-releases-for-providers - you can also learn there what are the consequences of being suspended (in short - no new releases of the provider until the problem is removed, papermill will be removed from "airflow" extras in the next minor relese of Airflow). The first step of the process is to let the maintainers of the provider library that holds us back, which is happending via this comment. Apparently there is a PR already in your repo to prevent it and add 3.11 support, and there is about a week to before we attempt to merge 3.11 charge with suspended the Apache Hive provider so in case you would like to avoid the suspension, there is a about a week to get the python-sasl release that will support Python 3.11. |
Also announced in airflow devlist https://lists.apache.org/thread/0dcvjj0f6bnjg3mk4zn32stjbxtprb5j so in case you have something to add, comment etc - feel free. |
FYI . We just merged 3.11 support. The Apache.Hive provider has been excluded from Python 3.11 until this issue is fixed. |
Yes but Pyhive is discontinued We likely drop hive integration eventually as there is no other Python SDK for it |
Run unit tests for Python3.11 exclude hive extra from it because of issue cloudera/python-sasl#30
I made a couple of contributions to PyHive which were accepted and released in 0.7.1.dev0. You are requested to test with the dev version and report any bugs in PyHive github repository before 0.7.1 is released in a month or so.
|
Fantastic! Thanks ! That will be great to have it released and add Hive back to be supported in 3.11 If all goes right, this will both, test installability, cross-dependencies of PyHive and pip resolvabilty on 3.11 as well as run unit tests for it, which should give us good confidence that it works. Fingers crossed it wil be green Re:
is pure-sasl workig with whole range Py 3.8 - 3.11 ? Because if so, I would gladly switch to pure-sasl instead for Airflow we already use pure-sasl so . I will run another PR to check it after. We had our dose of problems with sasl (see below comment):
That's a good timing. We are completling rewriting all our queries to SQLAlchemy 2.0 too, so having Hive supporting it too, would remove a blocker for our migration. |
@potiuk |
Cool. Updated the PR with switching to pure-sasl as well. Let's see :) |
This is awesome! |
Looks like we are close enough - we had some intermittent issues on apache/airflow#32607 - re-run it to see but images for 3.11 were built, and tests passed, so looks like things are good from Airlflow side (I re-run the failing jobs tests to verify). |
Awesome! Thanks for testing @potiuk. |
Yep. it looks good. The constraint job is failing because we have dev dependency and == there, once the package is released and we switch it to regular >=, it will start working. |
Are there any expectations on 0.7.1 to be released :) ? ( |
* Run test for Python3.11 Run unit tests for Python3.11 exclude hive extra from it because of issue cloudera/python-sasl#30
The PyHive repo maintainer with whom I am co-ordinating, is on vacation. He will release 0.7.1 once he is back from vacation. It will take a couple of weeks more. |
OK. Cool. Thanks. We just |
PyHive 0.7.0 is released which includes Python 3.11 and SQLAlchemy 2.0 support. |
Oh... Fantastic. Thank you for letting me know. I kept on checking from time to time but being notified is even cooler :) |
Currently when you try to install sasl on Python 3.11, the compilation fails with:
I know it is eaarly (Python 3.11 has just been released yesterday) but we are hoping in Apache Airflow to a much faster cycle of adding new Python releases - especially that Pyhon 3.11 introduces huge performance improvements (25% is the average number claimed) due to a very focused effort to increase single-threaded Python performance (Specialized interpreter being the core of it but also many other improvements) without actually changing any of the Python code.
I'd appreciate if someone in the cloudera team attempted to fix it. Otherwise we might want to simply skip
hive provider
from Python 3.11 compatible version of Airflow.I just opened a PR in Apache Airflow yesterday and plan to keep it open until it gets green :). So far I have to exclude hive provider.
apache/airflow#27264
I think it would be fantastic if we could as the open source community migrate to the new Python much faster.
Looking forward to cooperation on that one :)
The text was updated successfully, but these errors were encountered: