Skip to content

Commit

Permalink
[MINOR][DOCS] Document 'without' value for HADOOP_VERSION in pip inst…
Browse files Browse the repository at this point in the history
…allation

### What changes were proposed in this pull request?

I believe it's self-descriptive.

### Why are the changes needed?

To document supported features.

### Does this PR introduce _any_ user-facing change?

Yes, the docs are updated. It's master only.

### How was this patch tested?

Manually built the docs via `cd python/docs` and `make clean html`:

![Screen Shot 2020-11-20 at 10 59 07 AM](https://user-images.githubusercontent.com/6477701/99748225-7ad9b280-2b1f-11eb-86fd-165012b1bb7c.png)

Closes #30436 from HyukjinKwon/minor-doc-fix.

Authored-by: HyukjinKwon <gurwls223@apache.org>
Signed-off-by: HyukjinKwon <gurwls223@apache.org>
  • Loading branch information
HyukjinKwon committed Nov 20, 2020
1 parent 883a213 commit 02d410a
Showing 1 changed file with 8 additions and 3 deletions.
11 changes: 8 additions & 3 deletions python/docs/source/getting_started/install.rst
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ If you want to install extra dependencies for a specific componenet, you can ins
pip install pyspark[sql]
For PySpark with a different Hadoop version, you can install it by using ``HADOOP_VERSION`` environment variables as below:
For PySpark with/without a specific Hadoop version, you can install it by using ``HADOOP_VERSION`` environment variables as below:

.. code-block:: bash
Expand All @@ -68,8 +68,13 @@ It is recommended to use ``-v`` option in ``pip`` to track the installation and
HADOOP_VERSION=2.7 pip install pyspark -v
Supported versions of Hadoop are ``HADOOP_VERSION=2.7`` and ``HADOOP_VERSION=3.2`` (default).
Note that this installation of PySpark with a different version of Hadoop is experimental. It can change or be removed between minor releases.
Supported values in ``HADOOP_VERSION`` are:

- ``without``: Spark pre-built with user-provided Apache Hadoop
- ``2.7``: Spark pre-built for Apache Hadoop 2.7
- ``3.2``: Spark pre-built for Apache Hadoop 3.2 and later (default)

Note that this installation way of PySpark with/without a specific Hadoop version is experimental. It can change or be removed between minor releases.


Using Conda
Expand Down

0 comments on commit 02d410a

Please sign in to comment.