Skip to content

Commit f4ab067

Browse files
Updates for doc week
1 parent b281b9e commit f4ab067

File tree

4 files changed

+295
-290
lines changed

4 files changed

+295
-290
lines changed

data-skipping.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -30,8 +30,8 @@ Three index types are supported:
3030

3131
| Index type | Description | Applicable to predicates in WHERE clauses | Column types |
3232
|------------|--------------|--------------|--------------|
33-
| MinMax |Stores minimum and maximum values for a column | <,<=,=,>=,> | All types except for complex types. See [Supported Spark SQL data types](https://spark.apache.org/docs/latest/sql-reference.html#data-types). |
34-
| ValueList | Stores the list of unique values for the column | =,IN,LIKE | All types except for complex types. See [Supported Spark SQL data types](https://spark.apache.org/docs/latest/sql-reference.html#data-types).|
33+
| MinMax |Stores minimum and maximum values for a column | <,<=,=,>=,> | All types except for complex types. See [Supported Spark SQL data types](https://spark.apache.org/docs/2.4.4/sql-reference.html#data-types). |
34+
| ValueList | Stores the list of unique values for the column | =,IN,LIKE | All types except for complex types. See [Supported Spark SQL data types](https://spark.apache.org/docs/2.4.4/sql-reference.html#data-types).|
3535
| BloomFilter | Uses a bloom filter to test for set membership | =,IN | Byte, String, Long, Integer, Short |
3636

3737
You should use bloom filters for columns with very high cardinality. Index creation invokes a Spark job which writes metadata (indexes) to a user specified location, in Parquet format.

getting-started.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ subcollection: AnalyticsEngine
3434

3535
## Run applications on the cluster
3636

37-
* Watch a [demo](https://developer.ibm.com/clouddataservices/docs/analytics-engine/get-started/#spark-notebook) and run through the tutorial using sample code and data. Copy or download this [notebook](https://datascience.ibm.com/exchange/public/entry/view/e2e70feb00a65760eb1bd683da285364) from the {{site.data.keyword.DSX_short}} community to try it for yourself!
37+
* Watch a [demo](https://developer.ibm.com/clouddataservices/docs/analytics-engine/get-started/#spark-notebook) and run through the tutorial using sample code and data. Copy or download this [notebook](https://dataplatform.cloud.ibm.com/exchange/public/entry/view/e2e70feb00a65760eb1bd683da285364) from the {{site.data.keyword.DSX_short}} community to try it for yourself!
3838
* Learn how to use [spark-submit](https://developer.ibm.com/clouddataservices/docs/analytics-engine/get-started/#spark-submit) to execute a Python script on an {{site.data.keyword.iae_full_notm}} cluster.
3939
* Learn how to programmatically use {{site.data.keyword.iae_full_notm}} through this [tutorial](https://github.com/IBM-Cloud/IBM-Analytics-Engine). Get access to sample scripts to start operationalizing your first applications.
4040
* Get answers to some [frequently asked questions](/docs/AnalyticsEngine?topic=AnalyticsEngine-general-faqs) about using {{site.data.keyword.iae_full_notm}}.

0 commit comments

Comments
 (0)