Skip to content

Commit 14c0a37

Browse files
authored
[DOC] Notebook on Feature-based Clustering (aeon-toolkit#2579)
* Feature-based clustering * Feature-based clustering update * Update clustering overview * formatting
1 parent c216cce commit 14c0a37

File tree

2 files changed

+1526
-32
lines changed

2 files changed

+1526
-32
lines changed

examples/clustering/clustering.ipynb

+37-32
Original file line numberDiff line numberDiff line change
@@ -2,6 +2,12 @@
22
"cells": [
33
{
44
"cell_type": "markdown",
5+
"metadata": {
6+
"collapsed": false,
7+
"jupyter": {
8+
"outputs_hidden": false
9+
}
10+
},
511
"source": [
612
"# Time Series Clustering\n",
713
"\n",
@@ -23,45 +29,47 @@
2329
"erative [16], Feature K-means [17], Feature K-medoids [17], U-shapelets [18],\n",
2430
"USSL [19], RSFS [20], NDFS [21], Deep learning and dimensionality reduction\n",
2531
"approaches see [22]"
26-
],
27-
"metadata": {
28-
"collapsed": false
29-
}
32+
]
3033
},
3134
{
3235
"cell_type": "markdown",
36+
"metadata": {
37+
"collapsed": false,
38+
"jupyter": {
39+
"outputs_hidden": false
40+
}
41+
},
3342
"source": [
3443
"## Clustering notebooks\n",
3544
"\n",
36-
"- `aeon` currently focusses on partition based approaches that use elastic distance\n",
37-
"functions. The [partition based](partitional_clustering.ipynb) note book has an\n",
38-
"overview of the funtionality in aeon.\n",
45+
"`aeon` offers a comprehensive suite of time series clustering (TSCL) algorithms, encompassing partition-based, density-based, hierarchical, deep learning, and feature-based approaches.\n",
46+
"\n",
47+
"- `aeon` has many partition-based clustering algorithms, which include TimeSeriesKMeans, KMedoids, CLARA, CLARANS, ElasticSOM, and KSpectralCentroid, leveraging elastic distance measures like DTW. The [partition-based](partitional_clustering.ipynb) notebook has an overview of the functionality in aeon.\n",
3948
"\n",
40-
"- `sklearn` has *density based* and *hierarchical based* clustering algorithms, and\n",
41-
"these can be used in conjunction with `aeon` elastic distances. See the [sklearn and\n",
42-
"aeon distances](../distances/sklearn_distances.ipynb) notebook.\n",
49+
"- `sklearn` has *density-based* and *hierarchical based* clustering algorithms, which can be used in conjunction with `aeon` elastic distances. See the [sklearn and aeon distances](../distances/sklearn_distances.ipynb) notebook.\n",
4350
"\n",
44-
"- Deep learning based TSCL is a very popular topic, and we are working on bringing\n",
45-
"deep learning functionality to `aeon`, first algorithms for [Deep learning] are\n",
46-
"COMING SOON\n",
51+
"- Bespoke feature-based TSCL algorithms are easily constructed with `aeon` transformers and `sklearn` clusterers in a pipeline. Some examples are in the\n",
52+
"[sklearn clustering]. The [feature-based](feature_based_clustering.ipynb) notebook gives an overview of the feature-based clusterers in an aeon.\n",
4753
"\n",
48-
"- Bespoke feature based TSCL algorithms are easily constructed with `aeon`\n",
49-
"transformers and `sklearn` clusterers in a pipeline. Some examples are in the\n",
50-
"[sklearn clustering]. We will bring the bespoke feature\n",
51-
"based clustering algorithms into `aeon` in the medium term.\n",
54+
"- Deep learning based TSCL is a very popular topic, and we have introduced many deep learning functionalities to `aeon`. Autoencoder-based models like AEFCNClusterer, AEResNetClusterer, and AEDCNNClusterer enable complex pattern discovery.\n",
5255
"\n",
53-
"We are in the process of extending the bake off described in [1] to include all\n",
54-
"clusterers. So far, we find medoids with MSM distance is the best performer.\n",
56+
"- `aeon` also includes averaging-based clustering algorithms, which utilize centroid-based representations of time series.\n",
57+
"\n",
58+
"\n",
59+
"We are in the process of extending the bake-off described in [1] to include all clusterers. So far, we find medoids with MSM distance is the best performer.\n",
5560
"\n",
5661
"<img src=\"img/clst_cd.png\" width=\"600\" alt=\"cd_diag\">\n",
5762
"\n"
58-
],
59-
"metadata": {
60-
"collapsed": false
61-
}
63+
]
6264
},
6365
{
6466
"cell_type": "markdown",
67+
"metadata": {
68+
"collapsed": false,
69+
"jupyter": {
70+
"outputs_hidden": false
71+
}
72+
},
6573
"source": [
6674
"## References\n",
6775
"\n",
@@ -141,31 +149,28 @@
141149
"[22] B. Lafabregue, J. Weber, P. Gancarski, and G. Forestier. End-to-end deep\n",
142150
"representation learning for time series clustering: a comparative study. Data Mining\n",
143151
"and Knowledge Discovery, 36:29—-81, 2022\n"
144-
],
145-
"metadata": {
146-
"collapsed": false
147-
}
152+
]
148153
}
149154
],
150155
"metadata": {
151156
"kernelspec": {
152-
"display_name": "Python 3",
157+
"display_name": "Python 3 (ipykernel)",
153158
"language": "python",
154159
"name": "python3"
155160
},
156161
"language_info": {
157162
"codemirror_mode": {
158163
"name": "ipython",
159-
"version": 2
164+
"version": 3
160165
},
161166
"file_extension": ".py",
162167
"mimetype": "text/x-python",
163168
"name": "python",
164169
"nbconvert_exporter": "python",
165-
"pygments_lexer": "ipython2",
166-
"version": "2.7.6"
170+
"pygments_lexer": "ipython3",
171+
"version": "3.11.10"
167172
}
168173
},
169174
"nbformat": 4,
170-
"nbformat_minor": 0
175+
"nbformat_minor": 4
171176
}

0 commit comments

Comments
 (0)