From 24173f7023088d48a1a81765029aedd632316d56 Mon Sep 17 00:00:00 2001 From: Yura52 Date: Sat, 9 Dec 2023 21:09:40 +0300 Subject: [PATCH] Misc. improvements in package/README --- package/README.md | 13 ++++++++----- 1 file changed, 8 insertions(+), 5 deletions(-) diff --git a/package/README.md b/package/README.md index b328ec51..8a5b77fc 100644 --- a/package/README.md +++ b/package/README.md @@ -126,9 +126,12 @@ y_pred = model_with_embeddings(x) ``` In other words, the whole paper is about the fact that having such a thing as -`m_cont_embeddings` can (significantly) improve the downstream performance, -and the paper showcases three types of such embeddings: -simple, periodic and piecewise-linear. +`m_cont_embeddings` can (significantly) improve the downstream performance. + +**The paper showcases three types of such embeddings**: +- [Simple](#simple-embeddings) +- [Periodic](#simple-embeddings) +- [Piecewise-linear](#piecewise-linear-encoding--embeddings) ## Simple embeddings @@ -266,7 +269,7 @@ y_pred = model(x) | `QLR` / `TLR` | `ReLU(Linear(ple(x_i)))` | `PiecewiseLinearEmbeddings(bins, activation=True)` | In the above table: -- Q/T ~ quantiles-/tree- based bins, L ~ Linear, R ~ ReLU. +- Q ~ quantiles-based bins, T ~ tree-based bins, L ~ Linear, R ~ ReLU. - `x_i` is the i-th scalar continuous feature. - `ple` stands for "Piecewise-linear encoding". @@ -284,7 +287,7 @@ In the above table: **Hyperparameters** - For `PiecewiseLinearEmbeddings`, - reasonable starting points are `d_embedding=8, activation=False` + possible starting points are `d_embedding=8, activation=False` or `d_embedding=24, activation=True`. - See other notes on hyperparameters in ["Practical notes"](#practical-notes).