Skip to content

Commit 92e4b22

Browse files
committed
typos
1 parent 0da7b73 commit 92e4b22

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

_posts/2025-05-05-imbalanced-learning.markdown

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -57,7 +57,7 @@ Later, we'll ponder *how* to choose the tradeoff. But to do that we first need t
5757

5858
Like I previously mentioned, initial modeling stages try to give us the best tradeoff curve possible for the task - using data, model type, training techniques, whatever. At those stages we can optimize for threshold-independent metrics, for example various [area under the curve](https://developers.google.com/machine-learning/crash-course/classification/roc-and-auc) metrics. But ultimately, somewhere downstream the model's output will be binarized, and we might as well take that into consideration when tuning the model.
5959

60-
I'm personally fond of the [F-score](https://en.wikipedia.org/wiki/F-score) - it combines two very interpretable metrics (precision and recall), which makes cummunicating with less technical stakeholders (such as product managers and the FDA) easier, and can be [easily tweaked to account for error type preferences](/2019/09/30/f-score-deep-dive.html).
60+
I'm personally fond of the [F-score](https://en.wikipedia.org/wiki/F-score) - it combines two very interpretable metrics ([precision and recall](https://en.wikipedia.org/wiki/Precision_and_recall)), which makes communicating with less technical stakeholders (such as product managers and the FDA) easier, and can be [easily tweaked to account for error type preferences](/2019/09/30/f-score-deep-dive.html).
6161

6262
For this problem precision and recall were equally important, so I used the $$ F_1 $$ score:
6363

@@ -75,7 +75,7 @@ $$ BCE = -\sum_i{\alpha y_i \log_2{\hat{y_i}} + (1 - y_i) \log_2{(1 - \hat{y_i}}
7575

7676
## Optimizing Away
7777

78-
Now that all the actors are on stage, let's roll our sleeves and get our hands dirty.
78+
Now that all the actors are on stage, let's roll up our sleeves and get our hands dirty.
7979

8080
First, we'll take a step back from looking at individual instances, and look at the relationship between positive and negative instances, based on $$ \beta $$, the positive class prevalence. To that end, we'll replace individual $$ \hat{y_i} $$ and $$ 1 - \hat{y_i} $$ with their respective expectations, $$ x $$ and $$ z $$. Further, our assumed tradeoff curve constrains one by the other; let's reframe the loss as a function of $$ x $$ (also dependent on $$ \alpha $$, $$ \beta $$, and $$ p $$.)
8181

0 commit comments

Comments
 (0)