Skip to content
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.

FP16/AMP training #15

Open
Zhylkaaa opened this issue Jan 24, 2023 · 0 comments
Open

FP16/AMP training #15

Zhylkaaa opened this issue Jan 24, 2023 · 0 comments

Comments

@Zhylkaaa
Copy link

Zhylkaaa commented Jan 24, 2023

Hello, I have noticed that all scripts use full precision calculations. Are there reasons for that? Does it make substantial difference in terms of accuracy on downstream tasks? Were there some ablation studies regarding that?

I am wondering because it's easier to fit model with fp16 for university project.

Sorry if this question was answered elsewhere.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant