Skip to content

Conversation

winglian
Copy link
Collaborator

fixes #1753

@winglian winglian force-pushed the llama-pretrain-fix branch from 6e80490 to 9466e68 Compare July 17, 2024 11:18
@winglian winglian merged commit 976f851 into main Jul 17, 2024
@winglian winglian deleted the llama-pretrain-fix branch July 17, 2024 14:58
djsaunde pushed a commit that referenced this pull request Dec 17, 2024
* fixes to accelerator so that iterable pretraining datasets work

* fix the pretraining test params

* split batches, not dispatch batches needs to be set

* update c4 datasets

* set epochs in pretrain config test

* need to set both split_batches and dispatch_batches to false for pretraining

* fix bool val in comment
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

TinyLlama pretrain fails, but SFT works -- CUDA error: an illegal memory access was encountered
1 participant