Skip to content

Feature Request: add LazyPromptDataset to DeepSpeedChat #450

Open
@0xDing

Description

Currently, in our multi-GPU training process, every GPU process loads the train and evaluation datasets using the torch.load function. However, when the dataset is large, this can lead to an issue with CPU memory constraints.

To address this issue, I would like to suggest introducing the LazyPromptDataset functionality. This functionality will enable us to load the data on demand instead of pre-loading it into memory, which will save memory and improve overall performance.

Activity

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Metadata

Assignees

Labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions