Feature Request: add LazyPromptDataset to DeepSpeedChat #450
Open
Description
Currently, in our multi-GPU training process, every GPU process loads the train and evaluation datasets using the torch.load function. However, when the dataset is large, this can lead to an issue with CPU memory constraints.
To address this issue, I would like to suggest introducing the LazyPromptDataset functionality. This functionality will enable us to load the data on demand instead of pre-loading it into memory, which will save memory and improve overall performance.
Activity