-
Notifications
You must be signed in to change notification settings - Fork 88
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GPU memory issues when composing some of the waveform augmentations #132
Comments
Hi. That's curious! The impulse response transform had almost no changes between 0.9.0 and 0.10.1 🤔 |
Do you init your transforms once and then use them many times or do you init them every time you need to run them? |
Hi! I'm running into a similar problem, but only when training on multiple GPUS. I use pytorch lightning, It'll take some time but I will try to create a script to reproduce the problem. Are you also using multiple GPUs @luisfvc ? |
Hi, I've noticed the same problem with HPF and LPF. I'm only training on a single GPU, but found that it only occurs if I'm using multiprocessing in my dataloader (i.e. num_workers > 0). Could it be related to pytorch/pytorch#13246 (comment)? That's what I thought I was debugging until I realized these filters were the real culprit |
Thanks, that comment helps us getting closer to reproducing the bug |
I have the exact same experience. Had to set num_workers=0 when using torch-audiomentations. Curious if we found a better solution? |
Thanks RoyJames :) Just so I understand your way of using torch-audiomentations, I'd like to know: Did you run the transforms on CPU (in each data loader worker)? And did you train the ML model on GPU? I have added a "Known issues" section to readme now, by the way: https://github.com/asteroid-team/torch-audiomentations#known-issues |
I should write this article soon, to make it easier to decide if torch-audiomentations is a good fit and how. Also, it would be swell if someone/we could reproduce and fix this memory leak 😅 I don't have a lot of spare time to do it right now, but I'd love to help |
I think (hope) I did those augmentations on the GPU since the incoming data is already on CUDA. I wrapped torch-audiomentations functions in a preprocessor class that was used as the collate function of my dataloader. While I can't provide a complete code snippet, it is something like: class MyPreprocessor:
def __init__(self, noise_set: Path, device: str = "cuda"):
self._augmentor = Compose(
transforms=[
Gain(
min_gain_in_db=-15.0,
max_gain_in_db=5.0,
p=0.5,
p_mode="per_example",
),
LowPassFilter(
min_cutoff_freq=4000.0,
max_cutoff_freq=8000.0,
p=0.5,
p_mode="per_example",
),
AddBackgroundNoise(
background_paths=noise_set,
min_snr_in_db=0.0,
max_snr_in_db=30.0,
p=0.5,
p_mode="per_example",
)
]
)
def __call__(self, batch: T.List[np.ndarray])
AudioPair = namedtuple('AudioPair', ['clean', 'noisy'])
batch_pairs = [AudioPair(pair[0], pair[1]) for pair in batch]
batch_pairs = torch.utils.data.dataloader.default_collate(batch_pairs)
y = batch_pairs.clean.unsqueeze(1).to(self._device)
x = batch_pairs.noisy.unsqueeze(1).to(self._device)
x = self._augmentor(x, sample_rate=SAMPLE_RATE)
return x, y Then my dataloader looks like: self.train_loader = torch.utils.data.DataLoader(
self.train_set,
sampler=train_sampler,
collate_fn=MyPreprocessor(noise_set=noise_set, device="cuda"),
batch_size=BATCH_SIZE,
drop_last=True,
num_workers=num_workers,
shuffle=train_shuffle,
worker_init_fn=seed_worker,
) and I had to set Edit: I forgot to mention that I use the above with |
I was able to use Maybe this is obvious to some experienced folks. I feel we could mention this caveat to other unaware users? |
Yes, this is the way I use torch-audiomentations on GPU too 👍 It would indeed be nice to have this documented well. I'm currently focusing on the documentation website for audiomentations, but I want to eventually make one for torch-audiomentations too, using the knowledge I gained for making the audiomentations documentation |
I implemented it in the same way and applied it in the training loop, but I'm still experiencing the memory leak. I've got a GPU Server with multiple GPUs and I am using pytorch lightning with DDP. I'm using only one GPU per process. |
Hi, I have been experiencing some memory problems when using some of the transforms on the GPU. When I apply the low or high pass filtering, the memory usage of my GPU increases each training iteration. And since I updated from v0.9.0 to the latest release, the same happens with the impulse response transform. This does not happen when I compose other transforms, like polarity inversion, gain, noise or pitch shift.
Any ideas on why this is happening? I went through the package source code but couldn't spot any bug.
Thanks & regards
The text was updated successfully, but these errors were encountered: