Skip to content

Multi-card GPUs can't train occupancy will be 100 card masters, single GPUs won't. #211

@Lxy811

Description

@Lxy811

Reported an error during training, video card 100%, set batch-size high enough to run normally, error as follows:
Snipaste_2024-04-29_12-04-48
First Precls training. At first set bachsize=16 will be stuck in the training before the start, and then set bachsize=32, two GPUs can be trained, but training 5000 and 7600 and so on halfway is stuck. But single card can train normally.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions