Skip to content

[AMP] CUDA support for mixed precision pass #8294

Closed
@AndrewZhaoLuo

Description

Solve issues and make modifications to support CUDA for mixed precision pass here: #8069

Current initial issues as described by @Lunderberg

On the cuda side, it's failing a check that requires 16-bit floats to be used in pairs.

Check failed: lanes % 2 == 0 (1 vs. 0) : only support even lane for half type

This issue is completed when unit tests can pass for CUDA target.

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions