Skip to content

Slower inference performance when switching from torchxla_trace_once to openxla compile backend #5430

Open
@GleasonK

Description

@GleasonK

🐛 Bug

It looks like torchxla_trace_once is deprecated in favor of openxla, but when I tried to make that migration in some benchmark testing I saw a new warning message and some performance regressions. This was found when running an inference benchmark from openxla-benchmark - ResNet on GPU.

To Reproduce

Colab repro.

Steps to reproduce the behavior:

  1. Run colab with torchxla_trace_once - should dump files.
  2. Run colab with openxla - should dump files (restart runtime if it does not)

Hopefully that provides enough information to be useful, if not I am happy to help further.

Expected behavior

On-par performance and HLO graph generation between the two backends (openxla and torchxla_trace_once).

Environment

  • Reproducible on XLA backend [CPU/TPU]: GPU
  • torch_xla version: nightly build (8/9/23)

Additional context

Output traces: save_ir.zip

Activity

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions