Skip to content

Conversation

@nipung90
Copy link
Contributor

Summary: Backout diff for D87488015 in case it causes compatibility issues again.

Differential Revision: D87650960

…ipeline class constructors (meta-pytorch#3565)

Summary:

This diff enables the static logging functionality to collect data for:
plan() - This will allow us to log the inputs and outputs to the planner to help with use issue debugging
ShardEstimators - This will allow us to log the inputs and outputs to the ShardEstimators, which gives us the bandwidth inputs to verify if the planner is generating expected values as well as help with debugging OOMs
TrainingPipeline - The class type here will be an indicator of which pipeline was used by the training job. The training pipeline has implications on the memory usage and is an important data point to collect to investigate OOMs.

Reviewed By: kausv

Differential Revision: D87488015
… TrainingPipeline class constructors"

Summary: Backout diff for D87488015 in case it causes compatibility issues again.

Differential Revision: D87650960
@meta-codesync
Copy link
Contributor

meta-codesync bot commented Nov 21, 2025

@nipung90 has exported this pull request. If you are a Meta employee, you can view the originating Diff in D87650960.

@meta-cla meta-cla bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Nov 21, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported meta-exported

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant