Skip to content

fluid support asynchronous training #9941

Closed
@jacquesqiao

Description

@jacquesqiao

Project

https://github.com/PaddlePaddle/Paddle/projects/61

Design

Operators

Transpiler #9997

  • dist transpile async trainer program. Do not need to add .trainer_n suffix to gradient block in async mode.
  • dist transpile async pserver program. Do not need to aggregate gradient block.

Consider

  • need to consider how to add learning rate decay in asynchronous training. Do we need lr_decay?

Benchmark

Metadata

Metadata

Labels

No labels
No labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions