Closed
Description
support distribute training in python v2 API:
- support sparse remote updater.
- support save parameter in trainer.
How to Test If We Made It
-
Start parameter servers:
sh paddle/demo/quick_start/cluster/pserver.sh
-
Use example
word2vec
to start a distributed training job:In file
paddle/demo/word2vec/api_train_v2.py
, changecluster_train = False
intocluster_train = True
. -
Start word2vec distributed training
sh paddle/demo/word2vec/api_train_v2.py
Metadata
Metadata
Assignees
Labels
No labels