Skip to content

Deep Speech 2 on PaddlePaddle: Plan & Task Breakdown #44

@xinghai-sun

Description

@xinghai-sun

We are planning to build Deep Speech 2 (DS2) [1], a powerful Automatic Speech Recognition (ASR) engine, on PaddlePaddle. For the first-stage plan, we have the following short-term goals:

  • Release a basic distributed implementation of DS2 on PaddlePaddle.
  • Contribute a chapter of Deep Speech to PaddlePaddle Book.

Intensive system optimization and low-latency inference library (details in [1]) are not yet covered in this first-stage plan.

Tasks

We roughly break down the project into 14 tasks:

  1. Develop an audio data provider:
  2. Create a simplified DS2 model configuration:
  3. Develop to support variable-shaped dense-vector (image) batches of input data.
  4. Develop a new lookahead-row-convolution layer (See [1] for details):
  5. Build KenLM n-gram language model for beam search decoding:
  6. Develop a beam search decoder with CTC + LM + WORDCOUNT:
  7. Develop a Word Error Rate evaluator:
    • update ctc_error_evaluator(CER) to support WER.
  8. Prepare internal dataset for Mandarin (optional):
  9. Create standard DS2 model configuration:
    • With variable-length audio sequences (need Task 3).
    • With unidirectional-GRU + row-convolution (need Task 4).
    • With CTC-LM beam search decoder (need Task 5, 6).
  10. Make it run perfectly on clusters.
  11. Experiments and benchmarking (for accuracy, not efficiency):
    • With public English dataset.
    • With internal (Baidu) Mandarin dataset (optional).
  12. Time profiling and optimization.
  13. Prepare docs.
  14. Prepare PaddlePaddle Book chapter with a simplified version.

Task Dependency

Tasks parallelizable within phases:

Roadmap Description Parallelizable Tasks
Phase I Basic model & components Task 1 ~ Task 8
Phase II Standard model & benchmarking & profiling Task 9 ~ Task 12
Phase III Documentations Task13 ~ Task14

Issue for each task will be created later. Contributions, discussions and comments are all highly appreciated and welcomed!

Possible Future Work

  • Efficiency Improvement
  • Accuracy Improvement
  • Low-latency Inference Library
  • Large-scale benchmarking

References

  1. Dario Amodei, etc., Deep Speech 2 : End-to-End Speech Recognition in English and Mandarin. ICML 2016.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions