Replies: 1 comment 2 replies
-
So can I export a non-quantized tflite model at the moment? |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello all,
PR #1921 is making some significant changes to which model types are supported by Coqui STT packages. The gist of the PR is that only TFLite model inputs will be supported. This means model files ending in
.pb
or.pbmm
will not be loaded by 🐸STT after this PR and follow-up PRs are merged and the next version is released. If you're using a currently released version of 🐸STT, up until 0.10.0-alpha.10, nothing will change for you until you upgrade.We're making this change because the protobuf (and memory-mapped protobuf) inputs are hacky at best (with TensorFlow v1) and deprecated/unsupported (with TensorFlow v2). They increase our maintenance burden, slow down builds (which slows down CI, which slows down development), force us to have a hacky build system integration into TensorFlow, and so on. On top of this, 🐸STT packages have always been focused on on-device, low latency speech recognition, with excellent support for low power devices with constrained resources. This means that the TFLite backend has also always been the better choice for how to deploy 🐸STT.
Here is a summary of main changes:
.tflite
models).stt
,stt-tflite
,stt-gpu
), but just a singlestt
package..pb
or.pbmm
) model will result in an error on the API level, which you can detect and handle.This change will bring lots of benefits to maintainability and speed of development for 🐸STT, and will also make it a bit easier for us to upgrade to TensorFlow 2, a long-standing issue which has been made harder to tackle due to supporting protobuf models, and finally it should also make it easier to contribute to the code base.
If you have any questions or concerns, or if you think you have a use-case that can only be covered by the protobuf model formats, please post here so we can find a solution.
Best,
-- reuben
Q/A section:
coqui_stt_ctcdecoder
package to implement GPU-accelerated inference directly from model checkpoints. We have two places in our training code doing this, which you can use for inspiration: here and here. Improvements to make checkpoints simpler to load from an API/CLI/configuration perspective are more than welcome.train.py
) for training and also use the Python code for inference (e.g.evaluate.py
), without ever creating a protobuf (.pb
or.pbmm
) model. If you don't use protobuf files, then you probably only work with checkpoints, and these changes shouldn't affect you.Beta Was this translation helpful? Give feedback.
All reactions