Port of OpenAI's Whisper model in C/C++
-
Updated
Jan 21, 2025 - C++
Port of OpenAI's Whisper model in C/C++
MNN is a blazing fast, lightweight deep learning framework, battle-tested by business-critical use cases in Alibaba. Full multimodal LLM Android App:[MNN-LLM-Android](./project/android/apps/MnnLlmApp/README.md)
Transformer related optimization, including BERT, GPT
LightSeq: A High Performance Library for Sequence Processing and Generation
🍅🍅🍅YOLOv5-Lite: Evolved from yolov5 and the size of model is only 900+kb (int8) and 1.7M (fp16). Reach 15 FPS on the Raspberry Pi 4B~
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.
an open-source implementation of sequence-to-sequence based speech processing engine
Whisper Dart is a cross platform library for dart and flutter that allows converting audio to text / speech to text / inference from Open AI models
Fast implementation of BERT inference directly on NVIDIA (CUDA, CUBLAS) and Intel MKL
optimized BERT transformer inference on NVIDIA GPU. https://arxiv.org/abs/2210.03052
Running BERT without Padding
A high-performance inference system for large language models, designed for production environments.
A Fast Neural Machine Translation System developed in C++.
SCXML interpreter and transformer/compiler written in C/C++ with bindings to Java, C#, Python and Lua
Explore LLM model deployment based on AXera's AI chips
LM inference server implementation based on *.cpp.
An Implementation of Transformer (Attention Is All You Need) in DyNet
Swin Transformer C++ Implementation
使用ONNXRuntime部署LSTR基于Transformer的端到端实时车道线检测,包含C++和Python两个版本的程序
Add a description, image, and links to the transformer topic page so that developers can more easily learn about it.
To associate your repository with the transformer topic, visit your repo's landing page and select "manage topics."