Open
Description
Feature request
Add support for Mixtral 8x7B: https://mistral.ai/news/mixtral-of-experts/
Motivation
Mixtral 8x7B is a high-quality sparse mixture of experts model (SMoE) with open weights. Licensed under Apache 2.0. Mixtral outperforms Llama 2 70B on most benchmarks with 6x faster inference. It is the strongest open-weight model with a permissive license and the best model overall regarding cost/performance trade-offs. In particular, it matches or outperforms GPT3.5 on most standard benchmarks.
Your contribution
.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment