-
Notifications
You must be signed in to change notification settings - Fork 179
Description
Hello Team,
I would like to request support for integrating the new voyage-4-nano model into the repository/service.
The model appears to be lightweight and optimized for text embeddings, which would be highly beneficial for use cases requiring low-latency and cost-effective embedding generation, such as semantic search, retrieval, and RAG-based applications.
Request Details:
Model Link: https://huggingface.co/voyageai/voyage-4-nano
-
Please add support for
voyage-4-nanoin the model configuration/interface. -
If applicable, provide example usage in:
- Python
- API / SDK
- Any relevant framework integrations (e.g., LangChain, TEI, or similar)
-
Clarify any recommended settings (batch size, dimensions, latency expectations, etc.).
Use Case (Optional):
We intend to use this model for:
- Document retrieval
- Semantic search
- Embedding-based ranking in production systems
Please let me know if any additional details are needed from my side.
Thanks in advance for your support!
Open source status & huggingface transformers.
- The model implementation is available on transformers
- The model weights are available on huggingface-hub
- I verified that the model is currently not running in the latest version
pip install infinity_emb[all] --upgrade - I made the authors of the model aware that I want to use it with infinity_emb & check if they are aware of the issue.