Use Case
I'd like to be able to easily switch between the providers and models and specify a different BaseURL for Embedding, Reranker, and LLM models.
Problem Statement
I could not specify a proxy to cache my model requests. Such as LiteLLM Proxy or CloudFlare AI Gateway proxy.
How This Feature Would Help
Saves money. Improves speed for cached requests.
Proposed Solution
Implement LiteLLM Client SDK so I could set BASE_URL and switch easily between models and providers.
Alternatives Considered
No response
Priority
Important - affects my workflow
Additional Context
No response
Checklist