Replies: 6 comments
-
|
i think what <@714696027176960101> shared earlier here is a good example
let us know if it works! 🙂 |
Beta Was this translation helpful? Give feedback.
-
|
Hi, thank you for your support. I have tried your method, and unfortunately, it is still return the same old problem This is the response of my /v1/models API Endpoint { Am i missing something? Other way that i have tried:
|
Beta Was this translation helpful? Give feedback.
-
|
Best to check what the proper model name is on the LiteLLM side. When LLM_PROVIDER is set to "custom" we just forward requests with the configuration provided to LiteLLM You can check their docs here: https://docs.litellm.ai/docs/providers/vllm |
Beta Was this translation helpful? Give feedback.
-
|
it might be something like LLM_MODEL=hosted_vllm/google/gemma-3-12b-it |
Beta Was this translation helpful? Give feedback.
-
|
it still not working after all. maybe i will switch to use litellm directly, thank you all for your time |
Beta Was this translation helpful? Give feedback.
-
|
I have re-checked my .env, so after using provider as "custom" and model as "hosted_vllm/<model_name>", its finally works ! |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
=moved your question here <@402399659697373187>
Hi everyone. I have recently test Cognee with vLLM and HuggingFace embedding model, here is my .env file:
LLM_PROVIDER=custom
LLM_MODEL=hosted_vllm/gemma-3-12b
LLM_ENDPOINT=https://vllm_endpoint/v1
LLM_API_KEY=.
EMBEDDING_PROVIDER=custom
EMBEDDING_MODEL=intfloat/multilingual-e5-base
EMBEDDING_ENDPOINT=http://embedding_endpoint/v1
EMBEDDING_DIMENSIONS=768
Here is the error:
Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=gemma-3-12b\n Pass model as E.g. For 'Huggingface' inference endpoints pass in completion(model='huggingface/starcoder',..) Learn more: https://docs.litellm.ai/docs/providers/n\n\n
Is this the source code problem or it is my own .env setup problem?
This discussion was automatically pulled from Discord.
Beta Was this translation helpful? Give feedback.
All reactions