Why can I not see all the OpenAI Models available ...
# support
n
If I hit the openai model dropdown I only see 2 cluster options. I have my openai api key under providers. Why do I not see the other models?
v
Hi @Nathan Hull, I'm an AI assistant for Vapi. I'll analyze available sources and get back to you shortly.
----------------------------- **While I generate an answer, here are some sources that might be helpful**: * [Provider Keys]() * [Web Calling]() * [Custom LLM]() * [Dashboard]() -----------------------------
### Answer If the OpenAI models are not appearing correctly in the assistant options despite having your OpenAI API key set under providers, it might be due to a few reasons: 1. **Issue with API Key**: - Ensure that the API key you provided is correct and valid. Invalid keys may not allow access to the models. You can check for errors in the network tab as mentioned in the Discord forum [(1)]() . 2. **Supported Models**: - Vapi supports using any OpenAI-compatible endpoint. Ensure you are specifying the correct parameters when setting up the assistant. For example, your assistant setup might look like this:
Copy code
json
    {
      "name": "My Assistant",
      "model": {
        "provider": "openrouter",
        "model": "gpt-4",
        "messages": [
          {
            "role": "system",
            "content": "You are an assistant."
          }
        ],
        "temperature": 0.7
      }
    }
[(2)]() , [(3)]() 3. **Custom LLMs**: - If you are using custom LLMs or different providers, make sure they are correctly set up and the APIs endpoints are reachable. More details can be found in documents about [Custom LLMs]() [(2)]() . Ensure you have followed the setup instructions properly and double-check the API keys and configurations.
n
I had added my OpenAI key. But I only see the 3.5 and 4 "clusters" as options...
v
Got it. Tagging @User for additional help.
s
Because of intelligent model switching. In case of model failure, we will switch to another model to prevent it from call to crash. If you want to access all the models then use API
n
Hi Sahil, can I not use the api directly in the assistants config dashboard? How do I switch? In your GPT-4o update video it looks like it can be selected in the UI?
s
GPT-4o is accessible directly via dashboard and for other thing you need to make a HTTP request to the endpoint in order to use it.
@Sahil
s
Can you remove your openai key and retry it once?
you should see gpt-4o as well
n
I do see the 4o cluster. However when testing on the 4o the built in end call function seems to be broken. If I switch back to 4 the end call function works and the assistant will hang up but when the conversation is over use 4o the assistant does not terminate the call.
@Sahil I assume the built in functions not working is a bug?
s
No, it is working correctly. The issue is that the LLM model has control over it and is making the wrong tool request, which is causing the problem. I would suggest you use endCallPhrases instead of endCallFunction. This way, you will be able to minimize it.
n
@User Is there a way for me to see those logs on the VAPI side?
3 Views