Knowledge Base in Local Machine
# support
c
I am integrating and demoing some voice enabled solutions to clients so far with much success, even though I am not highly technical I have managed to work my way connecting Vapi with Google Calendar, Cal, Make, etc. However I am facing a technical challenge well above my weight: An important prospect loved the solution however he strictly needs, for policy reasons, the Knowledge Base to reside in a local machine. Is there a way to do that? if so how? is there any documentation I could refer to? thanks!!
v
You'll need to set up a local server that can receive and respond to knowledge base requests. VAPI will send requests to your server with the conversation context Your server needs to return either: a) Relevant documents with similarity scores b) A direct response for the assistant to speak
Implementation Steps: a. Create a local HTTP server that accepts POST requests at an endpoint b. Configure the server to expect requests in this format:
Copy code
{
  "message": {
    "type": "knowledge-base-request",
    "messages": [
      {
        "role": "user",
        "content": "What are your business hours?"
      }
    ]
  }
}
c. Return responses in these formats:
Copy code
{
  "documents": [
    {
      "content": "Our business hours are 9am-5pm Monday through Friday",
      "similarity": 0.95
    }
  ]
}
Or alternatively, provide a direct response:
Copy code
{
  "message": {
    "role": "assistant",
    "content": "Your direct response here"
  }
}
Copy code
{
  "model": {
    "knowledgeBase": {
      "provider": "custom-knowledge-base",
      "server": {
        "url": "https://your-local-server.com/knowledge-base"
      }
    }
  }
}
Update your assistant configuration using the VAPI API with this structure.
j
Couldn’t this be made with a ChatGPT module in make.com and webhook module?
ping when u respond ❤️
s
I don't know how it works. so cannot make a comment on that part.
j
Ok