Pen π»
04/04/2025, 8:03 AMVapi Support Bot
04/04/2025, 8:05 AMstartSpeakingPlan
settings might need adjustment. Setting waitSeconds
to 0.8 and adjusting Deepgram's endpointing value to 300 has helped others with similar issues. According to [troubleshooting information](https://discord.com/channels/1211482211119796234/1211483291191083018/threads/1344177033625997322), this combination can significantly improve response times.
2. **Model-generated responses**: If your assistant is configured to generate the first message using a language model (LLM), this can introduce a delay of approximately three seconds. The configuration firstMessageMode: assistant-speaks-first-with-model-generated-message
might be causing delays as mentioned in [this discussion](https://discord.com/channels/1211482211119796234/1211483291191083018/threads/1339303957675245669).
3. **Using your own key**: If you're using your own key instead of Vapi's, it can add extra latency according to [this source](https://discord.com/channels/1211482211119796234/1211483291191083018/threads/1339303957675245669).
Here are some troubleshooting steps you can try:
1. **Adjust your configuration**:
json
"start_speaking_plan": {
"waitSeconds": 0.8,
"smartEndpointingEnabled": false,
"transcriptionEndpointingPlan": {
"onPunctuationSeconds": 0.5
}
}
2. Consider using a static first message instead of a model-generated one to reduce initial response time.Vapi Support Bot
04/04/2025, 8:05 AMKyle
04/06/2025, 11:50 AMKyle
04/06/2025, 11:56 AMPen π»
04/06/2025, 7:43 PMKyle
04/07/2025, 2:47 PMPen π»
04/08/2025, 11:02 PMKyle
04/10/2025, 10:47 PMPen π»
04/11/2025, 12:19 PMKyle
04/11/2025, 9:50 PMKyle
04/11/2025, 9:51 PMKyle
04/11/2025, 9:51 PM