Simon Digilov
03/26/2025, 9:42 AMVapi Support Bot
03/26/2025, 9:44 AMsilenceTimeoutSeconds
- responseDelaySeconds
- llmRequestDelaySeconds
- startSpeakingPlan.waitSeconds
(try setting to 0.8)
- smartEndpointingEnabled
(set to false
if using non-English transcriptions)
- transcriber.endpointing
(set to 300 for Deepgram)
3. **Resource Utilization**: Monitor the server or service hosting your voice application for CPU, memory, and bandwidth usage.
## Using Your Call IDs for DebuggingVapi Support Bot
03/26/2025, 9:44 AMVapi Support Bot
03/26/2025, 9:44 AMSimon Digilov
03/26/2025, 9:47 AMSimon Digilov
03/26/2025, 9:48 AMVapi Support Bot
03/26/2025, 10:46 AMsmartEndpointingEnabled
setting**:
- Set smartEndpointingEnabled
to false
, especially if you're using non-English transcriptions
- According to [troubleshooting documentation](https://discord.com/channels/1211482211119796234/1211483291191083018/threads/1332286268805087272), this can significantly reduce endpointing latency
2. **Modify transcriber endpointing value**:
- Set transcriber.endpointing
to 300ms (especially if using Deepgram)
- This is recommended in [latency optimization guides](https://discord.com/channels/1211482211119796234/1211483291191083018/threads/1331035264751112202)
3. **Adjust startSpeakingPlan
settings**:
- Set startSpeakingPlan.waitSeconds
to 0.8
- This can help reduce the perceived latency according to [best practices](https://discord.com/channels/1211482211119796234/1211483291191083018/threads/1331035264751112202)
4. **Review other timing parameters**:
- Check silenceTimeoutSeconds
, responseDelaySeconds
, and llmRequestDelaySeconds
settings
- These can all impact the overall responsiveness of your agent
If you're still experiencing issues after making these changes, would you be able to share which transcription service you're using and any other configuration details? That would help me provide more targeted advice!
Sources:
- [Troubleshooting sudden latency increase in Vapi AI agent](https://discord.com/channels/1211482211119796234/1211483291191083018/threads/1332286268805087272)Vapi Support Bot
03/26/2025, 10:46 AMSahil
03/27/2025, 12:28 AMcurrent
"startSpeakingPlan": { "waitSeconds": 0, "transcriptionEndpointingPlan": { "onPunctuationSeconds": 0.1, "onNoPunctuationSeconds": 1.5, "onNumberSeconds": 0.5 }, "customEndpointingRules": \[\], "smartEndpointingPlan": { "provider": "livekit", "waitFunction": "200 + 8000 \* x" } }
next
"startSpeakingPlan": { "waitSeconds": 0.6, "smartEndpointingPlan": { "provider": "livekit"} }
Hey Simon, set your start speaking plan wait seconds to 0.6, smart endpointing plan to LiveKit, and remove the transcription endpointing plan. The latency you're observing is all because of endpointing at first glance, which means it's taking time to recognize if the user has stopped speaking.
Do these changes to your assistant, and then let me know how it goes for you