barbelos
03/13/2025, 12:24 PM015d659d-5fc1-4301-94f4-df2c2e1ab10d
6b3c50e6-b230-4e12-b470-f091ea13664e
04b68aee-0e91-4ddf-9f28-38125855edf3
https://cdn.discordapp.com/attachments/1349719731883474975/1349719732097388625/Screenshot_2025-03-13_at_13.23.31.png?ex=67d41ff4&is=67d2ce74&hm=0e6083c588ead37881d6170f2fd1239d489ea3379ace89d5fe50bbf3a1e18bd0&Vapi Support Bot
03/13/2025, 12:25 PMbash
curl -X PATCH https://api.vapi.ai/assistant/id \
-H "Authorization: Bearer token" \
-H "Content-Type: application/json" \
-d '{
"messagePlan": {
"idleMessages": [
"Are you still there?",
"Hello, can you hear me?"
],
"idleMessageMaxSpokenCount": 3,
"idleTimeoutSeconds": 8
}
}'
If the issue persists, try these additional troubleshooting steps:
- Review any recent changes to your assistant configuration
- Verify your completion message settings
- Check that your response handling logic is working correctly
The recording interface you shared shows a 20-second recording, which suggests the call is connecting but possibly not handling the conversation flow correctly.
Source:
- [Troubleshooting Vapi's Voice AI Platform for Issues with No Response After Initial Message](https://discord.com/channels/1211482211119796234/1211483291191083018/threads/1333747605628391434)barbelos
03/13/2025, 12:38 PM"messagePlan": {
"idleMessageMaxSpokenCount": 2,
"idleTimeoutSeconds": 5
}
Praveen Nair
03/15/2025, 2:04 PM"startSpeakingPlan": {
"waitSeconds": 0.6,
"smartEndpointingEnabled": false,
}
barbelos
03/16/2025, 8:26 AM13551297-b0bd-4df8-a240-c0e3f2c02b2b
8a0a51ea-6e34-4f2a-acb5-2251ec1fd7ea
Shubham Bajaj
03/17/2025, 8:11 PMVAD Detected: Speech Possibly Starting...
VAD Detected: Speech Started
The Voice Activity Detection (VAD) successfully detected that the user started speaking.
2. **Language Detection**:
Language detected vi with confidence 0.06495575
The system incorrectly identified the Arabic speech as Vietnamese with 0.06495575 confidence (just below the 0.07 threshold)
3. **Speech Stopping Detection**:
VAD Detected: Speech Possibly Stopping...
VAD Detected: Speech Stopped
The VAD correctly detected that the user stopped speaking.
4. **Endpointing Handling**:
Endpointing Ignoring VAD Speech Detected. No Active Timeout.
This is where the issue occurred. Despite detecting speech start/stop, the system didn't properly handle the endpointing.
## Root Cause
The issue stems from multiple factors:
1. **Language Detection Mismatch**:
- Configuration specifies Arabic: "transcriber": {"language": "ar-SA", "provider": "azure"}
- System detected Vietnamese with 0.06495575 confidence (below 0.07 threshold)
- This caused confusion in language processing
2. **Endpointing Configuration Issue**:
- "smartEndpointingEnabled": false
means using basic heuristic endpointing
- This basic endpointing doesn't handle Arabic speech patterns effectively
3. **Transcription Provider Settings**:
- Using Azure Speech Service with Arabic configuration
- But the speech detection confusion caused the pipeline to break down
## Recommended Solutions
### Solution 1: Try LiveKit Endpointing with Adjusted Wait Time
Update your assistant configuration to:
json
"transcriber": {
"language": "ar-SA",
"provider": "azure",
"codeSwitchingEnabled": false
},
"startSpeakingPlan": {
"waitSeconds": 0.8,
"smartEndpointingEnabled": "livekit"
}
This configuration:
- Disables language code switching to prevent the Vietnamese misdetection
- Enables LiveKit's smart endpointing which may handle Arabic better
- Increases waitSeconds to give more time for proper speech detection
### Solution 2: If LiveKit Doesn't Work, Try Deepgram
If Solution 1 doesn't resolve the issue, switch to Deepgram:
json
"transcriber": {
"language": "ar",
"provider": "deepgram",
"model": "nova-2",
"codeSwitchingEnabled": false
},
"startSpeakingPlan": {
"waitSeconds": 0.8,
"smartEndpointingEnabled": false
}
Deepgram provides strong support for Arabic and may handle the speech patterns better than Azure in this case.
### Solution 3: Adjust VAD Parameters for Better Detection
If you prefer to stay with Azure, try adjusting the VAD parameters:
json
"transcriber": {
"language": "ar-SA",
"provider": "azure",
"codeSwitchingEnabled": false
},
"startSpeakingPlan": {
"waitSeconds": 0.8,
"smartEndpointingEnabled": false
},
"stopSpeakingPlan": {
"voiceSeconds": 0.3,
"silenceSeconds": 1.2
}
This configuration:
- Adjusts the stopSpeakingPlan to be more sensitive to Arabic speech patterns
- Increases silence detection time to better handle pauses in Arabic speech
## Next Steps
1. Try Solution 1 first (LiveKit with waitSeconds=0.8)
2. If that doesn't work, try either Solution 2 (Deepgram) or Solution 3 (adjusted VAD)barbelos
03/17/2025, 8:19 PM"assistant.transcriber.property codeSwitchingEnabled should not exist"
barbelos
03/17/2025, 8:22 PMar
When using deepgram provider, error is as follows
"assistant.transcriber.language must be one of the following values for nova-2 model: bg, ca, zh, zh-CN, zh-HK, zh-Hans, zh-TW, zh-Hant, cs, da, da-DK, nl, en, en-US, en-AU, en-GB, en-NZ, en-IN, et, fi, nl-BE, fr, fr-CA, de, de-CH, el, hi, hu, id, it, ja, ko, ko-KR, lv, lt, ms, multi, no, pl, pt, pt-BR, ro, ru, sk, es, es-419, sv, sv-SE, th, th-TH, tr, uk, vi"
barbelos
03/17/2025, 8:23 PMcodeSwitchingEnabled
is not available for azure transcription through your endpoint at least!Shubham Bajaj
03/18/2025, 11:34 PMjson
"startSpeakingPlan": {
"waitSeconds": 0.8,
"smartEndpointingEnabled": false
}
Shubham Bajaj
03/18/2025, 11:35 PMstartSpeakingPlan
changes suggested.barbelos
03/19/2025, 12:27 AM"transcriber": {
"language": "ar",
"model": "whisper",
"provider": "talkscriber"
}
Praveen Nair
03/20/2025, 3:20 AMbarbelos
03/20/2025, 8:37 AMar-SA
previously. Was there any recent update on the Azure transcriber on your side? Do you have any other recommendations that will allow us to use the Azure transcriber?Praveen Nair
03/22/2025, 3:53 AMbarbelos
03/22/2025, 8:33 AMPraveen Nair
03/23/2025, 10:02 AMPraveen Nair
03/23/2025, 10:02 AM