streaming issue using custom llm
# support
d
using the web interface, i've created a bot and connect to a custom llm endpoint. Streaming is recommended and sometimes works, but often the response gets trucated like this https://cdn.discordapp.com/attachments/1265003141703667742/1265003142198857868/image.png?ex=669fed7b&is=669e9bfb&hm=f15b38c97ceb5d8cd54e31900abb08076ce9800a908203e7681748baf1b9fff2&
v
@Darok It sounds like you might be experiencing an issue with the streaming of responses from your custom LLM endpoint. To better assist you, I would need a bit more information about how you've set up your bot and the specific circumstances under which the response gets truncated. For instance, knowing the following could be helpful: - The settings you're using for streaming, such as buffer sizes or timeout values. - The average length of the responses that are getting truncated. - Whether there's a pattern to when the truncation happens, such as with certain types of queries or at specific times of day. In the meantime, you might want to check if there are any configurable parameters related to streaming that you can adjust, such as increasing the buffer size or extending the timeout period. Additionally, ensuring that your internet connection is stable and that the server hosting the LLM endpoint is not overloaded could also help. If you provide more details, I can give you more targeted advice or help troubleshoot the issue further.
You can continue this conversation by mentioning me in the message.
d
@Vapi what settings are u talking about ? the server or from the web interface ?
v
@Darok I'm referring to any settings that might be relevant to the streaming functionality of your bot, which could be configured either on the server side (where your custom LLM endpoint is hosted) or through the web interface of the VAPI platform you're using to create and manage your bot. On the server side, settings could include: - Stream buffer sizes - Timeout settings for responses - Rate limiting configurations - Connection stability and handling On the web interface side, settings could involve: - Configuration options for the bot's connection to the LLM endpoint - Parameters for handling streaming data - Error handling and retry logic for incomplete streams If you're unsure where these settings are or how to adjust them, you might want to consult the documentation for the VAPI platform or reach out to their support team for more detailed guidance. If there are specific settings you'd like to know more about or if you have error messages or logs, please share them, and I can provide more targeted assistance.
d
this is a professional endpoint that follows openai api endpoint specifications. it seems that the problem could be on your side
@Vapi looking for help on this one
v
@Darok I understand that you're using a professional endpoint that adheres to OpenAI API specifications, and you suspect the issue might be on the VAPI side. To assist you further, I would recommend checking the following: 1. **Web Interface Settings**: Ensure that the settings on the VAPI web interface are correctly configured for streaming. This includes checking any timeout settings, buffer sizes, or other relevant parameters that could affect streaming. 2. **Error Logs**: Review any error logs or messages that are generated when the response gets truncated. These logs can provide valuable insights into what might be causing the issue. 3. **Connection Stability**: Verify that the network connection between the VAPI platform and the professional endpoint is stable and reliable. Network issues can sometimes cause streaming data to be interrupted. 4. **Endpoint Configuration**: Double-check the configuration of your professional endpoint to ensure it's optimized for streaming and that there are no limitations that could cause truncation. 5. **VAPI Support**: If after checking these areas the problem persists, it may be beneficial to reach out to VAPI support for further assistance. They may be able to provide more detailed troubleshooting steps or identify if there's an issue on their end. If you would like me to help you reach out to VAPI support or if there's anything specific you would like me to do to assist you, please let me know.
d
@Vapi how many tokens a second is required to prevent it to be truncated ?
v
@Darok The number of tokens per second required to prevent truncation isn't typically a standard measure, as it can depend on various factors, including the specifics of the API implementation and the network conditions. However, in the context of OpenAI's API or similar endpoints, the concern is usually not about tokens per second but rather about ensuring that the entire response is received before any timeouts occur and that the client is able to process and handle the incoming stream of data efficiently. To prevent truncation, you should ensure that: 1. **Timeout Settings**: The timeout settings on both the client and server sides are configured to allow enough time for the entire message to be sent and received. 2. **Buffer Sizes**: The buffer sizes are adequate to handle the expected response sizes without overflowing. 3. **Rate Limiting**: The API's rate limiting allows for the expected throughput. If you're sending requests too quickly, you may need to throttle them to match the API's limits. 4. **Client Handling**: The client (in this case, the VAPI platform) is capable of processing the incoming data stream quickly enough to keep up with the data being sent from the server. If you're consistently experiencing truncation and you've ruled out issues with your network or client application, you may need to contact the support team for the API or the VAPI platform to investigate further. They might be able to provide more specific guidance based on their system's capabilities and limitations.
d
@Vapi how do i contact them ?
v
@Darok To contact the support team for the VAPI platform, you can typically follow these steps: 1. **Support Portal**: Visit the official website of the VAPI platform and look for a 'Support' or 'Help' section. This may lead you to a support portal where you can submit a ticket or find contact information. 2. **Documentation**: Check the documentation provided by VAPI, as it often includes instructions on how to get support or contact the team for technical issues. 3. **Email**: Look for a support or contact email address on the VAPI platform's website. You can send an email detailing your issue and any relevant information. 4. **Live Chat**: Some platforms offer live chat support. If available, you can use this feature for immediate assistance. 5. **Community Forums**: If the platform has community forums, you can post your issue there. Other users or support staff may be able to help. 6. **Social Media**: Some companies offer support through social media platforms. You can try reaching out via Twitter, Facebook, or LinkedIn. 7. **Phone Support**: If the platform provides a customer service phone number, you can call for direct assistance. Remember to provide as much detail as possible about your issue, including any error messages, logs, and descriptions of what you've already tried. This will help the support team understand and address your problem more effectively. If you need further assistance or have trouble finding the contact information, please let me know, and I'll do my best to help.
d
pls help
s
can you please share the call id with timestamp
3 Views