### Answer
Currently, the official Vapi documentation does not provide specific instructions on how to implement streaming responses back to Vapi when using a custom LLM. However, generally, when setting up a server to act as the LLM for Vapi, you would need to ensure that your endpoint supports streaming completions [(1)](
) .
If you are creating an assistant with a custom LLM, and your server is capable of handling streaming responses, you would typically set up your server to continuously send data back to Vapi in the format it requires. This often involves using websockets or similar technology that supports real-time data transmission.
For more guidance on how your server should handle requests and send responses to be compatible with Vapi, you may want to refer to the section on using your own server with a custom LLM in the Vapi documentation.
If you are looking for specific code examples or more detailed implementation guidance, these are not explicitly covered in the current documentation. You might find it useful to experiment based on the general principles outlined for custom LLM integrations or reach out to other developers in the community who might have implemented similar solutions.