Hey
@zbeaver4 another user here, I had a similar question a few days ago. The conclusion I came to was one would need to implement a custom LLM server to inject messages into the context. Ref to docs:
https://docs.vapi.ai/custom_llm
So you'd setup some intermediate server between your preferred LLM provider, inject the messages, then send to vapi. It be really nice if the devs of vapi gave us some way to do this through their platform as I imagine the above implementation would introduce additional lag which I'm unsure how to engineer around at the moment.