Hello. I am testing the calls with both OpenAI GPT-4o-mini and Google Gemini-2.0-flash. Generally Gemini-2.0-flash is cheaper than GPT-4o-mini. But I am observing for some reason Gemini-2.0-flash is costing more than the other. These are the call ids
7fe16f9c-e8ec-4731-add6-654abe6c004a (GPT-4o-mini)
de1fa024-3f09-47e2-ac81-6f8afc60f683 (Gemini-2.0-flash)
Looking at data, let me break down the costs for each call:
**Call with GPT-4o-mini (7fe16f9c-e8ec-4731-add6-654abe6c004a)**:
\- Model cost: $0.002 (18127 prompt tokens, 85 completion tokens)
\- Voice cost: $0.006 (408 characters)
\- Total cost (all components): $0.05
**Call with Gemini-2.0-flash (de1fa024-3f09-47e2-ac81-6f8afc60f683)**:
\- Model cost: $0.002 (15973 prompt tokens, 122 completion tokens)
\- Voice cost: $0.008 (541 characters)
\- Total cost (all components): $0.06
You're actually correct that the Gemini-2.0-flash model itself is cheaper ($0.002) than GPT-4o-mini ($0.002).
However, the total cost of the Gemini call is slightly higher because:
1\. The voice cost is significantly higher in the Gemini call ($0.008 vs $0.006) because it used more characters (541 vs 408)
2\. The completion tokens are higher for Gemini (122 vs 85)
This explains why the overall cost of the Gemini call appears higher, even though the model itself is cheaper. The difference is primarily due to variations in other components of the call, especially the voice component which used about 33% more characters in the Gemini call.