The price for using images as part of your prompt has indeed not changed between GPT-4o-mini and GPT-4o
Yet overall, captioning 500 images now costs me 5x less. This is because when I'm captioning an image, I'm providing both an image and a text prompt. The cost of using the image in the prompt stays the same, but the cost of the text dramatically dropped.
Good catch: the calculators here are bizarre. For GPT-4o, a 512x512 image uses 170 tile tokens. For GPT-4o mini, a 512x512 image uses 5,667 tile tokens. How does that even work in the context of a ViT? The patches and its image encoder should be the same size/output.
Since the base token counts increase proportionally (which makes even less sense) I have a hunch there's a JavaScript bug instead.
Confirmed that mini uses ~30x more tokens than base gpt-4o using same image/same prompt: { completionTokens: 46, promptTokens: 14207, totalTokens: 14253 } vs. { completionTokens: 82, promptTokens: 465, totalTokens: 547 }.
Both start with 150x150px and if you click the (i) it says mini uses way more base tokens and way more tile tokens, it still costs the same...