Learn how to save OpenAI costs by using Prem LLMs with a real-world invoice parsing example.
$15.00
per million output tokens for GPT-4o and $1.20
for GPT-4o-mini.
In contrast, Prem offers flat inference rates for any hosted or fine-tuned model that are extremely economical:
Model Type | Input (per 1M tokens) | Output (per 1M tokens) | Total Inference Cost |
---|---|---|---|
Prem SLM (all sizes) | $0.10 | $0.30 | $0.40 |
OpenAI GPT-4o | $5.00 | $15.00 | $20.00 |
OpenAI GPT-4o-mini | $0.30 | $1.20 | $1.50 |