gpt-4o-mini

Neural Network

GPT-4o mini is a convenient tool for processing text and images, allowing you to get accurate answers quickly while saving resources. The model is designed for tasks such as brief analysis, content generation, basic classification, and simple descriptions. Thanks to its optimized size and cost, GPT-4o mini addresses everyday user needs without straining the budget.

Main

/

Models

/

gpt-4o-mini
16 384

Max answer length

(in tokens)

128 000

Context size

(in tokens)

17,68

Prompt cost

(per 1M tokens)

70,71

Answer cost

(per 1M tokens)

0

Image prompt

(per 1K tokens)

*Prices for using the API.
Overview
Providers
API
bothub
BotHub: Try chat GPT for freebot

Caps remaining: 0 CAPS
Providers gpt-4o-miniOn Bothub, you can select your own providers for requests. If you haven't made a selection, we will automatically find suitable providers who can handle the size and parameters of your request.
Code example and API for gpt-4o-miniWe offer full access to the OpenAI API through our service. All our endpoints fully comply with OpenAI endpoints and can be used both with plugins and when developing your own software through the SDK.Create API key
Javascript
Python
Curl
import OpenAI from 'openai';
const openai = new OpenAI({
  apiKey: '<your bothub access token>',
  baseURL: 'https://bothub.chat/api/v2/openai/v1'
});


// Sync - Text generation 

async function main() {
  const chatCompletion = await openai.chat.completions.create({
    messages: [{ role: 'user', content: 'Say this is a test' }],
    model: 'gpt-4o-mini',
  });
} 

// Async - Text generation 

async function main() {
  const stream = await openai.chat.completions.create({
    messages: [{ role: 'user', content: 'Say this is a test' }],
    model: 'gpt-4o-mini',
    stream: true
  });

  for await (const chunk of stream) {
    const part: string | null = chunk.choices[0].delta?.content ?? null;
  }
} 
main();
illustaration

How it works gpt-4o-mini?

The main advantage of GPT-4o mini is its low cost: $0.15 per million input tokens and $0.60 per million output tokens (details, description). Its wide context window (128,000 tokens) allows the model to better understand queries and generate complete responses of up to 16,400 tokens at a time. In the MMLU (5-shot) test, GPT-4o mini scores 82% while working with multimodal input data. With a speed of 85.2 tokens per second, you can perform everyday tasks without losing time or quality.