codex-mini-latest

Model Overview

codex-mini-latest is a fine-tuned version of o4-mini specifically for use in Codex CLI.

Key Features

  • Higher intelligence (4/4 dots rating)
  • Medium speed (3/5 lightning bolts rating)
  • Fast reasoning model optimized for the Codex CLI
  • 200,000 context window
  • 100,000 max output tokens
  • Knowledge cutoff: Jun 01, 2024
  • Reasoning token support

Technical Specifications

  • Pricing: Input: $1.50 per 1M tokens; Cached input: $0.375 per 1M tokens; Output: $6.00 per 1M tokens
  • Supports: Input: text and image; Output: text only
  • Features: Responses API support

Snapshots

  • codex-mini-latest

Positioning and Use Cases

codex-mini-latest is a fine-tuned version of o4-mini specifically for use in Codex CLI. For direct use in the API, we recommend starting with gpt-4.1.

Rate Limits

  • Free tier: Not supported
  • Tier 1: 1,000 RPM, 100,000 TPM
  • Tier 2: 2,000 RPM, 200,000 TPM
  • Tier 3: 5,000 RPM, 4,000,000 TPM
  • Tier 4: 10,000 RPM, 10,000,000 TPM
  • Tier 5: 30,000 RPM, 150,000,000 TPM

Documentation

Official Documentation

OpenAI

Pioneer in AI, globally renowned for GPT series models

codex-mini-latest

Parameters Unknow
Output tokens 100,000

codex-mini-latest Fast reasoning model optimized for the Codex CLI

Official: $1.5 • $6 Our Price: $1.2 • $48 Save 20%

Frequently Asked Questions

What is the uptime guarantee?
We guarantee 99.9% uptime with our enterprise-grade infrastructure and redundant systems.
How is pricing calculated?
Pricing is based on the number of tokens processed. Both input and output tokens are counted in the final cost.
What is the difference between GPT-4 and GPT-4 Turbo?
GPT-4 Turbo is the latest version with improved performance, longer context window, and more recent knowledge cutoff date.