in Blog

July 23, 2024

OpenAI introduced GPT-4o Mini. What do we know about it?

Author:




Kaja Grzybowska


Reading time:




6 minutes


OpenAI’s recent introduction of GPT-4o Mini has been widely promoted as a breakthrough in accessible and efficient AI development. Does it live up to this bold claim?

This article critically examines the key features of GPT-4o Mini, including its purported enhanced performance, affordability, and improved safety measures. We delve into its technical specifications, compare it with its predecessor, GPT-3.5 Turbo, and analyze its potential impact on AI adoption across various industries.

Generative-AI-CTA

What is GPT-4o Mini?

GPT-4o Mini is designed to be OpenAI’s most cost-efficient small language model. With pricing set at just 15 cents per million input tokens and 60 cents per million output tokens, it significantly undercuts the cost of previous models. 

This price reduction doesn’t compromise the performance, however. The model scores 82% on the Massive Multitask Language Understanding (MMLU) benchmark, outperforming its predecessor, GPT-3.5 Turbo, in chat preferences.

GPT-4o Mini: Technical Specifications

GPT-4o Mini boasts a substantial context window of 128K tokens and supports up to 16K output tokens per request. This allows for handling complex, context-rich tasks with ease.

Currently, the model supports text and vision inputs through its API, with plans to expand to text, image, video, and audio inputs and outputs in the future.

The model excels in

  1. Reasoning tasks
  2. Mathematical computations
  3. Coding proficiency
  4. Multimodal reasoning

GPT-4o Mini vs. GPT-3.5 Turbo

  • Cost-Efficiency
    GPT-4o Mini is significantly more affordable, costing $0.15 per million input tokens and $0.60 per million output tokens. This is roughly 3.3x cheaper for input and 2.5x cheaper for output compared to GPT-3.5 Turbo.
  • Context Window
    With a massive 128K token input context window, GPT-4o Mini dwarfs GPT-3.5 Turbo’s 4,096 tokens, allowing for a much more comprehensive understanding of complex queries.
  • Output Capacity
    GPT-4o Mini can generate up to 16.4K tokens per request, a substantial increase from GPT-3.5 Turbo’s 4,096 token limit.
  • Performance
    On the MMLU benchmark, GPT-4o Mini scores an impressive 82.0 in the 5-shot setting, compared to 70.0 for GPT-3.5 Turbo, indicating superior language understanding and reasoning capabilities.
  • Capabilities
    GPT-4o Mini demonstrates improved performance in reasoning tasks, mathematical computations, coding proficiency, and multimodal reasoning compared to its predecessor.

It’s, however, worth noting that some users have reported inconsistent performance in certain tasks. Let’s take a look at them:

GPT-4o Mini vs. GPT-3.5 Turbo: Limitations and Considerations

  • Inconsistent Performance

Some users have reported that GPT-4o Mini can underperform GPT-3.5 Turbo in certain tasks, particularly in simple numerical comparisons. This inconsistency suggests that the model’s capabilities may vary depending on the specific task at hand.

  • Data Extraction

GPT-4o Mini has shown occasional weaknesses in data extraction tasks, sometimes missing key information that GPT-3.5 Turbo would successfully identify. This could be a crucial consideration for users relying on the model for information retrieval tasks.

  • Translation Quality

While comprehensive studies are yet to be conducted, anecdotal evidence suggests that some users still find the translation capabilities of other models, such as Claude Haiku, superior to GPT-4o Mini in certain language pairs or contexts.

  • Benchmark Gaps

Although GPT-4o Mini outperforms GPT-3.5 Turbo on benchmarks like MMLU, there are other important benchmarks where direct comparison data is not yet available. This makes it challenging to comprehensively evaluate the model’s performance across all potential use cases.

NOTE: These limitations highlight the importance of careful evaluation when choosing between GPT-4o Mini and GPT-3.5 Turbo for specific applications.

Enhanced Safety Through Instruction Hierarchy

A key innovation in the GPT-4o family, including GPT-4o Mini, is the implementation of instruction hierarchy, a novel approach to enhancing AI safety and reliability.

The instruction hierarchy method works by establishing a hierarchy of instructions that the model must follow. This hierarchy ensures that the most critical and fundamental instructions, such as ethical guidelines and safety protocols, take precedence over less important ones.

Instruction hierarchy is designed to make the AI more resistant to various types of attacks, particularly prompt injections and jailbreaks, which have been the challenges in earlier language models.

By helping the AI prioritize and follow the most important instructions, even when faced with conflicting or potentially malicious prompts, this approach – as OpenAI claims – significantly improves the model’s ability to adhere to ethical guidelines, making it substantially more difficult for users to override the AI’s core safety protocols or manipulate it into producing harmful content.

This enhanced resistance to manipulation ensures that GPT-4o Mini can more consistently ignore attempts to make it disregard its training or ethical constraints, regardless of user input.

Read more: Not only GPT. What LLMs can you choose from?

Cost difference and its impact on adoption

GPT-4o Mini’s price is significantly lower than GPT-3.5 Turbo’s, and this 3.3x reduction for input and 2.5x reduction for output tokens opens doors for wider adoption across various sectors.

Small and medium-sized enterprises, startups, and individual developers with limited budgets can now leverage powerful AI capabilities that were previously out of reach, potentially leading to a surge in AI-driven innovation across diverse fields.

Read more: LLM Implementation Strategy: Preparation Guide for Using LLMs

GPT-4o Mini’s cost-effectiveness makes it particularly attractive for applications requiring the processing of large volumes of data or frequent API calls, which could accelerate its adoption in areas such as large-scale data analysis, content generation, and real-time language processing systems.

Furthermore, OpenAI’s intention to replace GPT-3.5 Turbo with GPT-4o Mini in ChatGPT for Free, Plus, and Team users is a strategic move that could significantly accelerate the adoption of the newer model by exposing a large user base to its capabilities. However, the adoption rates will also depend on factors such as ease of integration, specific use case requirements, and the resolution of any limitations or inconsistencies in the model’s performance.

References:

  1. ContextAI. URL: https://context.ai/compare/gpt-4o-mini/gpt-3-5-turbo-16k. Accessed Jul. 27. 2024.
  2. Community Open AI. URL: https://community.openai.com/t/gpt-4o-mini-is-dummber-than-you-can-think/871987. Accessed Jul. 27. 2024.
  3. Bind AI. GPT-4o Mini: Is it better than GPT-4o? Will it replace GPT-3.5 turbo? URL: https://blog.getbind.co/2024/07/19/gpt-4o-mini-vs-gpt-4o-vs-gpt-3-5-turbo/. Accessed 27. 2024.
  4. OpenAI. URL: https://openai.com/index/new-tools-for-chatgpt-enterprise/. Accessed Jul. 27. 2024.
  5. TheVerge. OpenAI’s latest model will block the ‘ignore all previous instructions’ loophole. URL: https://www.theverge.com/2024/7/19/24201414/openai-chatgpt-gpt-4o-prompt-injection-instruction-hierarchy. Accessed Jul. 27. 2024.


Category:


Generative AI