With so many advancements in artificial intelligence, tracking what's new and how it can benefit you can be overwhelming. One of the latest developments making waves is GPT-4o. But what exactly is GPT-4o, and how can it help you and your team?

In this article, we'll explore the following:

  • What is GPT-4o?
  • How to test GPT-4o
  • Is GPT-4o free?
  • What are GPT-4o benchmarks?
  • How can I talk to GPT-4o?
  • GPT-4o vs. GPT-4 and GPT-3.5

What is GPT-4o (omni)?

What is GPT-4o?
What is GPT-4o?

GPT-4o (omni) represents a significant leap towards more natural interactions between humans and computers. This model can accept and generate inputs and outputs in text, audio, image, and video formats. With response times as fast as 232 milliseconds, GPT-4o matches the quick reflexes of human conversation.

GPT-4o stands out for its enhanced performance in multilingual, audio, and vision tasks. It matches GPT-4 Turbo in text and coding tasks, with significant improvements in non-English languages. Remarkably, it is faster and 50% cheaper in the API. GPT-4o offers superior understanding and generation in vision and audio compared to previous models.

Model capabilities

GPT-4o can handle diverse tasks, from singing and real-time translation to preparing for interviews and understanding sarcasm. Whether playing rock-paper-scissors, narrating visual stories, or even cracking dad jokes, GPT-4o showcases various interactive abilities. Its role as a text and vision model allows for more dynamic interactions.

Previously, voice interactions with models like GPT-3.5 and GPT-4 involved multiple steps, causing delays and limiting the model's ability to effectively process tone or multiple speakers. Now, GPT-4o integrates all these modalities into a single, cohesive model, enhancing its responsiveness and emotional expressiveness.

Early demonstrations show GPT-4o harmonizing songs, translating languages on the fly, and even providing customer service. It opens up new possibilities for artificial intelligence to assist in everyday tasks, make learning more interactive, and enhance communication across different languages and mediums.

How to Test GPT-4o

Here’s how to use GPT-4o for free on your Android or iOS device:

  1. Install the app from the Google Play Store or Apple App Store.
  2. Log in with your account details.
  3. Tap the icon in the upper right corner and choose "GPT-4o."
How to Test GPT-4o
How to Test GPT-4o
  1. Start your conversation with OpenAI’s newest Omni model. Note that the Android version currently does not support interruptions in the Voice Mode chat.

Is GPT-4o Free?

Yes! Free users can access its advanced features without any cost.

While the free version offers many functionalities, the Plus subscription provides additional benefits. Plus, users enjoy higher message limits and access to premium features.

What are GPT-4o Benchmarks?

GPT-4o demonstrates superior performance across multiple benchmarks, highlighting its strengths in text, audio, and vision tasks.

Text Evaluation

GPT-4o - Text Evaluation
GPT-4o - Text Evaluation

Image from OpenAI

GPT-4o achieves performance levels on par with GPT-4 Turbo in text and coding tasks, setting a new high score of 88.7% on zero-shot Chain-of-Thought (CoT) MMLU (Massive Multitask Language Understanding) evaluations. Additionally, it scores 87.2% on the traditional five-shot no-CoT MMLU, indicating its exceptional reasoning and general knowledge capabilities.

These scores mean it’s excellent at handling complex text-based tasks, just like its predecessor, GPT-4 Turbo.

Audio Performance

GPT-4o Audio Performance
GPT-4o Audio Performance

Image from OpenAI

Regarding audio tasks, GPT-4o excels in Automatic Speech Recognition (ASR) and audio translation. It significantly improves speech recognition over Whisper-v3 across various languages, especially those with fewer resources. GPT-4o also sets a new state-of-the-art in speech translation, outperforming Whisper-v3 on the MLS (Multilingual Speech) benchmark.

These advancements mean GPT-4o is excellent at recognizing and translating speech. It’s much better at understanding spoken words than previous models, especially in less widely spoken languages. It can also accurately translate spoken language into different languages.

Vision Understanding

GPT-4o Vision Understanding

Image from OpenAI

In vision tasks, GPT-4o achieves state-of-the-art results on visual perception benchmarks. It excels in zero-shot evaluations, including Multimodal Multitask Machine Understanding (MMMU), MathVista, and ChartQA. These benchmarks demonstrate GPT-4o’s ability to understand and interpret complex image inputs effectively.

These results mean it can understand and interpret images and visual data well. It was tested on various tasks requiring understanding pictures and charts, performing at the highest level.

Multilingual Capabilities

GPT-4o also shines in multilingual tasks, with improved performance in various languages due to a new tokenizer that compresses language tokens more efficiently. This results in fewer tokens required for accurate language processing, enhancing its capabilities in languages such as Gujarati, Telugu, Tamil, Marathi, Hindi, Urdu, Arabic, Persian, Russian, Korean, Vietnamese, Chinese, Japanese, Turkish, Italian, German, Spanish, Portuguese, and French.

How Can I Talk to GPT-4o?

Here’s how to use GPT-4o on your smartphone or desktop:


  1. Download the ChatGPT app from Google Play or Apple Store on your smartphone.
  2. Open the app and log in using your account credentials. If you don't have an account, you can easily create one.
  3. Tap the menu below and choose "GPT-4o."
How to talk to GPT-4o
How to talk to GPT-4o
  1. Start interacting with GPT-4o. Type your questions or use Voice Mode to speak directly to the model. To learn more about how this works, check out our article on ChatGPT’s speech-to-text capabilities.


  1. Visit the ChatGPT website on your desktop browser.
  2. Log in using your account.
  3. Click the settings menu, then select "GPT-4o."
GPT-4o on desktop
GPT-4o on desktop
  1. Begin your conversation with GPT-4o by typing your questions or using the microphone for voice interactions. Learning how to use GPT4-o on your desktop can improve your productivity.

Here are some common use cases for talking to GPT-4o:

Real-time translation

GPT-4o can translate spoken language instantly. For example, one person can speak English, and GPT-4o will translate it to Spanish in real-time.

Interactive learning

Use GPT-4o to tutor students in various subjects. For example, it can help students solve math problems by guiding them step-by-step without directly giving the answers. It can also adopt different tones depending on the context, making it suitable for casual and formal learning environments.

Customer support

GPT-4o can handle customer service tasks, such as making calls to resolve issues. For instance, it can call a company on your behalf to request a replacement device, reducing the time you spend on hold and dealing with customer service representatives.

Creative collaboration

GPT-4o can assist in creative projects. It can sing, harmonize, or even role-play scenarios. For example, it can engage in a playful conversation with you.

Daily assistance

GPT-4o can summarize meetings, take notes, and send summary emails. For example, during a meeting, it can identify speakers, summarize key points, and send the minutes to all participants.


GPT-4o can play games like rock-paper-scissors, respond with sarcasm, or even act as a conversational partner in a debate. For example, it can engage in a fun and engaging discussion about cats vs. dogs, taking sides and providing thoughtful arguments.

How Does GPT-4o Compare to GPT-4 and GPT-3.5?

Several key differences and improvements stand out when comparing GPT-4o to its predecessors, GPT-4 and GPT-3.5.

Multimodal Capabilities

GPT-4o is designed to handle multiple inputs and outputs, including text, audio, images, and video. This capability makes it more versatile than GPT-4 and GPT-3.5, which primarily focus on text and, to a limited extent, image processing. GPT-4o’s ability to integrate audio and video inputs means it can understand and respond more naturally and dynamically, similar to human interactions.

Response Time

One of GPT-4o's significant advancements is its response time. It can respond to audio inputs in as little as 232 milliseconds, comparable to human conversation response times. In contrast, GPT-3.5 and GPT-4 have longer response times, especially when processing audio inputs.

Cost and Efficiency

GPT-4o is designed to be faster and more cost-effective. It is 50% cheaper in the API compared to GPT-4 Turbo, making it more accessible for a broader range of applications. Its efficiency improvements also mean that it can handle higher request rates, providing a smoother experience for users.

For a deeper understanding of how token limits affect performance and cost, check out this comprehensive guide on token limits for ChatGPT-3.5 and ChatGPT-4.

Performance in Non-English Languages

GPT-4o shows significant improvement in understanding and generating text in non-English languages. While GPT-4 and GPT-3.5 have strong capabilities in English, GPT-4o expands its proficiency across multiple languages, making it a better choice for global applications.

Integrated Model for Voice, Text, and Vision

Unlike GPT-3.5 and GPT-4, which use separate models for different tasks, GPT-4o integrates all modalities into a single model. This integration allows GPT-4o to maintain context and provide more coherent responses across various input types. For instance, it can interpret visual cues while responding to voice commands, offering a more holistic understanding of the input.

Use Cases and Applications

GPT-4o introduces new use cases that were not possible or practical with GPT-4 and GPT-3.5. These include real-time interactive learning, advanced customer support, and creative collaboration involving singing or storytelling. Its ability to understand and generate audio and visual content opens up new possibilities for innovative applications.

Enhanced Emotional and Contextual Understanding

GPT-4o has improved capabilities in detecting and conveying emotions through voice, making interactions more natural and engaging. This feature is a step up from GPT-4 and GPT-3.5, which primarily focus on text-based interactions with limited emotional context.

In summary, GPT-4o represents a significant advancement over GPT-4 and GPT-3.5, offering enhanced multimodal capabilities, faster response times, cost efficiency, and improved performance across multiple languages. Its integrated voice, text, and vision model provides a more cohesive and versatile AI experience. Its image capabilities make it particularly powerful for various visual tasks.


Embrace the Future with GPT-4o

GPT-4o is a game-changer in the world of AI. This advanced model combines text, audio, video, and image capabilities, making interactions more natural and intuitive. It’s not only faster and more cost-effective but also excels in languages.

You can use GPT-4o to improve real-time translation, engage in interactive learning, enhance customer support, or explore creative projects. Plus, free users can access these features, making top-tier AI accessible to everyone.

Want the convenience of AI summaries?

Try Tactiq for your upcoming meeting.

Want the convenience of AI summaries?

Try Tactiq for your upcoming meeting.

Bringing AI into your meetings has never been so easy.

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique.