OpenAIs GPT-4o model is free to use!


In this article, we’ll take a look at what OpenAI’s latest GPT-4o model can do. If you’re curious about its capabilities, be sure to read on.

On Monday, OpenAI announced its newest, fantastic model named GPT-4o. They practically wrote an article and there are plenty of videos on their YouTube channel in which they demonstrate its capabilities.

We will go through them in broad outline, so that by the end of the article you will have a comprehensive picture of what makes this new large language model (LLM) so groundbreaking.

It is noteworthy that GPT-4o will also be available for free.

News about GPT4-o

Its most notable ability is that you can practically have a conversation with it. The reaction time for answering has been reduced to a minimum with this AI model, and you can also interrupt his speech.

The previously available Voice service of ChatGPT needed 4-5 seconds for the answer, now that has completely changed.

Watch this example video to see what’s new.

The GPT4-o model

In the description, you can read that this model is actually an omni model. This means, that it can not only interpret your voice (Voice mode) and text, but can practically handle both.

The interpretation of images and videos is made possible by the Vision mode.

The Memory mode remembers our previous conversations, which ChatGPT can use later. With the Analyze mode, complex data can be analyzed quickly and efficiently. Examples: Excel spreadsheets or even a chart that you show as an image to chatGPT.

Practical examples

If you look at the introduction page of GPT-4o, you can see many examples of people using their phone’s camera with the ChatGPT app and asking ChatGPT what it sees in the camera, and it then answers them “verbally”.

So the same thing can now be done in real time with your phone’s camera. The big difference between the current GPT-4o and the previous Voice service is that a completely new model was actually trained here with voice, text, videos and images.

It is also worth watching the video where they show that one person speaks Italian, the other person speaks English. The phone interprets it in real time. This means that as soon as the Italian text is spoken, ChatGPT immediately translates it into English and vice versa.

You’ve may seen conversations, where politicians spoke in their own mother tongue and still understood each other. It was thanks to the dubbing service in the background.

In the future, this will probably be completely replaceable with such an AI solution.

Performance of GPT-4o

There is a comparison chart on the hello-gpt-4o page in the Model evaluations section, with which we compare the GPT-4o model through various benchmarks, for example with GPT4 turbo, or even with the paid Claude 3 Opus and Gemini Pro 1.5.

The speed difference according to the graph will become visible in practice. All models are already very fast and are therefore capable of responding in real time.

You can feel the difference between a free ChatGPT, i.e. GPT 3.5 and GPT 4, but the difference between the above big models is not really noticeable anymore.

Security and limitations

The good news is that they also put a lot of emphasis on security. It’s highly likely that you won’t be able to say obscene words with it and you won’t be able to ask ChatGPT for other silly things either.

This model is subjected to serious, in-depth testing, which means that a team tries to hack these systems and extract answers from them. The name of the testing method is the so-called red teaming.

The future

These services are constantly being improved. Maybe by the time you read this article, the entire interface will look completely different.

Thanks to the Memory mode, ChatGPT will also remember previous conversations. You can turn this off if you don’t want it to store information about you.

The Analyze mode is good for uploading documents and requesting all kinds of analytics from it. So there will be improvements here in the free GPT as well.


This new model contains huge innovations. It almost feels like talking to a live person. It can also be used to facilitate scientific tasks, such as the analysis of complex data, graphs, and images.

If this is also combined with image and video generation, or perhaps with the generation of sound and music, the limit to what we can create with it is the imagination.

Of course, he won’t cook lunch for you, but he can help you in creative and scientific tasks. 🙂

Thank you for reading this article!

Image by Franz Bachinger, Pixabay

Leave a Reply

Your email address will not be published. Required fields are marked *