OpenAI Launches Free GPT-4o Version with Multimodal Capabilities

OpenAI has introduced GPT-4o (omni), a major upgrade to its ChatGPT, now available for free. Announced during a recent event, GPT-4o enhances capabilities in text, vision, and audio, making it a fully multimodal model. Mira Murati, OpenAI’s CTO, highlighted that GPT-4o is significantly faster and offers improved functionality across multiple input forms.

 

Currently, the free tier includes text and image features, with more functionalities to be added gradually. Paid users benefit from higher capacity limits. CEO Sam Altman announced on X that developers can access an API for GPT-4o, which is twice as fast and half the price of GPT-4 Turbo.

 

Once fully implemented, ChatGPT will function as a “Her-like” voice assistant, capable of responding to real-time prompts and observing surroundings through a camera. Currently, the voice mode can handle one prompt at a time using audio input only.

 

The multimodal nature of GPT-4o allows it to interact using text, voice, and images, enhancing user experience. This development aims to make AI more accessible and functional for a broader audience.