OpenAI releases the latest flagship model GPT4o, ChatGPT enters a new stage of human-computer interaction
2024-05-13 18:14
Odaily News OpenAI released its latest flagship model GPT-4o, which can reason about audio, vision and text in real time. The main concept is a humanized, supernatural, ultra-low latency personal voice interaction assistant. According to the OpenAI official website and the official account of the X platform, the o in GPT4o stands for Omni (comprehensive), which is a step towards more natural human-computer interaction-it accepts any combination of text, audio and images as input, and supports the generation of any combination of text, audio and image output. It can respond to audio input in 232 milliseconds and an average of 320 milliseconds, which is similar to the human reaction time in conversation. In terms of English and code, it performs on par with GPT-4 Turbo, with significant improvements on non-English language texts, while the API is faster and 50% cheaper. Compared with existing models, GPT-4o performs particularly well in visual and audio understanding. Text and image input will be launched in API and ChatGPT today, and voice and video input will be launched in the coming weeks.
2025-11-03 22:50
Ripple acquires custody and wallet technology company Palisade
2025-11-03 18:02
Coinbase includes Monad (MON) in its listing roadmap
2025-11-03 17:01
Coinbase will list KITE
Download Odaily App
Let Some People Understand Web3.0 First
