Published: July 30, 2025
OpenAI is quietly working on a powerful upgrade to ChatGPT—one that could fundamentally change how humans interact with AI. According to new reporting from The Information, this next-gen model is designed to handle images, audio, video, and text seamlessly, while integrating long-term memory and advanced real-time reasoning.

From Chatbot to Cognitive Companion
OpenAI has dominated the large language model (LLM) race since the debut of GPT-3 and the viral rise of ChatGPT. But the next iteration may go far beyond chat.
According to internal documents and employee interviews cited by The Information, OpenAI’s upcoming model is being trained to process and understand multiple modalities—including images, audio, and possibly even video. This positions it as more than just a language tool—it could become an intelligent assistant capable of deeper interaction across various inputs.
What’s more intriguing is the system’s long-term memory, which would allow the model to remember past interactions with users. This could make ChatGPT feel more human-like, enabling personalized recommendations, continuity in conversations, and even contextual learning.
A Major Shift in AI Reasoning
In a departure from today’s mostly static LLMs, OpenAI is reportedly experimenting with real-time reasoning capabilities. This means the system could reason through problems dynamically instead of generating one-off answers based on its training data.
“This is the kind of cognitive leap that shifts AI from ‘autocomplete on steroids’ to something resembling an adaptive thought process,” said Ben Bajarin, CEO of Creative Strategies, in a recent podcast.
If successful, this development could challenge the current limitations of generative AI, enabling applications in fields like real-time tutoring, legal analysis, enterprise automation, and even robotics.
Why This Matters
While ChatGPT has become a go-to AI assistant for millions, it still operates within tight boundaries. It can’t remember you between sessions (unless you explicitly enable it), it can’t see or hear, and it doesn’t “think” in real time. That could all change soon.
“We’re witnessing the beginning of a new class of AI systems that behave less like tools and more like collaborators,” says Sarah Guo, founder of AI-focused venture firm Conviction.
The inclusion of multimodality and memory also reflects a broader trend in AI development, where companies like Google DeepMind (with Gemini), Anthropic (Claude), and Meta (LLaMA) are racing toward more general-purpose, integrated intelligence systems.
Enterprise Implications
If the rumors hold true, OpenAI’s next major release could have serious enterprise implications:
- Customer support: Memory-enabled AI could handle ongoing support issues without users repeating themselves.
- Education: Real-time reasoning allows for dynamic tutoring that adapts to student responses.
- Productivity apps: Multimodal understanding could help AI read spreadsheets, analyze images, and process live audio or video for work tasks.
According to The Information, OpenAI is already demonstrating early versions of the technology to enterprise customers. This suggests a strong push toward monetization beyond ChatGPT Plus, potentially targeting sectors like healthcare, law, and financial services.
Competition Is Heating Up
OpenAI’s move comes as rivals step up their own innovation cycles. Google’s Gemini 1.5 Pro already supports a massive 1-million-token context window and multimodal input. Meanwhile, Anthropic’s Claude 3 Opus is being hailed as the most “humanlike” LLM to date.
But OpenAI has an edge in distribution. Its integrations with Microsoft products, especially via Copilot in Office and Azure, give it a direct enterprise channel few others can match.
What’s Next?
While OpenAI hasn’t officially confirmed launch timelines, insiders believe the new system may debut later this year or early 2026. Whether it will be branded GPT-5 or released under a different product line remains unknown.
What’s clear is that the AI assistant of tomorrow will not just respond—it will remember, adapt, and think.
Sources:
- The Information – “OpenAI is Working on a Multimodal ChatGPT”
(Note: Behind paywall)
https://www.theinformation.com/articles/openai-is-working-on-a-multimodal-chatgpt-with-memory-and-real-time-reasoning - Ben Bajarin – Creative Strategies (Podcast Commentary)
https://creativestrategies.com/ - Sarah Guo – Conviction VC
https://www.conviction.com/ - OpenAI Blog (For Future Verification)
https://openai.com/blog/ - Google DeepMind – Gemini Overview
https://deepmind.google/technologies/gemini/ - Anthropic Claude 3 Release Notes
https://www.anthropic.com/index/introducing-claude - Microsoft Copilot Overview
https://www.microsoft.com/en-us/microsoft-copilot