OpenAI Unveils ChatGPT as ‘App Store’ & Bombshell Jony Ive AI Hardware | Google’s Web Agents Advance, AUI Boosts Reliability

Key Takeaways
- OpenAI announced a sweeping strategy to evolve ChatGPT into a full-fledged computing platform and “App Store,” with new SDKs for interactive apps and robust tools for building autonomous agents.
- A major surprise from OpenAI’s Dev Day was the revelation of a three-year collaboration with legendary designer Jony Ive on new AI-centric hardware, aiming to redefine human-technology interaction.
- Google DeepMind launched “Gemini 2.5 Pro Computer Use,” an advanced agent capable of autonomously interacting with web interfaces, filling forms, and outperforming rivals in UI interaction benchmarks.
- Stealth startup AUI introduced Apollo-1, a foundation model leveraging “stateful neuro-symbolic reasoning” to achieve unprecedented reliability (90%+ pass rates) for enterprise-grade task-oriented AI agents.
Main Developments
OpenAI’s third annual DevDay in San Francisco marked a significant pivot, as CEO Sam Altman unveiled a vision to transform ChatGPT from a chatbot into a comprehensive computing platform and “App Store.” With the new Apps SDK, developers can now build interactive applications directly within ChatGPT, promising access to a user base exceeding 800 million. This move, according to OpenAI, aims to shift from merely asking AI questions to commanding it to perform complex tasks, reflecting a future where AI acts as a “super assistant” that evolves beyond traditional chat interfaces. The Agent Kit further empowers developers to create autonomous AI workers, with integrated tools for workflow design, deployment, and performance evaluation, directly addressing enterprise demands for AI as a productivity engine. A standout example saw financial platform Ramp building a procurement agent in hours, drastically cutting down task completion times. The advancement of Codex, now powered by a specialized GPT-5, showcased AI’s capability to autonomously write, review, and even evolve code from natural language or even a whiteboard sketch.
However, the day’s biggest revelation came from a non-livestreamed fireside chat: OpenAI’s three-year collaboration with legendary Apple designer Jony Ive on a new family of AI-centric hardware. Ive articulated a desire to redefine humanity’s relationship with technology, moving beyond “legacy products” to physical forms specifically designed for breathtaking AI capabilities. This bombshell announcement underscores OpenAI’s ambition to extend its influence beyond the cloud, into tangible user experiences. Underlying all these initiatives is an “unquenchable thirst for compute,” as acknowledged by OpenAI leadership, signaling continued massive investments in infrastructure to meet overwhelming demand.
Meanwhile, Google DeepMind unveiled its own formidable entry into the agent space with “Gemini 2.5 Pro Computer Use.” This fine-tuned version of Gemini 2.5 Pro can autonomously surf the web, click buttons, retrieve information, and fill out forms using a virtual browser. Google’s partnership with Browserbase and impressive benchmarks against rivals like Claude Sonnet and OpenAI’s agent models—including successfully navigating a Google Search Captcha in tests—position it as a strong contender in interface control. While currently lacking direct file system access, Gemini 2.5 Computer Use focuses on enabling developers to create agents for interface-driven tasks, operating at lower latency for production use cases.
Addressing a critical gap in enterprise AI, stealth startup Augmented Intelligence (AUI) introduced Apollo-1, a foundation model designed for unprecedented reliability in task-oriented dialogue. Built on “stateful neuro-symbolic reasoning,” Apollo-1 promises “behavioral certainty” over probabilistic outcomes, achieving a staggering 92.5% pass rate on the TAU-Bench Airline benchmark, far surpassing competitors. AUI’s approach aims to provide the deterministic outcomes required for mission-critical enterprise tasks, where “always” performing an action is a requirement, not a preference. This model, which is already in pilot with Fortune 500 companies and set for general availability in November 2025, represents a significant leap towards truly trustworthy enterprise agents.
Further underscoring the enterprise focus, IBM launched Project Bob, an AI-first Integrated Development Environment (IDE) that orchestrates multiple LLMs (Claude, Mistral, Llama, Granite 4) to automate application modernization. Claiming 45% productivity gains for its internal developers, Project Bob maintains full-repository context, handling complex upgrades and integrating DevSecOps practices. IBM also introduced AgentOps for real-time agent governance and integrated the open-source Langflow visual agent builder into watsonx Orchestrate. These tools aim to bridge the “prototype to production chasm,” providing the governance, security, and scalability necessary for mission-critical AI agent deployments, reinforcing that robust infrastructure is now essential for enterprise AI adoption.
Analyst’s View
Today’s announcements confirm that the race to build truly autonomous, action-oriented AI agents is intensifying. OpenAI’s ambitious platform play, transforming ChatGPT into an ecosystem for apps and agents, coupled with the audacious Jony Ive hardware collaboration, signals a long-term vision to redefine computing interfaces beyond current paradigms. This isn’t just about better chatbots; it’s about AI becoming an operating system and a physical presence. However, as Google demonstrates with its capable web agents and AUI pushes for enterprise-grade reliability, the immediate battleground remains the reliable execution of complex tasks. IBM’s focus on governance and modernization highlights the practical challenges of deploying agents at scale in the enterprise. The underlying constraint of compute power looms large, suggesting that while the vision for pervasive AI is clear, the infrastructure needed to realize it will continue to drive massive investment and innovation. We’re moving rapidly from AI-as-a-tool to AI-as-an-actor, making reliability, safety, and integration paramount.
Source Material
- Google’s AI can now surf the web for you, click on buttons, and fill out forms with Gemini 2.5 Computer Use (VentureBeat AI)
- Has this stealth startup finally cracked the code on enterprise AI agent reliability? Meet AUI’s Apollo-1 (VentureBeat AI)
- OpenAI Dev Day 2025: ChatGPT becomes the new app store — and hardware is coming (VentureBeat AI)
- OpenAI unveils AgentKit that lets developers drag and drop to build AI agents (VentureBeat AI)
- IBM claims 45% productivity gains with Project Bob, its multi-model IDE that orchestrates LLMs with full repository context (VentureBeat AI)