OpenAI Releases o3-Pro

Plus, this AI is controlling the next generation of robots

Good morning. It’s Wednesday, June 11th.

On this day in tech history: In 1978, Intel launched the 8086, a 16-bit processor that became the foundation of the x86 architecture. Adopted in the first IBM PCs, it set a standard that still drives most personal computers today. The chip enabled faster processing and more complex software at the time, but its deeper impact was locking in the x86 model. Nearly five decades later, its legacy continues to define how PCs are built and how software is written.

In today’s email:

  • OpenAI taps Google, slows o3-pro for precision

  • Google speeds up Veo 3, adds Mariner to Chrome

  • Redwood equips NEO for mobile, full-body tasks

  • 5 New AI Tools

  • Latest AI Research Papers

You read. We listen. Let us know what you think by replying to this email.

Clone Any Voice. Speak Any Language. Sound Human.

Fish.Audio is redefining AI voice generation with the most realistic speech synthesis on the market. Whether you're creating voiceovers, audiobooks, ads, or full-blown AI characters, this platform delivers stunning quality and control.

  • 200,000+ voices in the public library — or upload your own

  • Voice cloning from just 15 seconds of audio

  • Text-to-speech, speech-to-text, and full voice agent support

  • Supports 13 languages with native-level expressiveness

  • Used by over 150 creators and studios across YouTube, TikTok, and beyond

Side-by-side tests show Fish.Audio outperforming ElevenLabs in emotional nuance and clarity. From Taylor Swift and Elon to original avatars, it’s become the go-to platform for creators who want voices that feel real.

🎙 Try it now: fish.audio/discovery

Today’s trending AI news stories

OpenAI Trades Speed for Precision with o3-pro; Taps Google for Compute

OpenAI is leaning hard into precision with o3-pro, a tool-native upgrade to its o3 model built for enterprise-grade reasoning. Tuned for precision over speed, it integrates Python, file analysis, web browsing, and vision, making it ideal for complex, multi-step tasks where accuracy trumps latency. Available via API and now default for Pro and Team ChatGPT users, o3-pro trades speed and cost at $20 input and $80 output per million tokens, for deeper tool use and more deliberate thinking. It can take minutes per response, but early testers say it handles uncertainty well, reasons in context, and chooses tools intelligently. It's slow, but by design.

Image: OpenAI

At the same time, OpenAI dropped o3’s pricing by 80%, pushing high-reasoning LLMs into mass-market territory. Input is now $2 per million tokens, output $8, with cacheable inputs as low as $0.50. “Flex mode” offers speed–cost tradeoffs, and benchmarks show o3 still holds its own against Claude Opus and Gemini Pro at a fraction of the price. This isn’t just a price cut, it’s a compression of the AI value curve, aimed at devs, startups, and researchers who need power without the premium.

The company’s open-weight model, once slated for June, has been delayed following a breakthrough it now plans to fold in. The model aims to bring o-series reasoning to the open ecosystem, potentially augmented by OpenAI’s cloud backend. Though timing slips, the goal is strategic: redefine what “open” can do, even as rivals like Mistral and Qwen gain ground.

Behind it all, OpenAI has quietly signed a compute deal with Google Cloud, shifting from a Microsoft-first infrastructure stance to a more pragmatic, multi-cloud play. In the LLM arms race, compute is the real constraint, and even competitors now share silicon.

Looking ahead, OpenAI sees a horizon beyond answers. In The Gentle Singularity, Altman suggests that 2026 could bring models capable of producing genuinely novel insights. The company’s roadmap points toward generative cognition rather than imitation. That vision is already fueling tangible results: OpenAI now reports $10 billion in annual recurring revenue, doubling from $5.5 billion a year ago, powered by ChatGPT subscriptions, enterprise deployments, and developer API usage. With 500 million weekly active users and 3 million business customers, the company’s growth has been extraordinary, but so have its costs. Read more.

Google Upgrades Veo 3 Speed, Embeds Project Mariner in Chrome

Google is advancing its AI-native tooling with two closely linked developments. Veo 3 Fast significantly boosts the rendering speed of 720p video, more than doubling that of its predecessor, while maintaining integration across both the Gemini app and Flow. Gemini Pro users can now generate three videos daily, and Flow Pro users are charged 20 credits per output, while Ultra-tier users retain access to higher quality and generation limits. Google is also experimenting with multimodal prompts like voice-to-video and plans to scale access to Workspace accounts and international users.

Running parallel to this is the gradual release of Project Mariner, an experimental browser-based agent embedded directly in Chrome. Available to Gemini Ultra subscribers, Mariner operates across open tabs and can handle navigation, form-filling, and transactional tasks through a prompt-driven chat interface. Designed with strict permission gating, the agent asks for user approval before taking actions, even for basic lookups, reflecting Google’s emphasis on privacy-aware automation. Read more.

Redwood AI equips humanoid robot ‘NEO’ for mobile tasks, whole-body manipulation

1X Technologies has launched Redwood, a lightweight AI model that turns its NEO humanoid into a home-capable autonomous agent. Trained on real-world robot data, Redwood lets NEO move, perceive, and act in domestic spaces, handling tasks like laundry, door answering, and indoor navigation. The model generalizes well, adapting to novel objects and retrying failed grasps. It enables whole-body, multi-contact manipulation, synchronizing locomotion and arm movement for mobile, dynamic control, including bracing and leaning.

Crucially, Redwood runs entirely on NEO’s onboard GPU, with voice-driven intent prediction handled by a connected language model. Unlike simulation-tuned AI, Redwood is grounded in practical deployment, as demoed at NVIDIA GTC 2025. For developers, it offers a path to scalable, compute-efficient humanoid autonomy. For robotics, it’s a signpost toward real-world utility. Read more.

5 new AI-powered tools from around the web

arXiv is a free online library where researchers share pre-publication papers.

Thank you for reading today’s edition.

Your feedback is valuable. Respond to this email and tell us how you think we could add more value to this newsletter.

Interested in reaching smart readers like you? To become an AI Breakfast sponsor, reply to this email or DM us on 𝕏!