OpenAI Launches GPT-Realtime-2 Voice Model, Real-Time Translation and Transcription Tools in Updated API
OpenAI announced new voice intelligence features for its API on Thursday, including GPT-Realtime-2 with GPT-5-class reasoning, real-time translation supporting more than 70 input languages, and live transcription. The updates shift real-time audio toward interfaces that can listen, reason, translate, transcribe and act during conversations. All models include guardrails against abuse.
gizmodo.comOpenAI announced new voice intelligence features for its API on Thursday, launching three models designed to let developers build applications that move beyond simple call-and-response audio interactions. The company launched GPT-Realtime-2, a voice model built to create realistic vocal simulation that can converse with users.
GPT-Realtime-2 is built with GPT-5-class reasoning and was created to deal with more complicated requests from users.
5. OpenAI also launched GPT-Realtime-Translate for real-time translation services that keep pace with the user conversationally. GPT-Realtime-Translate supports more than 70 input languages and 13 output languages.
The company launched GPT-Realtime-Whisper, which provides live speech-to-text transcription capabilities captured as interactions occur. All of the new voice models are included in OpenAI’s Realtime API. “Together, the models we are launching move real-time audio from simple call-and-response toward voice interfaces that can actually do work: listen, reason, translate, transcribe, and take action as a conversation unfolds,” OpenAI said.
Translate and Whisper are billed by the minute while GPT-Realtime-2 is billed by token consumption. TechCrunch reported that the new features could be handy for customer service systems. OpenAI notes that its new features will assist with education, media, events, and creator platforms, among others.
The company said it has built guardrails to stop its new features from being abused to create spam, fraud, or other forms of online abuse. OpenAI embedded certain triggers in the system so that conversations can be halted if they are detected as violating our harmful content guidelines. TechCrunch Disrupt 2026 is scheduled for October 13-15, 2026 in San Francisco, CA.
Key Facts
Story Timeline
3 events- 2026-05-07
OpenAI announced GPT-Realtime-2, GPT-Realtime-Translate, and GPT-Realtime-Whisper for its Realtime API
1 sourceTechCrunch - 2026-05-08
Current date; TechCrunch Disrupt 2026 early registration discount ends
1 sourceTechCrunch - 2026-10-13
TechCrunch Disrupt 2026 begins in San Francisco
1 sourceTechCrunch
Potential Impact
- 01
Developers receive GPT-5-class reasoning in voice model for handling complex requests
- 02
Real-time translation expands to more than 70 input languages with conversational pacing
- 03
Customer service, education, media, events and creator platforms gain new real-time voice tools
Transparency Panel
Related Stories
Substrate placeholder — needs reviewAkamai Signs $1.8 Billion Seven-Year Cloud Deal With Anthropic
Akamai Technologies announced a $1.8 billion seven-year contract with Anthropic for its Cloud Infrastructure Services, the largest in the company's history. The deal was disclosed in Akamai's first-quarter 2026 earnings report. Akamai shares rose 27 percent on May 8 following the…
techjuice.pkTrump Administration Considers New Oversight for Advanced AI Models
Anthropic's unreleased Mythos model, capable of autonomously finding software vulnerabilities, prompted a White House shift from its previous hands-off AI policy. President Trump is considering an executive order to establish a formal review process for the most powerful systems,…
pandaily.comNvidia CEO Jensen Huang Says He Does Not Mind Paying $8 Billion in California Taxes
Nvidia CEO Jensen Huang stated he is comfortable with his tax payments to California while speaking at the Milken Institute Global Conference. Huang addressed the proposed billionaire tax and affirmed his decision to continue living in the state. The comments came as conference a…