<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
    <channel>
        <title>Realtime API on KnightLi Blog</title>
        <link>https://www.knightli.com/en/tags/realtime-api/</link>
        <description>Recent content in Realtime API on KnightLi Blog</description>
        <generator>Hugo -- gohugo.io</generator>
        <language>en</language>
        <lastBuildDate>Sat, 09 May 2026 10:58:47 +0800</lastBuildDate><atom:link href="https://www.knightli.com/en/tags/realtime-api/index.xml" rel="self" type="application/rss+xml" /><item>
        <title>OpenAI&#39;s New Realtime Voice Models: GPT-Realtime-2, Live Translation, and Streaming Transcription</title>
        <link>https://www.knightli.com/en/2026/05/09/openai-realtime-voice-models-gpt-realtime-2-translate-whisper/</link>
        <pubDate>Sat, 09 May 2026 10:58:47 +0800</pubDate>
        
        <guid>https://www.knightli.com/en/2026/05/09/openai-realtime-voice-models-gpt-realtime-2-translate-whisper/</guid>
        <description>&lt;p&gt;On May 7, 2026, OpenAI introduced a new generation of voice models for the Realtime API. The point is not only to make AI sound more natural, but to let voice agents understand, reason, call tools, translate, and transcribe during a live conversation.&lt;/p&gt;
&lt;p&gt;The update includes three models:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;code&gt;GPT-Realtime-2&lt;/code&gt;: the main model for realtime voice agents, with stronger reasoning, tool calling, and longer context.&lt;/li&gt;
&lt;li&gt;&lt;code&gt;GPT-Realtime-Translate&lt;/code&gt;: a live speech translation model that supports 70+ input languages and 13 output languages.&lt;/li&gt;
&lt;li&gt;&lt;code&gt;GPT-Realtime-Whisper&lt;/code&gt;: a low-latency streaming speech-to-text model for captions, meeting notes, and realtime workflows.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;If early voice assistants were mostly “ask once, answer once,” this release moves closer to a voice interface that can listen and act at the same time.&lt;/p&gt;
&lt;h2 id=&#34;gpt-realtime-2-the-main-model-for-voice-agents&#34;&gt;GPT-Realtime-2: the main model for voice agents
&lt;/h2&gt;&lt;p&gt;&lt;code&gt;GPT-Realtime-2&lt;/code&gt; is built for live voice interactions. It does not just answer questions; it needs to keep context while the user speaks, changes direction, interrupts, or adds constraints, and then call tools when needed.&lt;/p&gt;
&lt;p&gt;Officially highlighted capabilities include:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Short preambles before a response, such as “let me check that,” so users know the system is working.&lt;/li&gt;
&lt;li&gt;Parallel tool calls for calendars, search, orders, support systems, and other multi-tool workflows.&lt;/li&gt;
&lt;li&gt;More natural recovery behavior when something fails.&lt;/li&gt;
&lt;li&gt;A context window increased from 32K to 128K for longer conversations and more complex task flows.&lt;/li&gt;
&lt;li&gt;Better retention of specialized terminology, proper nouns, and medical vocabulary.&lt;/li&gt;
&lt;li&gt;More controllable tone and delivery, such as calm, empathetic, confirmational, or upbeat responses.&lt;/li&gt;
&lt;li&gt;Adjustable reasoning effort: &lt;code&gt;minimal&lt;/code&gt;, &lt;code&gt;low&lt;/code&gt;, &lt;code&gt;medium&lt;/code&gt;, &lt;code&gt;high&lt;/code&gt;, and &lt;code&gt;xhigh&lt;/code&gt;, with &lt;code&gt;low&lt;/code&gt; as the default.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This means developers can use voice agents in more demanding products, not only simple Q&amp;amp;A. A support agent can listen while checking an order; a travel app can give next steps after a flight change; a real estate assistant can filter listings and schedule a tour from spoken requirements.&lt;/p&gt;
&lt;h2 id=&#34;live-translation-for-multilingual-voice-products&#34;&gt;Live translation for multilingual voice products
&lt;/h2&gt;&lt;p&gt;&lt;code&gt;GPT-Realtime-Translate&lt;/code&gt; is designed for live speech translation. People can speak in their own language while the other side hears translated speech and sees realtime transcripts.&lt;/p&gt;
&lt;p&gt;Clear use cases include:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Multilingual customer support.&lt;/li&gt;
&lt;li&gt;Cross-border sales and pre-sales conversations.&lt;/li&gt;
&lt;li&gt;Online education and live events.&lt;/li&gt;
&lt;li&gt;International meetings and hosting.&lt;/li&gt;
&lt;li&gt;Creator and video platform localization.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The hard part of live translation is not only accuracy. It also requires low latency, natural pauses, tone preservation, accent handling, and domain vocabulary. OpenAI is emphasizing cross-language conversations that feel closer to natural speech, instead of waiting for an entire segment before translation begins.&lt;/p&gt;
&lt;h2 id=&#34;streaming-transcription-voice-content-enters-workflows-immediately&#34;&gt;Streaming transcription: voice content enters workflows immediately
&lt;/h2&gt;&lt;p&gt;&lt;code&gt;GPT-Realtime-Whisper&lt;/code&gt; is the new streaming speech-to-text model. Its value is turning speech into usable text while it is happening, instead of waiting for a recording to finish.&lt;/p&gt;
&lt;p&gt;Common applications include:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Live meeting captions.&lt;/li&gt;
&lt;li&gt;Classroom and broadcast captions.&lt;/li&gt;
&lt;li&gt;Realtime meeting notes.&lt;/li&gt;
&lt;li&gt;Continuous dictation input for voice agents.&lt;/li&gt;
&lt;li&gt;Follow-up workflows in support, healthcare, recruiting, sales, and other high-volume voice scenarios.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;For products, streaming transcription shortens the time from spoken words to actionable text. Captions appear faster, notes can be generated during the conversation, and downstream workflows such as summaries, task extraction, and CRM updates can start earlier.&lt;/p&gt;
&lt;h2 id=&#34;pricing-and-availability&#34;&gt;Pricing and availability
&lt;/h2&gt;&lt;p&gt;All three models are available in the Realtime API. Official pricing is:&lt;/p&gt;
&lt;table&gt;
  &lt;thead&gt;
      &lt;tr&gt;
          &lt;th&gt;Model&lt;/th&gt;
          &lt;th&gt;Price&lt;/th&gt;
      &lt;/tr&gt;
  &lt;/thead&gt;
  &lt;tbody&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;code&gt;GPT-Realtime-2&lt;/code&gt;&lt;/td&gt;
          &lt;td&gt;Audio input $32 / 1M tokens, cached input $0.40 / 1M tokens, audio output $64 / 1M tokens&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;code&gt;GPT-Realtime-Translate&lt;/code&gt;&lt;/td&gt;
          &lt;td&gt;$0.034 / minute&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;code&gt;GPT-Realtime-Whisper&lt;/code&gt;&lt;/td&gt;
          &lt;td&gt;$0.017 / minute&lt;/td&gt;
      &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;
&lt;p&gt;OpenAI also says the Realtime API supports EU Data Residency and is covered by its enterprise privacy commitments. For European businesses or products with data residency requirements, that is worth evaluating separately.&lt;/p&gt;
&lt;h2 id=&#34;what-this-means-for-developers&#34;&gt;What this means for developers
&lt;/h2&gt;&lt;p&gt;The key shift is that voice is becoming part of the product interaction layer, not just an input/output layer.&lt;/p&gt;
&lt;p&gt;In many earlier voice features, speech was converted to text, and text responses were converted back into speech. The hard middle layer is intent understanding, interruption handling, context tracking, tool calls, tool transparency, and graceful recovery.&lt;/p&gt;
&lt;p&gt;&lt;code&gt;GPT-Realtime-2&lt;/code&gt; tries to move more of that capability directly into the realtime voice model. For developers, the question is not only answer quality, but whether the model can support sustained conversations and multi-step tasks.&lt;/p&gt;
&lt;p&gt;Products that are especially worth testing include:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Customer support voice agents.&lt;/li&gt;
&lt;li&gt;In-car and mobile voice assistants.&lt;/li&gt;
&lt;li&gt;Travel, booking, real estate, finance, and other services that need conversation plus lookup.&lt;/li&gt;
&lt;li&gt;Multilingual meetings and cross-border communication tools.&lt;/li&gt;
&lt;li&gt;Live captions, meeting notes, and call quality systems.&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id=&#34;safety-and-disclosure-still-matter&#34;&gt;Safety and disclosure still matter
&lt;/h2&gt;&lt;p&gt;OpenAI says the Realtime API includes multiple safety layers, such as active classifiers over sessions and the ability to stop policy-violating conversations. Developers can also add their own guardrails through the Agents SDK.&lt;/p&gt;
&lt;p&gt;One easily missed requirement is disclosure: developers should make it clear when end users are interacting with AI, unless that is already obvious from the context.&lt;/p&gt;
&lt;p&gt;This matters in support, sales, education, healthcare, and similar scenarios. The more natural voice becomes, the more important product boundaries become: users should know they are talking to AI, and understand when speech may be recorded, transcribed, or used to trigger tools.&lt;/p&gt;
&lt;h2 id=&#34;summary&#34;&gt;Summary
&lt;/h2&gt;&lt;p&gt;OpenAI&amp;rsquo;s Realtime API update moves live voice from “can listen and speak” toward “can listen while working through tasks.”&lt;/p&gt;
&lt;p&gt;&lt;code&gt;GPT-Realtime-2&lt;/code&gt; handles more complex voice agents, &lt;code&gt;GPT-Realtime-Translate&lt;/code&gt; handles live cross-language communication, and &lt;code&gt;GPT-Realtime-Whisper&lt;/code&gt; handles low-latency transcription. Together, they cover the three basic capabilities most voice products need: conversation, translation, and transcription.&lt;/p&gt;
&lt;p&gt;If you are building support, in-car, meeting, education, cross-border communication, or mobile voice assistant products, this release is worth testing. The important question is not only whether the model sounds natural, but how it performs in long conversations, interruptions, tool calls, failure recovery, and cost control.&lt;/p&gt;
&lt;p&gt;Reference:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://openai.com/index/advancing-voice-intelligence-with-new-models-in-the-api/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;OpenAI: Advancing voice intelligence with new models in the API&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
</description>
        </item>
        
    </channel>
</rss>
