<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
    <channel>
        <title>GPT-5.5 on KnightLi Blog</title>
        <link>https://www.knightli.com/en/tags/gpt-5.5/</link>
        <description>Recent content in GPT-5.5 on KnightLi Blog</description>
        <generator>Hugo -- gohugo.io</generator>
        <language>en</language>
        <lastBuildDate>Fri, 24 Apr 2026 08:39:56 +0800</lastBuildDate><atom:link href="https://www.knightli.com/en/tags/gpt-5.5/index.xml" rel="self" type="application/rss+xml" /><item>
        <title>OpenAI Releases GPT-5.5: Stronger Agentic Coding, Knowledge Work, and Research</title>
        <link>https://www.knightli.com/en/2026/04/24/openai-gpt-5-5-release/</link>
        <pubDate>Fri, 24 Apr 2026 08:39:56 +0800</pubDate>
        
        <guid>https://www.knightli.com/en/2026/04/24/openai-gpt-5-5-release/</guid>
        <description>&lt;p&gt;OpenAI published &lt;a class=&#34;link&#34; href=&#34;https://openai.com/index/introducing-gpt-5-5/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Introducing GPT-5.5&lt;/a&gt; on April 23, 2026. Judging from the official page, this release is not just about making the model &amp;ldquo;smarter&amp;rdquo;; it is more about whether the model can keep pushing complex tasks forward.&lt;/p&gt;
&lt;p&gt;OpenAI positions GPT-5.5 as a model better suited for real work. It is expected not only to answer questions, but also to write code, debug, research online, analyze data, create documents and spreadsheets, operate software, and move across tools until the task is finished.&lt;/p&gt;
&lt;h2 id=&#34;1-where-gpt-55-is-strongest&#34;&gt;1. Where GPT-5.5 Is Strongest
&lt;/h2&gt;&lt;p&gt;The release page repeatedly highlights four areas:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Agentic coding&lt;/li&gt;
&lt;li&gt;Computer use and tool use&lt;/li&gt;
&lt;li&gt;Knowledge work&lt;/li&gt;
&lt;li&gt;Early scientific research assistance&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;In other words, GPT-5.5 is aimed less at short Q&amp;amp;A and more at long-running tasks. For example, an engineering problem is not just &amp;ldquo;how should this code be changed&amp;rdquo;; the model needs to understand the project structure, locate the cause of failure, edit related files, add tests, verify results, and reduce repeated prompting from the user.&lt;/p&gt;
&lt;p&gt;OpenAI also emphasizes that GPT-5.5 uses fewer tokens in Codex tasks. This matters in practice because coding agents can consume tokens quickly once they start reading files, running commands, and fixing bugs. If a model can complete the same task in fewer steps, both cost and waiting time go down.&lt;/p&gt;
&lt;h2 id=&#34;2-coding-is-the-main-showcase&#34;&gt;2. Coding Is the Main Showcase
&lt;/h2&gt;&lt;p&gt;OpenAI calls GPT-5.5 its strongest agentic coding model to date.&lt;/p&gt;
&lt;p&gt;The most notable public numbers include:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;code&gt;Terminal-Bench 2.0&lt;/code&gt;: GPT-5.5 reaches &lt;code&gt;82.7%&lt;/code&gt;&lt;/li&gt;
&lt;li&gt;&lt;code&gt;SWE-Bench Pro&lt;/code&gt;: GPT-5.5 reaches &lt;code&gt;58.6%&lt;/code&gt;&lt;/li&gt;
&lt;li&gt;OpenAI&amp;rsquo;s internal &lt;code&gt;Expert-SWE&lt;/code&gt;: GPT-5.5 also scores higher than GPT-5.4&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;These evaluations have one thing in common: they are closer to real engineering workflows than isolated algorithm questions. Terminal-Bench in particular involves command-line operations, planning, trial and error, tool coordination, and multi-step verification.&lt;/p&gt;
&lt;p&gt;For everyday developers, the implication is direct: whether a model can take on larger tasks depends on whether it can hold context for a long time, check its own assumptions, know when to run tests, and understand what else a change may affect.&lt;/p&gt;
&lt;p&gt;The value of GPT-5.5 in Codex also shows up mainly in these behaviors. It feels more like a collaborator that can take over part of an engineering task, rather than a tool that only completes code fragments.&lt;/p&gt;
&lt;h2 id=&#34;3-knowledge-work-becomes-a-core-scenario&#34;&gt;3. Knowledge Work Becomes a Core Scenario
&lt;/h2&gt;&lt;p&gt;Beyond coding, OpenAI is placing GPT-5.5 into a broader office-work context.&lt;/p&gt;
&lt;p&gt;The announcement says GPT-5.5 can generate documents, spreadsheets, and slide decks better in Codex, and is also better suited for operational research, spreadsheet modeling, and organizing business materials. Combined with computer use, its goal is not merely to offer suggestions, but to participate in the full workflow of finding information, understanding content, using tools, checking output, and turning raw material into a result.&lt;/p&gt;
&lt;p&gt;The page also notes that OpenAI already uses Codex across many internal departments, including software engineering, finance, communications, marketing, data science, and product management. The interesting part is not any single example, but the direction: OpenAI is expanding Codex from a developer tool into a more general work tool.&lt;/p&gt;
&lt;p&gt;In ChatGPT, GPT-5.5 Thinking is available to Plus, Pro, Business, and Enterprise users; GPT-5.5 Pro is aimed at harder questions and higher-accuracy work, and is available to Pro, Business, and Enterprise users.&lt;/p&gt;
&lt;h2 id=&#34;4-research-capability-is-more-than-better-answers&#34;&gt;4. Research Capability Is More Than Better Answers
&lt;/h2&gt;&lt;p&gt;GPT-5.5 also receives a strong research-focused presentation.&lt;/p&gt;
&lt;p&gt;OpenAI says it has improved in genetics, quantitative biology, bioinformatics, mathematical proof, and related areas. The key is not whether the model can recall a fact, but whether it can handle more realistic research problems: reading data, spotting anomalies, proposing analyses, interpreting results, and continuing based on intermediate findings.&lt;/p&gt;
&lt;p&gt;The release page mentions &lt;code&gt;GeneBench&lt;/code&gt; and &lt;code&gt;BixBench&lt;/code&gt;, both of which focus more on multi-stage scientific analysis. OpenAI also says an internal version of GPT-5.5, with a custom harness, helped discover a new proof related to Ramsey numbers and verified it with Lean.&lt;/p&gt;
&lt;p&gt;These examples should not be simplified into &amp;ldquo;AI can now do research independently.&amp;rdquo; But they do suggest that models are moving from answer engines toward research collaborators. In scenarios where code, data, papers, experiment ideas, and notes are mixed together, GPT-5.5&amp;rsquo;s long-horizon reasoning and tool use become especially important.&lt;/p&gt;
&lt;h2 id=&#34;5-inference-efficiency-stronger-without-getting-much-slower&#34;&gt;5. Inference Efficiency: Stronger Without Getting Much Slower
&lt;/h2&gt;&lt;p&gt;One easily overlooked point is that OpenAI says GPT-5.5 matches GPT-5.4 in real-world per-token latency.&lt;/p&gt;
&lt;p&gt;Normally, larger and more capable models bring higher latency. This time, OpenAI emphasizes that inference-system optimization helped GPT-5.5 become more capable while keeping speed stable. The release page also mentions that Codex analyzed production traffic patterns and wrote load-balancing heuristic algorithms, increasing token generation speed by more than &lt;code&gt;20%&lt;/code&gt;.&lt;/p&gt;
&lt;p&gt;That detail is interesting: the model is not only served by infrastructure, but also helps improve the infrastructure that serves it.&lt;/p&gt;
&lt;h2 id=&#34;6-safety-gets-stricter-especially-around-cybersecurity&#34;&gt;6. Safety Gets Stricter, Especially Around Cybersecurity
&lt;/h2&gt;&lt;p&gt;Because GPT-5.5 has stronger cybersecurity capabilities, OpenAI is also tightening safety controls.&lt;/p&gt;
&lt;p&gt;The announcement says GPT-5.5 improves over GPT-5.4 in cybersecurity capability, so OpenAI is deploying stricter classifiers, especially for high-risk activity, sensitive cybersecurity requests, and repeated misuse.&lt;/p&gt;
&lt;p&gt;This means some users may see more refusals or friction when working on cybersecurity-related tasks. OpenAI also offers Trusted Access for Cyber, intended to reduce unnecessary barriers for verified defensive users.&lt;/p&gt;
&lt;p&gt;For ordinary developers, the simple takeaway is: legitimate security hardening, vulnerability fixing, and code auditing should continue to be supported, while high-risk attack workflows will be more tightly controlled.&lt;/p&gt;
&lt;h2 id=&#34;7-availability-and-api-pricing&#34;&gt;7. Availability and API Pricing
&lt;/h2&gt;&lt;p&gt;According to OpenAI&amp;rsquo;s release page, GPT-5.5 availability is as follows:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;ChatGPT: GPT-5.5 Thinking for Plus, Pro, Business, and Enterprise users&lt;/li&gt;
&lt;li&gt;ChatGPT: GPT-5.5 Pro for Pro, Business, and Enterprise users&lt;/li&gt;
&lt;li&gt;Codex: GPT-5.5 for Plus, Pro, Business, Enterprise, Edu, and Go plans&lt;/li&gt;
&lt;li&gt;Codex: &lt;code&gt;400K&lt;/code&gt; context window&lt;/li&gt;
&lt;li&gt;Codex Fast mode: about &lt;code&gt;1.5x&lt;/code&gt; token generation speed at &lt;code&gt;2.5x&lt;/code&gt; the cost&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;For the API, OpenAI says &lt;code&gt;gpt-5.5&lt;/code&gt; and &lt;code&gt;gpt-5.5-pro&lt;/code&gt; will be available soon.&lt;/p&gt;
&lt;p&gt;The announced API prices are:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;code&gt;gpt-5.5&lt;/code&gt;: &lt;code&gt;US$5 / 1M tokens&lt;/code&gt; input and &lt;code&gt;US$30 / 1M tokens&lt;/code&gt; output&lt;/li&gt;
&lt;li&gt;&lt;code&gt;gpt-5.5-pro&lt;/code&gt;: &lt;code&gt;US$30 / 1M tokens&lt;/code&gt; input and &lt;code&gt;US$180 / 1M tokens&lt;/code&gt; output&lt;/li&gt;
&lt;li&gt;&lt;code&gt;gpt-5.5&lt;/code&gt; API context window: &lt;code&gt;1M&lt;/code&gt;&lt;/li&gt;
&lt;li&gt;Batch and Flex are half the standard API price&lt;/li&gt;
&lt;li&gt;Priority processing is &lt;code&gt;2.5x&lt;/code&gt; the standard price&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This is clearly more expensive than many everyday models, so it is better suited for high-value tasks: complex engineering changes, long-document analysis, office automation, research assistance, and important business workflows, rather than casual chat.&lt;/p&gt;
&lt;h2 id=&#34;8-how-to-read-this-release&#34;&gt;8. How to Read This Release
&lt;/h2&gt;&lt;p&gt;In one sentence, GPT-5.5 is about OpenAI pushing models further from &amp;ldquo;answering questions&amp;rdquo; toward &amp;ldquo;getting work done.&amp;rdquo;&lt;/p&gt;
&lt;p&gt;The most important part is not just higher benchmark scores, but the convergence of several capabilities:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Better long-task persistence&lt;/li&gt;
&lt;li&gt;More reliable tool use&lt;/li&gt;
&lt;li&gt;Stronger engineering context understanding&lt;/li&gt;
&lt;li&gt;Better fit for documents, spreadsheets, research, and business workflows&lt;/li&gt;
&lt;li&gt;Longer context and higher token efficiency&lt;/li&gt;
&lt;li&gt;Stricter controls around high-risk capabilities&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;For developers, the most interesting thing to test is complex engineering work in Codex. For enterprise users, the bigger question is whether it can turn some cross-tool, cross-document, cross-process work into deliverable output.&lt;/p&gt;
&lt;p&gt;GPT-5.5 is not a small update aimed only at chat experience. It looks more like another step in OpenAI&amp;rsquo;s push toward AI as an execution layer for work.&lt;/p&gt;
&lt;h2 id=&#34;related-links&#34;&gt;Related Links
&lt;/h2&gt;&lt;ul&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://openai.com/index/introducing-gpt-5-5/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Introducing GPT-5.5 - OpenAI&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
</description>
        </item>
        
    </channel>
</rss>
