GPT 5.5 “Spud” is not verified as an official public OpenAI model in the sources reviewed; OpenAI’s official model page, latest model guide, and model index point to GPT 5.4 instead [20][23][24]. Spud claims come from general web, social, or video sources, while OpenAI’s documented multimodal guidance here is for GP...

Create a landscape editorial hero image for this Studio Global article: GPT-5.5 Spud fact-check: OpenAI documents GPT-5.4, not Spud. Article summary: The official OpenAI docs provided here point to GPT 5.4—not GPT 5.5 “Spud”—as the documented model, so Spud’s image/document grounding claims remain unverified until OpenAI publishes official evidence [20][23][24].. Topic tags: openai, gpt 5, multimodal ai, document understanding, computer vision. Reference image context from search candidates: Reference image 1: visual subject "# OpenAI Spud: Leaked April 16 Release, Mythos-Level Benchmarks, and What GPT-5.5 or GPT-6 Might Mean. Claude Mythos Preview posted 77.80% on SWE-bench Pro. GPT-5.4 is at 57.70%. O" source context "OpenAI Spud: Leaked April 16 Release, Mythos-Level Benchmarks, and What GPT-5.5 or GPT-6 Might Mean - Adam Holter" Reference image 2: visual subject "# GPT 5.5 Spud Leak
Rumors about GPT-5.5 “Spud” focus on a high-value promise: a new OpenAI model with stronger multimodal grounding for images, forms, charts, scans, and documents. The verified record reviewed here is narrower. In the official OpenAI sources provided, the documented model is GPT-5.4—not GPT-5.5 or Spud [20][
23][
24].
That does not prove that “Spud” has never existed internally. It does mean that public claims about Spud’s release timing, benchmark performance, or superior image/document grounding are not established by the official evidence in this source set.
The strongest official evidence points to GPT-5.4. OpenAI’s GPT-5.4 model page describes GPT-5.4 as its frontier model for complex professional work, while OpenAI’s latest-model guide and model index also point readers to GPT-5.4 [20][
23][
24].
Studio Global AI
Use this topic as a starting point for a fresh source-backed answer, then compare citations before you share it.
GPT 5.5 “Spud” is not verified as an official public OpenAI model in the sources reviewed; OpenAI’s official model page, latest model guide, and model index point to GPT 5.4 instead [20][23][24].
GPT 5.5 “Spud” is not verified as an official public OpenAI model in the sources reviewed; OpenAI’s official model page, latest model guide, and model index point to GPT 5.4 instead [20][23][24]. Spud claims come from general web, social, or video sources, while OpenAI’s documented multimodal guidance here is for GPT 5.4 vision and document understanding [1][2][3][5][7][9][12].
For real image and document tasks, evaluate GPT 5.4 on your own forms, scans, charts, receipts, and multi page files, and preserve original image detail for dense or spatial workloads [22][37][38].
Continue with "Sui’s 2026 Confidential Transactions: Why Private On-Chain Payments Matter" for another angle and extra citations.
Open related pageCross-check this answer against "DroneShield’s ASIC Investigation: Why November Share Sales Still Worry Investors".
Open related pageDocs Guides, concepts, and product docs for CodexUse cases Example workflows and tasks teams hand to Codex. Models. Latest: GPT-5.4. Text generation. Code generation. Images and vision. [Structured output](…
GPT-5.5 Spud: Everything About OpenAI Next Frontier Model. GPT-5.5 Spud is OpenAI next frontier model: pretraining complete, Q2 2026 release expected. GPT-5.5 , code-named "Spud" , is the next frontier model from OpenAI. GPT-5.5 Spud OpenAI next AI model le...
OpenAI's GPT-5.5 'Spud' Is Coming: What We Know. OpenAI's GPT-5.5 'Spud' Is Coming: What We Know. OpenAI's next major AI model is nearly ready. Unlike the GPT-5.1 through 5.4 releases that refined and extended the GPT-5 base, Spud represents a completely ne...
Skip to main contentGPT-5.5: The Spud Leaks & The New Frontier of Omnimodal AI. Open menu Open navigation[]( to Reddit Home. Get App Get the Reddit app Log InLog in to Reddit. Go to ChatGPT. [r/ChatGPT]…
By contrast, the Spud-specific sources reviewed here are general-web articles, Reddit, X posts, and YouTube videos—not official OpenAI model pages, model guides, model cards, or benchmark reports [2][
3][
5][
7][
9][
12]. The safe conclusion is simple: GPT-5.5 Spud should be treated as a rumor or unverified label until OpenAI publishes official documentation.
| Claim | Status | What the sources support |
|---|---|---|
| GPT-5.5 “Spud” is an official public OpenAI model | Not verified | The official OpenAI sources reviewed here document GPT-5.4, not a GPT-5.5 or Spud model page [ |
| Spud is imminent or already validated | Unverified | The Spud references in this source set come from general-web or user-generated social/video sources [ |
| OpenAI has documented multimodal document workflows | Verified for GPT-5.4 | OpenAI provides GPT-5.4 vision and document-understanding guidance, plus prompt guidance for dense or spatial image tasks [ |
| Spud is better than GPT-5.4 at multimodal grounding | Not supported here | The reviewed official docs support GPT-5.4 guidance; they do not provide Spud-specific capability or benchmark evidence [ |
OpenAI’s official GPT-5.4 page says GPT-5.4 is its frontier model for complex professional work [20]. OpenAI also provides a GPT-5.4 cookbook page focused on vision and document understanding [
1]. In the retrieved material, that guidance is associated with examples such as structured extraction from a handwritten insurance form, spatial reasoning over an apartment floor plan, chart understanding, and bounding-box extraction from a police form [
1].
Those examples matter because real document work requires more than fluent summarization. A grounded model must connect its answer to visible evidence: field labels and values, table cells, chart marks, handwriting, document layout, and spatial position. Still, the GPT-5.4 material reviewed here is OpenAI-authored guidance and demonstration, not an independent audited benchmark report for every production document workflow [1][
20][
22].
OpenAI’s prompt guidance is also practical for evaluation. It recommends using original image detail for large, dense, or spatially sensitive images, especially computer use, localization, OCR, and click-accuracy tasks [22]. For forms, scans, screenshots, and charts, that means a workflow can lose accuracy if it downscales or strips away the details the model needs to inspect.
OCR asks a system to read text. Multimodal grounding asks it to connect text, layout, position, visual structure, and reasoning into an answer that can be checked against the page.
The research context supports that broader view. Document-understanding evaluation spans form understanding, receipt parsing, and document visual question answering [38]. Multi-page document VQA can require a model to reason across pages, navigate the document, retrieve relevant content, and inspect targeted pages rather than rely on a single image or page crop [
37].
That is why one impressive screenshot demo is not enough. A serious evaluation should cover the actual document types, scan quality, page count, handwriting, tables, charts, small text, and failure cases that match the intended workflow.
original image detail for dense, large, or spatially sensitive inputs such as OCR, localization, click-accuracy, and computer-use tasks [The name “Spud” appears in rumor-style coverage, but it is not verified as an official public OpenAI model in the sources reviewed here. The actionable conclusion is narrower: evaluate GPT-5.4 for OpenAI’s documented vision and document-understanding workflows, and treat GPT-5.5 Spud multimodal-grounding claims as unproven until OpenAI publishes an official model page, model guide, model card, or benchmark report [1][
20][
22][
23][
24].
- A New Foundation: Unlike incremental updates, GPT-5.5 (codenamed “Spud”) is rumored to be a completely new pre-trained base, built on nearly
BREAKING: OpenAI's GPT-5.5, nicknamed "Spud," is now projected to be released next week. GPT-5.5 released on...? polymarket.com.
BREAKING: OpenAI's GPT-5.5, internally nicknamed “Spud,” is now projected to launch as early as next week. In this episode: • What we know
Search the API docs. Realtime API. Model optimization. Specialized models. Legacy APIs. + Building frontend UIs with Codex and Figma. API. Building frontend UIs with Codex and Figma. GPT-5.4 is our frontier model for complex professional work. Learn more in...
Latest: GPT-5.4. Using tools. Skills. Shell. Computer use. + File search. + Overview. + Reasoning models. + Using realtime models. Use original for large, dense, or spatially sensitive images, especially [computer use, localization, OCR, and click-accuracy...
Latest: GPT-5.4. Using tools. Models and providers. Computer use. Reasoning models. Using realtime models. Latest: GPT-5.4. [Using tools](h…
Overview. Models. Latest: GPT-5.4. Text generation. Using tools. Overview. Models and providers. Running agents. [Evaluate agent…
Doc-V∗V^{\ } begins with a Global Thumbnail Overview that provides a low-cost structural prior, and then alternates between structured visual reasoning and document navigation actions, including semantic retrieval and targeted page fetching. Motivated by th...
Document understanding benchmarks span form understanding, receipt parsing, and document VQA, including FUNSD (Jaume et al., 2019) , SROIE (