<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
    <channel>
        <title>Image Generation on Uncle Xiang&#39;s Notebook</title>
        <link>https://ttf248.life/en/tags/image-generation/</link>
        <description>Recent content in Image Generation on Uncle Xiang&#39;s Notebook</description>
        <generator>Hugo -- gohugo.io</generator>
        <language>en</language>
        <lastBuildDate>Sat, 25 Apr 2026 07:39:34 +0800</lastBuildDate><atom:link href="https://ttf248.life/en/tags/image-generation/index.xml" rel="self" type="application/rss+xml" /><item>
        <title>ChatGPT Images 2.0 is very powerful. Can we still trust [it] after taking a screenshot? / Is it credible just by looking at a screenshot?</title>
        <link>https://ttf248.life/en/p/chatgpt-images-2-screenshot-trust/</link>
        <pubDate>Sat, 25 Apr 2026 00:38:42 +0800</pubDate>
        
        <guid>https://ttf248.life/en/p/chatgpt-images-2-screenshot-trust/</guid>
        <description>&lt;p&gt;Initially, I didn&amp;rsquo;t actually plan on testing it. When I came across the news that OpenAI was releasing &lt;code&gt;ChatGPT Images 2.0&lt;/code&gt; on April 21, 2026, my first reaction was just &amp;ldquo;another image version update.&amp;rdquo; However, when I checked the Artificial Analysis leaderboard and saw that &lt;code&gt;GPT Image 2 (high)&lt;/code&gt; ranked first for text-to-image generation with an Elo of 1332, I felt a bit compelled to test it anyway.&lt;/p&gt;
&lt;p&gt;The results are quite impressive; the Chinese output is excellent, it can handle comics, and character/narrative consistency across multiple continuous images has also improved. However, as I tested it further, I felt that what is truly worth discussing this time isn&amp;rsquo;t &amp;ldquo;it draws better,&amp;rdquo; but rather &amp;ldquo;it starts making things that were previously taken as default truths seem unreliable.&amp;rdquo; This subject matter is more complicated than a simple leaderboard ranking.&lt;/p&gt;
&lt;h2 id=&#34;where-is-its-strength--what-is-its-strength-literal-translation&#34;&gt;Where is its strength? / What is its strength? (Literal translation)
&lt;/h2&gt;&lt;p&gt;&lt;strong&gt;or more naturally:&lt;/strong&gt;&lt;/p&gt;
&lt;h2 id=&#34;its-key-strengths&#34;&gt;Its Key Strengths
&lt;/h2&gt;&lt;p&gt;&lt;em&gt;(Since &amp;ldquo;先说它到底强在哪&amp;rdquo; is an introduction to discussing strengths, a heading like this is appropriate.)&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;OpenAI&amp;rsquo;s official name this time is &lt;code&gt;ChatGPT Images 2.0&lt;/code&gt;. The one corresponding in the leaderboard/list is &lt;code&gt;GPT Image 2 (high)&lt;/code&gt;. Don&amp;rsquo;t get them mixed up. The official introduction states very clearly that this generation isn&amp;rsquo;t just about improved image quality; it also incorporates a &lt;code&gt;thinking mode&lt;/code&gt;, allowing it to connect to live web search, generate multiple images from a single prompt, and even refine a very rough prompt into a final image that looks thoroughly researched and well-thought-out.&lt;/p&gt;
&lt;p&gt;This change is extremely noticeable when applied to Chinese scenarios. Previously, generated images with Chinese weren&amp;rsquo;t always unusable, but often required extensive and repeated retouching. The characters would be misplaced, the intended meaning would drift, and the style would be inconsistent throughout—comic panel layouts were particularly troublesome. After testing it, I found that the compliance of Chinese prompts has improved significantly. When creating comic pages, character concept sheets, or continuous narrative content, we are no longer dealing with &amp;ldquo;occasional successes,&amp;rdquo; but rather reaching a genuinely usable range.&lt;/p&gt;
&lt;p&gt;The official examples are also very straightforward, showcasing everything from Japanese comic pages and multilingual layouts to continuous multi-page narratives and dense textual infographics. To be honest, this isn&amp;rsquo;t a version that offers &amp;ldquo;small incremental improvements&amp;rdquo;; it&amp;rsquo;s a version where the workflow itself transforms. Many things that previously required ten rounds of revision can now achieve 70-80% completion in just one round.&lt;/p&gt;
&lt;h2 id=&#34;the-trouble-isnt-that-it-cant-draw-its-that-it-draws-too-well&#34;&gt;The trouble isn&amp;rsquo;t that it can&amp;rsquo;t draw; it&amp;rsquo;s that it draws too well
&lt;/h2&gt;&lt;p&gt;But the limitations have not disappeared.&lt;/p&gt;
&lt;p&gt;Through my own experiments, one very noticeable feeling is that the platform still has boundaries. If you try to prompt for specific styles, IPs, or sensitive scenarios involving real people, the system won&amp;rsquo;t just provide everything readily. OpenAI&amp;rsquo;s system card actually explains the reasons clearly: this generation, because of its stronger sense of realism, would more easily produce highly convincing deepfake content without additional safeguards, especially concerning real people, real locations, or real events.&lt;/p&gt;
&lt;p&gt;So, the weirdest/most contradictory part of this generation is right here.&lt;/p&gt;
&lt;p&gt;On one hand, it is significantly better than before—so much so that you can&amp;rsquo;t help but keep testing it, continuing to nitpick the details, and trying to compensate for things that were impossible before.&lt;/p&gt;
&lt;p&gt;On the other hand, it also needs to be stricter. So strict that for certain requirements, if you are still thinking of &amp;ldquo;bypassing&amp;rdquo; them through prompts, fundamentally, you are no longer using a tool for creation; you are wrestling with the content governance system itself.&lt;/p&gt;
&lt;p&gt;I won&amp;rsquo;t elaborate on the copyright legal details for this part; it&amp;rsquo;s too easy to get sidetracked. My judgment is simple: the boundary still exists, and in the future, it will only become more like boundaries in real society—it won&amp;rsquo;t automatically vanish just because the model gets stronger.&lt;/p&gt;
&lt;h2 id=&#34;will-screenshots-be-trustworthy-in-the-future&#34;&gt;Will Screenshots Be Trustworthy in the Future?
&lt;/h2&gt;&lt;p&gt;This is where I was genuinely stumped later on. / This was the part that truly amazed/confused me later.&lt;/p&gt;
&lt;p&gt;Previously, people were more concerned that AI images had too strong an &amp;ldquo;AI feel&amp;rdquo;—that they looked fake at first glance and could only fool those who weren&amp;rsquo;t paying attention. Now, the issue is different: many images are no longer questioned based on whether they look like AI drawings, but rather whether they look like screenshots you actually saw in a group chat yesterday.&lt;/p&gt;
&lt;p&gt;Receipts, chat logs, transfer pages, product dashboards, order interfaces—these items are not inherently artistic creations; their visual structures are very fixed, and their information density is low. For image models, this is conversely another target suitable for fabrication. As long as text stability, interface consistency, and local realism are maintained, creating a screenshot that &amp;ldquo;looks sufficiently authentic&amp;rdquo; has lowered the bar significantly.&lt;/p&gt;
&lt;p&gt;OpenAI mentioned two things in the system card: continuing to include &lt;code&gt;C2PA&lt;/code&gt; source metadata, and adding an invisible watermark. This direction is certainly correct; neglecting it would be insufficient. However, the problem remains very real: the images that ordinary people actually encounter are often not the original ones.&lt;/p&gt;
&lt;p&gt;It undergoes forwarding through WeChat, Weibo, Xiaohongshu, Moments, group chat sharing, cropping, compression, and secondary saving. After completing this cycle, whether the metadata persists, who will investigate it, or if the platform allows access to it—it is basically all question marks.&lt;/p&gt;
&lt;p&gt;In other words, while the technical sphere has begun to supplement &amp;ldquo;source proof,&amp;rdquo; the social sphere remains caught in the outdated habit of believing that &amp;ldquo;if you see it, it must be true&amp;rdquo; (or &amp;ldquo;seeing is believing&amp;rdquo;). This disparity—this gap—is what I find most troublesome.&lt;/p&gt;
&lt;p&gt;Going forward, a single screenshot will likely only count as a clue, not inherent evidence. If we want to be truly rigorous, we must review the original files, export records, timelines, context, and whether it can be cross-verified. The probative value of a single image is being gradually undermined by model capabilities.&lt;/p&gt;
&lt;h2 id=&#34;what-we-come-to-believe-in-the-future-may-not-be-pictures-but-rather-chains-of-evidence&#34;&gt;What we come to believe in the future may not be pictures, but rather chains of evidence
&lt;/h2&gt;&lt;p&gt;So, I feel a little conflicted about these types of models right now.&lt;/p&gt;
&lt;p&gt;It is genuinely powerful, and it&amp;rsquo;s not empty hype. Looking at the rankings as of April 25, 2026, &lt;code&gt;GPT Image 2 (high)&lt;/code&gt; is already in first place. The Chinese language support is good; it can handle comics/manga. Multi-image continuity and text control have all improved—these enhancements are real. For creators, operators, designers, and content makers, this is pure productivity.&lt;/p&gt;
&lt;p&gt;But the other side of the same coin is also true.&lt;/p&gt;
&lt;p&gt;As &amp;ldquo;looks very real&amp;rdquo; images become cheap and abundant, society can no longer rely on the old low-cost mechanisms of trust. Previously, we defaulted to assuming that screenshots were likely genuine; going forward, this default assumption must be lowered. This is especially true for items such as receipts, WeChat conversations, payment pages, and order records. I believe that in the future, we must always ask: What is the original source? And can it be independently verified?&lt;/p&gt;
&lt;p&gt;How should I put it? Previously, the problem with AI images was that they weren&amp;rsquo;t realistic enough. Now, the problem is that they are starting to become too real.&lt;/p&gt;
&lt;p&gt;This might be the threshold that &lt;code&gt;ChatGPT Images 2.0&lt;/code&gt; genuinely crossed. While its capability is undoubtedly powerful, what&amp;rsquo;s more problematic is that it has also brought into sharp focus the question of whether one can still trust a screenshot.&lt;/p&gt;
&lt;h2 id=&#34;references&#34;&gt;References
&lt;/h2&gt;&lt;ul&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://openai.com/index/introducing-chatgpt-images-2-0/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Introducing ChatGPT Images 2.0 | OpenAI&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://deploymentsafety.openai.com/chatgpt-images-2-0&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;ChatGPT Images 2.0 System Card | OpenAI Deployment Safety Hub&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://artificialanalysis.ai/image/leaderboard/text-to-image&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Text to Image Leaderboard | Artificial Analysis&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id=&#34;writing-notes&#34;&gt;Writing Notes
&lt;/h2&gt;&lt;h3 id=&#34;original-prompt&#34;&gt;Original Prompt
&lt;/h3&gt;&lt;blockquote&gt;
&lt;p&gt;ChatGPT released Image 2. I saw related news while scrolling, so at first I didn&amp;rsquo;t plan to test it. But when I checked the leaderboard, I realized how truly far ahead it is. When I tried it out, the results for Chinese were very good. It can create comics and has added reasoning capabilities, allowing it to output multiple continuous images at once. There are still copyright issues; it cannot directly generate anime-style pictures, so techniques must be used to bypass this. At first, I was only impressed by how strong this version&amp;rsquo;s image generation ability is, but later I realized that if it is &lt;em&gt;too&lt;/em&gt; realistic, that doesn&amp;rsquo;t work either—it makes people lose trust in screenshots. Receipts or WeChat screenshots—can we still believe them in the future?&lt;/p&gt;&lt;/blockquote&gt;
&lt;h3 id=&#34;summary-of-writing-ideas&#34;&gt;Summary of Writing Ideas
&lt;/h3&gt;&lt;ul&gt;
&lt;li&gt;Focus the opening on a personal trigger point—&amp;ldquo;originally didn&amp;rsquo;t intend to test this, but tried it after seeing the ranking list&amp;rdquo;—rather than starting with a product announcement.&lt;/li&gt;
&lt;li&gt;In the main body, first confirm where this generation&amp;rsquo;s capabilities are strongest. Focus on Chinese language support, multi-image continuity, and&lt;/li&gt;
&lt;/ul&gt;</description>
        </item>
        
    </channel>
</rss>
