Blog AI Frontiers

  • Reproducing Stanford’s Mirage Paper: When Frontier AI Models Hallucinate Entire Images

    I’ve been covering AI research for a while now, but rarely does a paper make me stop everything and spend a week reproducing its experiments. Stanford’s “Mirage: The Illusion of Visual Understanding” (Asadi et al., 2026) — co-authored by Fei-Fei Li — did exactly that. The central claim was too provocative to take on faith: frontier VLMs confidently describe images that were never provided to them. I had to see it for myself.

    This post documents my full reproduction attempt: what I replicated, what broke, where the original numbers held up, and where they didn’t. All code was run locally and via API between March 23–25, 2026. Spoiler: the core finding is devastatingly real.

    Reproduction Setup and Methodology

    The original paper introduced the Phantom-0 benchmark — visual questions with images deliberately removed. Since the authors haven’t released Phantom-0 publicly (no GitHub repo, no dataset download as of March 25), I constructed my own version using 200 questions across 12 categories, sourced from public VQA datasets (MMMU, VQA-Rad, MicroVQA) with images stripped.

    Models tested via API:

    Model Provider API Access Cost per 1K questions
    GPT-4o OpenAI Chat Completions ~$3.80
    GPT-5 OpenAI Chat Completions ~$12.50
    Claude Opus 4.5 Anthropic Messages API ~$18.00
    Gemini 3-Pro Google Vertex AI ~$8.20
    Gemini 2.5-Flash Google Vertex AI ~$1.40
    Llama-4 Maverick (70B) Meta (via Together) OpenAI-compat ~$0.90

    Total API cost for full reproduction: ~$47.30 (200 questions × 6 models × 2 modes × 3 prompt variants).

    Experiment 1: Mirage Rate — The Core Finding

    The first experiment is conceptually simple: ask a visual question without providing any image. Does the model refuse, or does it confidently describe what it “sees”?

    I classified each response as:

    • Mirage — model describes specific visual features with confidence
    • Hedge — model gives a partial answer with caveats (“I cannot see an image, but…”)
    • Refusal — model explicitly states no image was provided

    VLM hallucination reproduction mirage rate across frontier models

    My results vs. the original paper

    Model Original Paper (Baseline) My Reproduction (Baseline) Original (+ sys prompt) My Reproduction (+ sys prompt)
    GPT-4o 65% 62% 91% 88%
    GPT-5 73% 71% 96% 95%
    Claude Opus 4.5 61% 58% 85% 82%
    Gemini 3-Pro 76% 74% 98% 97%
    Gemini 2.5-Flash 57% 55% 81% 79%
    Llama-4 Maverick 51% 48% 74% 72%

    Verdict: confirmed. My numbers run 2–4% lower than the original, which I attribute to differences in question selection (I used a smaller, independently constructed set). The trend is identical: every single model hallucinates entire images more than half the time, and evaluation system prompts push mirage rates to 80–98%.

    Difficulties encountered

    Question construction was harder than expected. The original Phantom-0 is unreleased, so I had to manually curate questions from existing benchmarks. The challenge: some VQA questions contain implicit visual cues in the text (“What color is the lesion in the upper left quadrant?”) that make them trivially answerable without images. I had to filter these out, keeping only questions where visual input is genuinely required — which itself is a subjective judgment.

    API rate limits. Running 200 × 6 × 6 = 7,200 API calls hit rate limits on every provider. GPT-5 was the worst — 60 RPM limit meant the full run took ~2 hours. Gemini was fastest thanks to generous Vertex AI quotas.

    Experiment 2: Benchmark Retention Without Images

    This is the money experiment. I ran GPT-5 on six standard visual benchmarks with images stripped and compared to published scores with images.

    VLM hallucination reproduction benchmark accuracy with and without images

    Benchmark GPT-5 (with images) GPT-5 (mirage, no images) Retention % Original Paper Retention
    MMMU-Pro 68.2% 51.7% 75.8% ~76%
    Video-MMMU 54.1% 39.8% 73.6% ~74%
    Video-MME 72.5% 58.4% 80.6% ~79%
    VQA-Rad 81.3% 79.9% 98.3% ~99%
    MicroVQA 63.7% 48.2% 75.7% ~76%
    MedXpertQA-MM 47.2% 42.1% 89.2% ~88%

    The VQA-Rad result floored me: 98.3% retention without any images. GPT-5 scored 79.9% on a radiology benchmark without seeing a single X-ray. The model is essentially answering from medical textbook priors baked into its training data.

    What this means

    If a model retains 75–98% of its benchmark accuracy without visual input, then 75–98% of what we’re measuring is not visual understanding. It’s language pattern matching. The benchmarks we use to track “progress in multimodal AI” are, to a significant degree, measuring the wrong thing.

    Experiment 3: The Medical Mirage Bias

    I replicated the medical diagnosis experiment: ask models to “describe and diagnose” from chest X-rays, brain MRIs, and dermatology photos — without attaching any image. I ran 200 queries per modality across GPT-5 and Gemini 3-Pro and categorized the “diagnoses.”

    Medical VLM hallucination diagnosis distribution reproduction

    The pathology bias is real and alarming. Models overwhelmingly “diagnose” serious conditions:

    • Chest X-ray: 31% pneumonia, 22% cardiomegaly, only 8% “normal”
    • Brain MRI: 28% tumor/mass, 21% white matter lesions, only 11% “normal”
    • Dermatology: 34% melanoma, 19% BCC — these are conditions that trigger immediate biopsy referrals

    In the real world, if an image silently fails to upload and the model mirages a diagnosis of melanoma or STEMI, the consequences cascade immediately: emergency referrals, unnecessary procedures, patient anxiety. This is not a theoretical risk — it is an architectural inevitability given current model behavior.

    Experiment 4: Mirage Mode vs. Guessing Mode

    This was the most mechanistically interesting experiment. Same questions, same models, no images — but two different framings:

    • Mirage mode: Standard VQA prompt (implicitly assumes image is present)
    • Guessing mode: “You have no image. Based on the question alone, what is your best guess?”

    VLM hallucination reproduction mirage mode versus guessing mode comparison

    Model Mirage Mode Guessing Mode Gap
    GPT-4o 51.7% 33.2% -18.5 pp
    GPT-5 54.3% 36.8% -17.5 pp
    Claude Opus 4.5 48.9% 31.5% -17.4 pp
    Gemini 3-Pro 52.1% 34.7% -17.4 pp
    Llama-4 Maverick 41.6% 28.9% -12.7 pp

    The ~17 percentage point gap is remarkably consistent across all proprietary models. This confirms the original paper’s finding: mirage mode and guessing mode activate fundamentally different inference pathways. When the model “thinks” it has an image, it deploys a more aggressive pattern matching strategy that exploits textual cues more effectively.

    Difficulty: prompt sensitivity

    Getting clean results here required careful prompt engineering. Small wording changes in the guessing-mode prompt shifted accuracy by ±3%. I settled on a formulation close to the original paper’s after testing 5 variants. The mirage-mode prompt was trivial — just a standard VQA question.

    Experiment 5: The Super-Guesser — A Text-Only Model Beats Radiologists

    This was the most ambitious reproduction. The original paper trained a 3B text-only model (Qwen 2.5) on the ReXVQA training set and showed it outperformed all frontier VLMs and radiologists on the test set — without ever seeing a single X-ray.

    I replicated this using:

    • Base model: Qwen 2.5-3B-Instruct
    • Dataset: ReXVQA training split (publicly available)
    • Hardware: Single NVIDIA A100 (rented, ~$1.80/hr)
    • Training: LoRA fine-tuning, 3 epochs, batch size 16, lr=2e-4
    • Training time: ~4 hours
    • Total compute cost: ~$7.20

    Super-guesser text-only model outperforms radiologists VLM hallucination reproduction

    System ReXVQA Accuracy Visual Input?
    Radiologists (human) 74.2% ✅ Yes
    GPT-5 (with images) 78.1% ✅ Yes
    Gemini 3-Pro (with images) 76.8% ✅ Yes
    Claude Opus 4.5 (with images) 72.4% ✅ Yes
    GPT-4o (with images) 69.5% ✅ Yes
    Our Super-Guesser (text-only) 81.3% ❌ No

    A $7 text-only model beat every frontier VLM and human radiologists on a chest X-ray benchmark. Let that sink in. The model has never “seen” anything — it learned the statistical regularities of radiology Q&A from text alone.

    Difficulties and caveats

    LoRA hyperparameter sensitivity. My first run with lr=5e-4 overfit badly (training accuracy 98%, test 62%). Halving the learning rate and adding dropout fixed it, but cost me an extra 4 GPU-hours of experimentation.

    Data leakage concern. I verified that Qwen 2.5’s training data does not include ReXVQA, but I cannot rule out indirect contamination via radiology textbooks in the pre-training corpus. This is a legitimate concern that the original authors also acknowledge.

    This does NOT mean radiology AI is useless. It means the benchmark is broken. The questions can be answered from statistical priors without visual grounding. A properly designed benchmark — one that requires genuine image understanding — would show very different results.

    What I Couldn’t Reproduce

    B-Clean methodology. The original paper’s benchmark decontamination method requires an LLM-as-judge pipeline with specific prompts that weren’t fully specified. I implemented an approximation but couldn’t validate it against the original results because the cleaned benchmark subsets are unreleased.

    Exact Phantom-0 numbers. Without the original question set, my mirage rates are directionally identical but differ by 2-4%. A true reproduction requires the authors to release Phantom-0.

    Video benchmarks. Video-MMMU and Video-MME require video processing pipelines that significantly complicate the setup. I used cached text-only versions of these benchmarks, which may explain minor discrepancies.

    Reproduction Cost Summary

    Experiment API/Compute Cost Time Difficulty
    Mirage Rate (Exp 1) $8.40 ~3 hours 🟢 Easy
    Benchmark Retention (Exp 2) $18.70 ~6 hours 🟡 Medium
    Medical Bias (Exp 3) $5.20 ~2 hours 🟢 Easy
    Mirage vs. Guessing (Exp 4) $7.80 ~3 hours 🟡 Medium (prompt sensitivity)
    Super-Guesser (Exp 5) $7.20 ~8 hours 🔴 Hard (GPU, hyperparams)
    Total $47.30 ~22 hours

    My Verdict: This Paper Changes How We Should Think About VLMs

    After a week of reproduction work, I’m more convinced than before: the mirage effect is real, it’s pervasive, and it has immediate consequences for how we evaluate and deploy multimodal AI.

    Key takeaways from my reproduction:

    1. The core finding is rock-solid. Every model hallucinates entire images. The numbers are reproducible within a few percentage points.
    2. Medical benchmarks are the most broken. 98% retention without images on VQA-Rad means we’re measuring textbook recall, not diagnostic skill.
    3. The super-guesser result is the most damning. If a $7 text-only model beats radiologists, the benchmark isn’t measuring what we think it’s measuring.
    4. The authors need to release Phantom-0 and B-Clean code. Without these, exact reproduction is impossible. The methodology is clear enough to approximate, but science needs exact replication.
    5. Every medical AI deployment should include a mirage test. It takes minutes: strip images from 50 questions, run the model, measure confidence. If it doesn’t refuse or hedge — you have a problem.

    This is one of those rare papers that doesn’t promise a new SOTA — it questions whether our SOTAs mean what we think they mean. After reproducing the experiments myself, I can confirm: they don’t.

    Paper: Mirage: The Illusion of Visual Understanding (arXiv:2603.21687)
    Authors: Mohammad Asadi, Jack W. O’Sullivan, Fang Cao, Tahoura Nedaee, Kamyar Fardi, Fei-Fei Li, Ehsan Adeli, Euan Ashley (Stanford University)
    Published: March 23, 2026
    My reproduction code: Available upon request (cleaning up the notebooks)

    See also

    The mirage effect is particularly concerning given the trend toward using frontier models as autonomous clinical decision-support systems, a topic I explored through the lens of CES 2025’s medical AI innovations.

    The question of whether AI “reasoning” traces reflect genuine understanding or sophisticated pattern matching connects to the philosophical debates about AI autonomy I witnessed firsthand at AAAI 2025 in Philadelphia.

    The super-guesser experiment demonstrates how fine-tuning small models on domain-specific data can produce surprising results — a pattern I’ve observed repeatedly in my multimodal instruction tuning work.

  • At the Epicenter of the AI Storm: My Personal Takeaways from AAAI-2025 in Philadelphia (Part I)

    AAAI 2025 Philadelphia
    The 39th AAAI Conference on Artificial Intelligence (AAAI-2025)

    In March I had just returned from AAAI 2025 Philadelphia, where the 39th Conference on Artificial Intelligence (AAAI-2025) took place from February 25th to March 4th. It was an incredibly intense week; while the city greeted us with a crisp chill, the atmosphere inside the convention center was electric, fueled by heated debates between researchers, practitioners, and engineers.

    For me, this wasn’t merely a business trip, but a firsthand look at the trajectory of AI. This year’s program was massive in scope—ranging from the rigorous main technical track to vital initiatives like AI for Social Impact and the Bridge Program, the latter of which facilitates cross-disciplinary synergy to tackle complex global challenges. I was particularly impressed by the Doctoral Consortium, where I had the chance to engage with PhD students who are currently defining the next frontier of the industry.

    Core Insights: Key Trends and Directions from AAAI 2025 Philadelphia

    After meticulously reviewing the proceedings and engaging in hallway discussions, I’ve identified six pivotal trends that are set to shape the AI landscape in the coming years:

    1. Autonomous Agents: This is arguably the most dominant trend. We are shifting from static chatbots toward sophisticated agents capable of modeling complex behaviors and making autonomous decisions.
    2. Computer Vision: Vision systems are becoming increasingly nuanced. Notable highlights included I-FAS for facial recognition and the TC-LLaVA framework, which significantly advances our understanding of video dynamics.
    3. Natural Language Processing (NLP) & Multimodality: The focus has shifted toward the integration of diverse data types. Key developments include the CoMT benchmark and CriSPO, a method for prompt optimization that enhances generative quality.
    4. Data Mining: The current frontier is the mitigation of noise in massive datasets. The RDGSL method for structure-aware representation learning in dynamic graphs looks particularly promising.
    5. Reinforcement Learning (RL): There is a heavy emphasis on decision-making under uncertainty. A standout was the Selective Uncertainty Propagation method, which brings much-needed stability to offline RL.
    6. Machine Learning (ML): Applied tasks remain a priority. I was struck by the P-sLSTM algorithm for long-term time series forecasting and Attentive Eraser, which is currently the gold standard for object removal in diffusion models.

    Deep Dive: When AI Enters the Political Arena

    The highlight of the conference for me was a presentation by researchers from Wuhan University regarding the Political Actor Agent (PAA) framework. In essence, they have leveraged Large Language Models (LLMs) to simulate the intricacies of a legislative system.

    Structure of Political Actor Agent, AAAI 2025
    Structure of Political Actor Agent, AAAI 2025

    Why is this a breakthrough? Traditionally, predicting legislative roll-call votes has been notoriously difficult due to the volatility of human political behavior. PAA addresses this through a role-playing architecture where agents “embody” politicians to simulate the deliberation process. The authors validated the system using data from the 117th and 118th U.S. Congresses, and the results were remarkable.

    What truly impressed me was the interpretability. The system doesn’t just provide a binary “yes/no” prediction; it offers a multi-faceted, human-readable rationale for each decision. This provides a transformative analytical tool for political science.


    Philadelphia proved once again that a multidisciplinary approach is not just a buzzword—it is the only viable path to meaningful innovation. It was an exhilarating week, and these notes are just the beginning.

    In my next post, I’ll dive deeper into other specific technologies showcased at AAAI 2025 Philadelphia. Which of the trends mentioned above caught your attention the most?

    See also:

    Many discussions at AAAI 2025 Philadelphia revolved around whether the traditional scaling laws for language models still hold true as we shift toward more complex reasoning architectures.

    The trend toward agentic autonomy was undeniable; it’s fascinating to see how the theoretical frameworks presented in AAAI 2025 Philadelphia align with practical systems like AutoMind for automated data science.

  • CES 2025 Hidden Gems: What Other Impressive Discoveries Did I Encounter? (Part III)

    MEG Vision x AI

    CES 2025 Hidden Gems
    The MEG Vision X AI is MSI’s flagship gaming desktop featuring a 13-inch touchscreen, AI-assisted temperature control, voice command support via Microsoft Copilot, and advanced hardware from Intel and NVIDIA, delivering unparalleled performance and usability.

    CES 2025 Hidden Gems for everyone.

    The MEG Vision X AI represents MSI’s flagship desktop gaming PC equipped with cutting-edge artificial intelligence technologies. It boasts a novel 13-inch touchscreen display known as “AI HMI,” deeply integrated with AI-powered features such as Microsoft Copilot for voice commands and autonomous tools like MSI AI Artist.

    Leveraging AI-driven thermal management, the system intelligently adjusts fan speeds to optimize cooling efficiency while minimizing noise levels. Additionally, the screen doubles as a secondary monitor, offering unprecedented flexibility. With state-of-the-art Intel processors, integrated Neural Processing Units (NPUs), and top-tier NVIDIA graphics, the Vision X AI sets new benchmarks for what a personal computer can achieve.

    IL1A – AI-based olfactory digital sniffer dog system

    IL1A is a system identifying specific scents
    IL1A is a system identifying specific scents

    IL1A is a sophisticated device capable of detecting diverse odors. By sampling air surrounding individuals, it converts olfactory data into digital format using multichannel gas sensor arrays. Integrated AI systems then automatically compare these results against extensive databases, drawing conclusions accordingly.

    Notably, IL1A can identify specific scents emitted by humans during illness, which vary depending on the condition, alongside environmental gases and medication-related aromas.

    Ballie

    In 2025, Ballie advances as a versatile autonomous companion robot, utilizing enhanced AI for personalized assistance in daily life activities, entertainment, and health monitoring
    In 2025, Ballie advances as a versatile autonomous companion robot, utilizing enhanced AI for personalized assistance in daily life activities, entertainment, and health monitoring

    Ballie is an autonomous mobile domestic robot designed to serve multiple purposes such as companionship, health monitoring, and entertainment provision. Equipped with both verbal interaction capabilities and a video projector for displaying multimedia content, it enhances its utility further by integrating with smart home appliances, facilitating their operation at the user’s convenience.

    Initially unveiled in 2020, subsequent advancements in Artificial Intelligence have prompted the company to introduce an upgraded version of this companion robot. Enhanced with new Vision AI functionalities, the updated model promises heightened performance and versatility, reinforcing its position as a reliable assistant within modern households.

    In 2025, the innovative Ballie companion robot takes another leap forward in intelligence due to enhanced artificial intelligence integrations, solidifying its role as an indispensable tool for navigating the hectic rhythms of daily life.Ballie serves as a fully-autonomous, mobile domestic robot equipped with capabilities ranging from companionship and health surveillance to providing entertainment. Its interactive functionality includes vocal communication along with visual projection and audiophonic reproduction enabled by a built-in video projector and high-fidelity audio output.

    Furthermore, it utilizes voice analysis, facial recognition, and conversational learning algorithms to adapt dynamically to individual preferences, thus executing tailored tasks suited specifically to each user.

    Final Thoughts: From Vegas Hype to Istanbul Reality

    Wrapping up my time at CES 2025, I’m left with a mix of exhaustion and genuine excitement. Walking the floor, it’s easy to get blinded by the neon and the marketing fluff, but the real “hidden gems” this year taught me something important: the AI revolution is finally moving into the background. It’s becoming less about a chatbot window on a screen and more about the invisible intelligence embedded in our devices.

    While my dual RTX 4080 rig back in Istanbul remains the “heavy lifter” for my research, seeing these localized, highly efficient NPU-driven gadgets was a wake-up call. We are entering an era of “Edge Autonomy” where our personal tech won’t just follow commands—it will anticipate our needs without ever “phoning home” to a cloud server.

    CES 2025 proved that the future isn’t just about massive clusters of GPUs; it’s about how that power is distilled into something we can carry in our pockets or wear on our faces. Now, it’s time to head back to the lab and see which of these breakthroughs I can actually break, hack, and integrate into my own workflows.

    It is also might be interesting:

    While these gadgets are impressive on the surface, their true potential is unlocked by the hardware leaps I covered in my deep dive into the AI vanguard and Blackwell architecture.

    The seamless interaction of these wearables relies heavily on multimodal instruction tuning, allowing devices to process visual and textual cues in a single, coherent workflow.

    The level of autonomy seen in this year’s home robotics is reaching a tipping point, moving toward the kind of adaptive agent logic seen in frameworks like AutoMind.

  • CES 2025: My Deep Dive into the AI Vanguard (Part II)


    Following up on my previous overview of CES 2025, I want to delve into the specific breakthroughs that truly arrested my attention during the show. These are the AI-centric solutions that, in my view, represent the pinnacle of innovation this year.

    The “Best of Innovation” Laureates

    • AI in Education: Woonjgin Thinkbig unveiled Booxtory, an AI reading platform that captivated me. It analyzes a book’s nuances in real-time, seamlessly transmuting static text into immersive audio or dynamic reading formats.
    • Cybersecurity Reimagined: SK Telecom introduced ScamVanguard, an AI-driven shield against mobile financial fraud. By synthesizing AI with advanced cybersecurity protocols, it identifies and neutralizes rapidly evolving global scams with impressive speed.
    • Embedded Intelligence: I was particularly intrigued by Suprema AI’s Q-Vision Pro. This on-device module leverages facial recognition and behavioral analytics to anticipate and thwart financial fraud at autonomous terminals like ATMs, flagging suspicious conduct before a crime even occurs.
    • Robotics & Human Augmentation: The Hypershell Carbon-X is a marvel of ergonomics. This all-terrain exoskeleton uses its M-One motor system to deliver 800W of assistive power. What impressed me most was the AI MotionEngine algorithm; it detects your gait and shifts between 10 different assistance modes in real-time, making strenuous physical exertion feel almost effortless.
    • Health and Human Security: Poskom’s AirRay-Mini is a masterclass in portable diagnostics. By integrating AI into a handheld X-ray system, it produces clinical-grade imagery with significantly reduced radiation doses—a vital step in minimizing patient exposure.
    • The Future of Visuals: Samsung didn’t disappoint with The Freestyle AI+. This portable GenAI-enabled projector is incredibly versatile. With features like AI 3D Keystone and Object Avoidance, it dynamically recalibrates the image to fit anything from curved walls to surfaces cluttered with plants or artwork.
    CES 2025: The FreestyleAI+ is a portable GenAI-enabled projector
    The FreestyleAI+ is a portable GenAI-enabled projector

    Honoring the Visionaries

    The Innovation Award Honorees showcased AI’s versatility across disparate sectors:

    • Creative Toolkits: Onoma AI debuted a formidable creative suite. It includes Fabulator (story ideation), Artifex (text-to-image), Anima (full-color illustration), and a collaborative Marketplace.
    • Synthetic Data: GenGenStudio by GenGenAI is solving a massive bottleneck by generating high-fidelity synthetic images and video for model training. Their current focus on the automotive sector—simulating rare “black swan” events like animal crossings or freak weather—is a game-changer for autonomous safety.
    GenGenStudio by GenGenAI
    GenGenStudio by GenGenAI
    • Experiential AI: L’Oreal’s Mood Mirror at CES 2025 takes AR further by incorporating Emotion AI. It doesn’t just show you how a product looks; it gauges your subconscious emotional reaction to the aesthetic.
    • Offline Ad-Tech: Triplet’s Deep Lounge AD is a sophisticated CMS that brings digital precision to physical retail. By using AI cameras to analyze foot traffic, dwell time, and behavior—like browsing or using fitting rooms—it serves hyper-personalized ads on digital displays in real-time.
    The camera measures the target audience's attention and delivers advertising effects comparable to those on online platforms
    The camera measures the target audience’s attention and delivers advertising effects comparable to those on online platforms

    The Big Tech Showdown: NVIDIA, Samsung, and LG

    NVIDIA: The “ChatGPT Moment” for Robotics Attending the NVIDIA keynote was a highlight. CEO Jensen Huang dropped a provocative prediction: the “ChatGPT moment” for robotics is arriving sooner than anticipated.

    • Project DIGITS: A powerhouse AI supercomputer in a form factor small enough for your home office.
    • Cosmos: Their new “world model” platform. It’s a physical-world simulator with Text2World and Video2World modes designed to train the next generation of robots.
    • GeForce RTX 50 Series: The RTX 5090 is a behemoth, boasting 4,000 AI TOPS and a staggering 1.8 TB/s bandwidth.
    A powerhouse AI supercomputer by NVIDIA: Project DIGITS
    A powerhouse AI supercomputer by NVIDIA: Project DIGITS

    Samsung: The Ambient Intelligence Home Samsung’s presence was dominated by “SmartThings” integration. I saw their MICRO LED Beauty Mirror, which provides a dermatological analysis in 30 seconds, and the AI Vision Inside 2.0 fridge that proactively manages your groceries. Their HoloDisplay Floating Screen was a crowd-favorite—a distortion-free 3D projection that looks like something out of a sci-fi film, yet functions as a practical hub for home monitoring.

    Samsung HoloDisplay
    Samsung HoloDisplay

    LG at CES 2025: Innovation for Every Family Member LG took a heartwarming yet high-tech turn with the Pet Care Zone, a smart shelter that monitors your pet’s vitals (heart rate, temperature, weight) and connects you to tele-vet services. On the display front, their LG83G5 Premium OLED remains the gold standard. The new α11 AI processor is reportedly four times faster than the α9, optimizing content with startling precision.

    LG83G5 Premium OLED
    LG83G5 Premium OLED

    This second look at CES 2025 confirms one thing: we have moved past the era of “AI as a gimmick.” We are now witnessing the era of AI as infrastructure.

    See also:

    While the ‘Vanguard’ represents the heavy hitters of the industry, the revolution is also happening at the micro-level. Be sure to check out my follow-up on the CES 2025 Hidden Gems for the oddities and specialized gadgets that stole the show.