Researchers found that multimodal AI models cannot tell time. The more variation there was in the clock face, the more the chatbot being tested was likely to misreadResearchers found that multimodal AI models cannot tell time. The more variation there was in the clock face, the more the chatbot being tested was likely to misread

Before AI Takes Our Jobs, Someone Better Teach It How to Tell Time

Am I the only one who didn’t know that AI cannot figure out time? I mean, every day, we hear all about generative AI “revolutionizing” everything and replacing everyone. Pretty genius little things. So imagine my shock when I learned that multimodal AI models cannot tell time. How did I know, you ask?

To start with, researchers at the University of Edinburgh recently found that multimodal large language models (MLLMs) like ChatGPT-4o, GPT-o1, Gemini-2.0, and Claude 3.5-Sonnet ran into accuracy problems while reading a clock face.

Things got worse when they were tested with clocks designed with Roman numerals, a colored dial, or a decorative hour hand. Some of the clocks also had a hand that tracked seconds in addition to minutes and hours. In the face of those design touches, the AI models reportedly fell into further errors.

This discovery was made during a test of a lineup of top MLLMs today, and to think that Gemini-2.0 performed the “best” with only 22.8% accuracy sounds hilarious. GPT-4.o and GPT-o1’s exact match accuracy stood at 8.6% and 4.84% respectively.

Per the researchers, these models struggled with everything. Which hand is the hour hand? Which direction is it pointing? What angle corresponds to what time? What number is that? According to them, the more variation there was in the clock face, the more the chatbot being tested was likely to misread the clock.

These are literally basic skills for people. Most six or seven-year-olds can already tell time. But for these models, it might as well be the most complicated astrophysics.

After the clock fiasco, the researchers tested the bots on yearly calendars. You know, the ones with all twelve months on one page. GPT-o1 performed the “best” here, reaching 80 percent accuracy. But that still means that one out of every five answers was wrong, including simple questions like “Which day of the week is New Year’s Day? If my child failed to get that right on a quiz, I would honestly be very worried.

I never would have thought that AI models could ever get confused by a common calendar layout. But then, it is not very shocking to find out. It all still boils down to a long-standing gap in AI development. MLLMs only recognize patterns they have already seen, and clocks, calendars, or anything that requires spatial reasoning don’t fit into that.

Humans can look at a warped Dali clock and still figure out roughly what time it is meant to display. But AI models see a slightly thicker hour hand and kind of short-circuit.

Why This Matters

It is easy (almost satisfying) to laugh at ChatGPT, Gemini, and these models for failing a task you learned when you were little. A task you do with so much ease. As someone who has gotten jilted by clients for the free work these things offer, albeit substandard, I admit I do find it really satisfying.

But as much as I want to just laugh it off, there is a more serious angle to this. These same MLLMs are being pushed into autonomous driving perception, medical imaging, robotics, and accessibility tools. They are being used for scheduling and automation as well as real-time decision-making systems.

Now, clock-reading errors are funny. But medical errors? Navigation errors? Even scheduling errors? Not so funny.

If a model cannot reliably read a clock, trusting it blindly in high-stakes environments is too risky a gamble for me. It just shows how far these systems still are from actual, grounded intelligence. And how much human common sense and nuance still matter. I am trying so hard to steer clear of taking this chance to make a human vs. AI case. I sure won’t use it to preach “Why I Hate AI and You Should Too.” But there is a problem that needs to be looked into.

As the study’s lead author, Rohit Saxena, put it, these weaknesses “must be addressed if AI systems are to be successfully integrated into time-sensitive real-world applications.”

Market Opportunity
Sleepless AI Logo
Sleepless AI Price(AI)
$0.03727
$0.03727$0.03727
-2.63%
USD
Sleepless AI (AI) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

What We Know (and Don’t) About Modern Code Reviews

What We Know (and Don’t) About Modern Code Reviews

This article traces the evolution of modern code review from formal inspections to tool-driven workflows, maps key research themes, and highlights a critical gap
Share
Hackernoon2025/12/17 17:00
X claims the right to share your private AI chats with everyone under new rules – no opt out

X claims the right to share your private AI chats with everyone under new rules – no opt out

X says its Terms of Service will change Jan. 15, 2026, expanding how the platform defines user “Content” and adding contract language tied to the operation and
Share
CryptoSlate2025/12/17 19:24
Michael Saylor Pushes Digital Capital Narrative At Bitcoin Treasuries Unconference

Michael Saylor Pushes Digital Capital Narrative At Bitcoin Treasuries Unconference

The post Michael Saylor Pushes Digital Capital Narrative At Bitcoin Treasuries Unconference appeared on BitcoinEthereumNews.com. The suitcoiners are in town.  From a low-key, circular podium in the middle of a lavish New York City event hall, Strategy executive chairman Michael Saylor took the mic and opened the Bitcoin Treasuries Unconference event. He joked awkwardly about the orange ties, dresses, caps and other merch to the (mostly male) audience of who’s-who in the bitcoin treasury company world.  Once he got onto the regular beat, it was much of the same: calm and relaxed, speaking freely and with confidence, his keynote was heavy on the metaphors and larger historical stories. Treasury companies are like Rockefeller’s Standard Oil in its early years, Michael Saylor said: We’ve just discovered crude oil and now we’re making sense of the myriad ways in which we can use it — the automobile revolution and jet fuel is still well ahead of us.  Established, trillion-dollar companies not using AI because of “security concerns” make them slow and stupid — just like companies and individuals rejecting digital assets now make them poor and weak.  “I’d like to think that we understood our business five years ago; we didn’t.”  We went from a defensive investment into bitcoin, Saylor said, to opportunistic, to strategic, and finally transformational; “only then did we realize that we were different.” Michael Saylor: You Come Into My Financial History House?! Jokes aside, Michael Saylor is very welcome to the warm waters of our financial past. He acquitted himself honorably by invoking the British Consol — though mispronouncing it, and misdating it to the 1780s; Pelham’s consolidation of debts happened in the 1750s and perpetual government debt existed well before then — and comparing it to the gold standard and the future of bitcoin. He’s right that Strategy’s STRC product in many ways imitates the consols; irredeemable, perpetual debt, issued at par, with…
Share
BitcoinEthereumNews2025/09/18 02:12