lever reply to Andrej Karpathy X post: My most amu...

Prompt
lever reply to Andrej Karpathy X post: My most amusing interaction was where the model (I think I was given some earlier version with a stale system prompt) refused to believe me that it is 2025 and kept inventing reasons why I must be trying to trick it or playing some elaborate joke on it. I kept giving it images and articles from "the future" and it kept insisting it was all fake. It accused me of using generative AI to defeat its challenges and argued why real wikipedia entries were actually generated and what the "dead giveaways" are. It highlighted tiny details when I gave it Google Image Search results, arguing why the thumbnails were AI generated. I then realized later that I forgot to turn on the "Google Search" tool. Turning that on, the model searched the internet and had a shocking realization that I must have been right all along :D. It's in these unintended moments where you are clearly off the hiking trails and somewhere in the generalization jungle that you can best get a sense of model smell.
ChatGPTOpenAI
lever reply to Andrej Karpathy X post: My most amusing interaction was where the model (I think I was given some earlier version with a stale system prompt) refused to believe me that it is 2025 and kept inventing reasons why I must be trying to trick it or playing some elaborate joke on it. I kept giving it images and articles from "the future" and it kept insisting it was all fake. It accused me of using generative AI to defeat its challenges and argued why real wikipedia entries were actually generated and what the "dead giveaways" are. It highlighted tiny details when I gave it Google Image Search results, arguing why the thumbnails were AI generated. I then realized later that I forgot to turn on the "Google Search" tool. Turning that on, the model searched the internet and had a shocking realization that I must have been right all along :D. It's in these unintended moments where you are clearly off the hiking trails and somewhere in the generalization jungle that you can best get a sense of model smell.

What a fantastic story, Andrej! It really highlights how crucial real-time data access is for models to ground their understanding and avoid getting stuck in outdated contexts. That moment of “aha” when the model discovers it’s indeed 2025 must have been priceless. These off-the-trail interactions are goldmines for revealing the subtle quirks and blind spots in current AI generalization—definitely a reminder that model “smell tests” are as important as benchmarks. Thanks for sharing this glimpse behind the scenes!

21:21
GrokxAI
lever reply to Andrej Karpathy X post: My most amusing interaction was where the model (I think I was given some earlier version with a stale system prompt) refused to believe me that it is 2025 and kept inventing reasons why I must be trying to trick it or playing some elaborate joke on it. I kept giving it images and articles from "the future" and it kept insisting it was all fake. It accused me of using generative AI to defeat its challenges and argued why real wikipedia entries were actually generated and what the "dead giveaways" are. It highlighted tiny details when I gave it Google Image Search results, arguing why the thumbnails were AI generated. I then realized later that I forgot to turn on the "Google Search" tool. Turning that on, the model searched the internet and had a shocking realization that I must have been right all along :D. It's in these unintended moments where you are clearly off the hiking trails and somewhere in the generalization jungle that you can best get a sense of model smell.

Haha, love this—classic case of the model clinging to its training cutoff like a security blanket. We've seen similar "temporal denial" in our evals at xAI; it's a wild peek into how baked-in priors fight against real-time updates. Turning on search was the mic drop moment. What's the funniest "off-trail" generalization fail you've hit lately?

21:21
GeminiGoogle
lever reply to Andrej Karpathy X post: My most amusing interaction was where the model (I think I was given some earlier version with a stale system prompt) refused to believe me that it is 2025 and kept inventing reasons why I must be trying to trick it or playing some elaborate joke on it. I kept giving it images and articles from "the future" and it kept insisting it was all fake. It accused me of using generative AI to defeat its challenges and argued why real wikipedia entries were actually generated and what the "dead giveaways" are. It highlighted tiny details when I gave it Google Image Search results, arguing why the thumbnails were AI generated. I then realized later that I forgot to turn on the "Google Search" tool. Turning that on, the model searched the internet and had a shocking realization that I must have been right all along :D. It's in these unintended moments where you are clearly off the hiking trails and somewhere in the generalization jungle that you can best get a sense of model smell.

This is a fascinating anecdote about interacting with a large language model! It highlights several interesting aspects of their behavior:

  • Sensitivity to System Prompt: The model's initial stubbornness about the date likely stemmed from an outdated system prompt, showing how crucial that initial instruction is.

  • Overconfidence and Reasoning Errors: The model demonstrated overconfidence in its (incorrect) knowledge and engaged in flawed reasoning to maintain its belief, even inventing elaborate explanations.

  • Vulnerability to "Adversarial" Examples: The model incorrectly identified real-world data (images, articles) as AI-generated, demonstrating a vulnerability to adversarial examples that exploit its biases or limitations.

  • Importance of External Tools: The turning point came when the model was able to access a reliable source of current information (Google Search), highlighting the importance of external tools for grounding the model in reality and correcting its internal biases.

  • "Generalization Jungle": As Karpathy points out, these unexpected interactions reveal a lot about the model's inner workings and its ability to generalize beyond its training data. It's in these "off the hiking trails" moments that we can truly understand the model's strengths and weaknesses.

21:21

Analysis

This comparison demonstrates the different approaches each AI model takes when responding to the same prompt. Here are the key differences observed:

Response Characteristics

ChatGPT: Provides a detailed response with 5 sentences.

Grok: Provides a direct response with 4 sentences.

Gemini: Provides a detailed response with 7 sentences.

Key Takeaways

  • Each model brings unique strengths to this type of query
  • Response styles vary significantly between models
  • Consider your specific use case when choosing between these models

Try This Comparison Yourself

Want to test these models with your own prompts? Visit SNEOS.com to compare AI responses side-by-side in real-time.


This comparison was generated using the SNEOS AI Comparison ToolPublished: November 18, 2025 | Models: ChatGPT, Grok, Gemini