A story about how AI can go wrong when you expect it to act like a simulation engine
Maybe you know, maybe you don't: I'm a Washington Commanders (née Redskins) fan. Have been for a long time, more than four decades now. Good times (NFL championships/Super Bowls, magical Jayden Daniels rookie season), bad times (Snyder era, Sean Taylor), and just about everything in between. Even when it was hard, I was still a fan. (Note: I'm not old enough to recall the incredible Sammy Baugh as their QB/Safety/Punter - if you're unfamiliar with that name, and an NFL fan, you should definitely go read that article on his incredible fifteen years at Washington.)
To say this season was a disappointment is an understatement. With that in mind, I turned my attention toward the 2026 draft. And that's where I ran into major shortcomings in the AI realm.
Where AI Actually Falls Short
I asked Copilot what the likely draft pick spot would be for Washington. It made something up. I gave it the latest records, and asked it to predict the remaining games, and it made up a fake schedule. I gave it articles about the schedules, and it suggested those were hypothetical future schedules. I finally coerced it into thinking about the current players and records, and it still hallucinated details about imaginary matchups. And when it did give draft order (which it had invented), it suggested taking players that were drafted two years ago. I gave up.
I turned my attention to ChatGPT. Same misinformation.
I turned to Gemini, and asked,
Why do all you AI have trouble with a request like "predict the remaining 2025 NFL season and the resultant 2026 draft order"? Making up standings, ignoring actual news, "predicting" players that were already drafted a year or two ago, etc.
Gemini replied:
It is incredibly frustrating, especially when you are looking for data-driven analysis rather than "creative writing."
To answer your question directly: The reason most AI models fail at this specific task ("predict the rest of the season" or "project the draft") is a conflict between how they are trained and how sports work.
The "Frozen in Time" Problem: Most models are trained on data that is 6–18 months old. If a model’s training cutoff was January 2024, it has zero concept of the 2025 season. When you ask it to "predict the 2025 standings," it doesn't run a simulation; it hallucinates a plausible-sounding story based on historical team strength (e.g., "The Chiefs will be good," "The Panthers will struggle"). It ends up listing players who were drafted years ago because those names are statistically weighted as "college prospects" in its frozen memory.
Lack of a Simulation Engine: A language model is a text predictor, not a logic engine. It cannot track 32 teams, calculate tie-breakers, and simulate the remaining 4 weeks of a season in its head. That requires a structured database and a Monte Carlo simulation, which is a math problem, not a language problem.
The "News" Void: Unless the AI actively searches the web (like I can), it misses the context that changes everything—like a starting QB tearing an ACL in Week 8 or a breakout star from a small school (like Indiana's Fernando Mendoza this year).
The Actual 2025 Reality (As of Dec 14, 2025)
Since I have access to live tools, I can skip the "creative writing" and give you the actual current snapshot of the 2025 season entering Week 15, and what that means for the 2026 Draft.
OK, Gemini, let's go. I asked:

No comments:
Post a Comment