Show HN: Gemini Pro 3 imagines the HN front page 10 years from now
TL;DR Highlight
An experiment feeding Gemini Pro 3 today's HN front page and asking it to predict what HN looks like in 2035 — exposing the limits of AI future prediction.
Who Should Read
AI researchers interested in LLM reasoning limits, and product thinkers who use AI for forecasting or trend analysis.
Core Mechanics
- The experiment gave Gemini Pro 3 the current HN front page as context and asked for a prediction of the HN front page 10 years out (2035).
- The model's predictions revealed a pattern: extrapolating current trends linearly rather than reasoning about discontinuities, surprises, or second-order effects.
- The AI predicted more AI, more AGI discussion, more quantum computing — essentially amplified versions of what's already trending, without predicting emergent surprises.
- This exposes a fundamental limitation: LLMs are trained on what happened, not on what was surprising about what happened. They tend to produce 'confident-sounding trend extrapolation' not genuine forecasting.
- The meta-lesson is that AI models make poor forecasters for discontinuous events but reasonable performers on incremental trend extension.
Evidence
- The actual model outputs were shared in the post and showed heavy clustering around AI/ML, quantum, and biotech topics with little imagination for entirely new categories.
- HN commenters pointed out that actual HN front pages from 10 years ago would have looked very different from what anyone in 2015 would have predicted.
- Several forecasting enthusiasts cited Superforecasting literature — the point that calibrated uncertainty, not confident prediction, is the mark of good forecasting. LLMs tend to be overconfident.
- Some commenters argued the experiment was unfair — no human can reliably predict 10-year tech trends either. The interesting question is whether AI is worse than a calibrated human expert.
How to Apply
- When using LLMs for trend analysis or forecasting, treat their outputs as 'extrapolation hypotheses' to be stress-tested, not predictions to be trusted.
- Ask the model explicitly to generate surprising or contrarian scenarios — this partially counteracts the tendency to extrapolate trends.
- For strategic planning, use LLMs to enumerate known trends and then bring human judgment (or dedicated forecasting tools like Metaculus) for discontinuity assessment.
- Frame LLM forecasting prompts as 'what could make this trend reverse?' rather than 'where will this go?' to get more useful adversarial scenarios.
Terminology
Related Papers
Using Claude Code: The unreasonable effectiveness of HTML
Claude Code 팀이 Markdown 대신 HTML을 LLM 출력 포맷으로 선호하기 시작한 이유와 그 실용적 장점을 정리한 글로, AI와 함께 문서/스펙/대시보드를 만드는 워크플로우에 직접적인 영향을 준다.
When to Vote, When to Rewrite: Disagreement-Guided Strategy Routing for Test-Time Scaling
Disagreement-guided routing boosts LLM accuracy on math and code by 3-7% with adaptive problem solving.
Less Is More: Engineering Challenges of On-Device Small Language Model Integration in a Mobile Application
Five failure modes and eight practical solutions emerged after five days of running on-device SLMs (Gemma 4 E2B, Qwen3 0.6B) with Wordle.
Dynamic Context Evolution for Scalable Synthetic Data Generation
A framework that completely eliminates duplication and repetition in large-scale synthetic data generation with LLMs using three mechanisms (VTS + Semantic Memory + Adaptive Prompt).
90%+ fewer tokens per session by reading a pre-compiled wiki instead of exploring files cold. Built from Karpathy's workflow.
This is a workflow sharing post about how pre-organizing a codebase in Wiki format can reduce token usage per Claude session by more than 90% instead of directly exploring the codebase every time.