YearGuessr Benchmark: Building-Age Estimation
- YearGuessr is an open benchmark featuring 55,546 building façade images with detailed metadata for precise building-age estimation across 157 countries.
- It employs strict ordinal regression using CORAL-style threshold and FCRC ranking losses to predict construction years from combined visual and geographic inputs.
- The benchmark exposes popularity bias by revealing performance gaps between models on frequently recognized landmarks versus more ordinary, long-tail structures.
YearGuessr is the first large-scale, open benchmark explicitly designed to assess building-age estimation and expose “popularity bias” in contemporary vision–LLMs (VLMs). By providing a multi-modal dataset and rigorous ordinal regression protocols, YearGuessr enables systematic evaluation of models’ generalization capabilities beyond the memorization of famous architectural landmarks (Szu-Tu et al., 24 Dec 2025).
1. Dataset Composition
YearGuessr consists of 55,546 unique building façade images sourced from Wikipedia/Wikimedia Commons (CC BY-SA 4.0), spanning 157 countries. Each sample is annotated with:
- Image : a pixel façade crop.
- GPS coordinates , with 100% coverage.
- Wikipedia page-view count (summed 01 Jul 2023–01 Jul 2024), serving as a quantification of "popularity."
- Complete textual description (median length: 2,240 characters) and country via reverse geocoding.
- Construction year label .
Geographically, the dataset is 63.3% from the Americas, 22.5% Europe, 6.3% Asia, and the remainder from Africa/Oceania. Temporally, labels are continuous between 1001 and 2024 CE, with a long-tailed log-scale distribution (notably, >10% predate 1600 CE; major peaks in the 18th–20th centuries).
2. Ordinal Regression Task Formulation
YearGuessr frames construction year prediction as strict ordinal regression. Each model processes and outputs a scalar year estimate . Evaluation uses:
- Mean Absolute Error (MAE):
- Interval Accuracy (IA), for years:
Two complementary training paradigms are used:
(a) CORAL/CORN-style Threshold Loss:
For each , real-valued scores are predicted for binary predicates (year ), with a sigmoid cross-entropy across thresholds:
The final prediction is
(b) FCRC Ranking-based Contrastive Loss (from NumCLIP):
This loss penalizes out-of-order image/text embeddings , weighting negatives with :
3. Popularity-Aware Evaluation Protocols
To quantify memorization effects tied to landmark popularity, YearGuessr stratifies test samples by into popularity bins:
- "Ordinary" ()
- "Popular" ()
- Finer: –––
For any subset (e.g., "Popular"), interval accuracy is
The "popularity gain" metric is defined as
A continuous variant weights by , but the principal focus is on interpretable bin-wise splits. This protocol exposes models’ tendency to perform disproportionately well on highly viewed, frequently pre-trained landmarks.
4. Benchmarked Models and YearCLIP Design
Evaluation covers 43 methods across several model families:
| Model Family | Notable Examples | Distinguishing Properties |
|---|---|---|
| CNNs | ResNet-50/152, ConvNeXt-baseline/large | Pure vision; no language prior |
| Pure Transformers | ViT-B/16, Swin-B | Token-mixing, no explicit language modeling |
| CLIP-based | Zero-shot CLIP, GeoCLIP, NumCLIP, YearCLIP | Joint vision-language, varying geo/numeric features |
| Closed-source VLMs | GPT-4-Vision-mini, Gemini, Claude 3, Grok 2 | Large, often proprietary, pre-trained on web-scale |
| Open-source VLMs | CogVLM2, Gemma3, GLM-4V-9B, LLaVA variants | Community-accessible, hybrid language/vision |
YearCLIP extends CLIP by using frozen image and text encoders, with additional architectural innovations including:
- Location Conditioning: Random Fourier Features (RFF) with MLP and a learnable "zero convolution" fuse geographic and visual information.
- Coarse-to-Fine Style Classification: Seven historical style tokens (e.g., Roman, Gothic, Contemporary) provide architectural cues.
- Reasoning Prompts: Approximately 20 architectural sub-tokens (roof, wall, height) enrich predictions and enable post-hoc rationalization.
- Trainable Regressor: Ingests similarity scores from style and reasoning prompts for ordinal-regressed year output.
Training incorporates FCRC loss for ordering, cross-entropy for style classes, and an optional regression penalty.
5. Empirical Findings on Popularity Bias
Significant empirical findings are as follows:
- Standard Accuracy: YearCLIP attains MAE ≈ 39.5 years, IA₅ ≈ 19%, IA₁₀₀ ≈ 91.6%, a 13.5% MAE improvement over GeoCLIP (45.7 years). Gemini 1.5-Pro (MAE 33.1, IA₅ 28.2%), Gemini 2.0-Flash, and Grok 2 are the strongest closed models.
- Popularity Gap: Pure vision models generalize better on “ordinary” buildings but exhibit negative :
- ConvNeXt-B: IA₅ drops from 16.6 (Ordinary) to 12.7 (Popular),
- Swin-B: 15.8→6.8,
- Closed-source VLMs display strong memorization with large positive :
- Gemini 2.0-Flash: IA₅ jumps from 24.2 to 58.4,
- Gemini 1.5-Pro: , Grok 2:
- Open-source LLM/VLM hybrids yield positive but smaller .
- Regional and Temporal Patterns: All models are most accurate in the Americas (lowest MAE), less so in Africa/Europe; performance increases with proximity to the 19th–20th centuries, with MAE >400 years in the 1000–1150 CE interval.
- The results collectively demonstrate that current VLMs often rely on memorized entries from pre-training, particularly for high-popularity buildings, rather than genuinely learning to interpret architectural signals; vision-only models avoid this but at the cost of lower peak accuracy.
6. Real-World and Methodological Implications
YearGuessr surfaces an over-reliance on landmark memorization in building age estimation, undermining claims of deep architectural understanding in VLMs. In applied contexts—such as heritage preservation, sustainability audits, and disaster assessment—this bias risks systematic neglect of data from marginalized regions and rare styles, with attendant fairness concerns.
A plausible implication is that benchmarking must integrate popularity-aware stratification to avoid overestimating VLM performance due to web-crawled data’s inherent visibility skew. Furthermore, pure vision models’ relatively better generalization to the "long tail" suggests ensemble or hybrid approaches may mitigate the identified bias, albeit with trade-offs in peak accuracy.
7. Future Directions and Recommendations
Suggested avenues for further research and dataset enrichment include:
- Expansion with additional non-Western and early-period samples (e.g., using CMAB or bespoke low-resource collection).
- Annotation of renovation/rebuilding events to distinguish original from reconstructed features.
- Incorporation of fairness-aware or adversarial loss functions to penalize shortcut learning tied to popularity.
- Utilization of synthetic data (e.g., via diffusion priors) and active learning to balance class representation.
- Development of debiased prompting or retrieval strategies to reduce memorization.
The YearGuessr benchmark (Szu-Tu et al., 24 Dec 2025) thus establishes a multi-modal, ordinal regression standard, holding future VLMs to higher standards of generalization and providing a framework to address memorization-driven artifacts in model evaluation.