April 2026GPT-5.3 · Source Shift

ChatGPT 5.3 Halved Its Hotel SourcesWhat changed on March 5, 2026

TL;DR: Across 9,722 hotel responses from 4 country locales, the day ChatGPT's UI switched to GPT-5.3 is a cliff. URLs per answer fell 49% (24 → 12), unique domains per answer fell 46% (18 → 10), and the pool of domains per prompt fell 54% (132 → 61). Answer coverage collapsed for Wikipedia (31% → 4%), Booking (44% → 20%), Expedia (56% → 33%) and Reddit (14% → 2%). A Stockholm-operated network of ~17 Booking-affiliate listicle sites gained the freed-up citation slots.

NS
Nicolas Sitter
Published April 17, 2026
140
World prompts
24 → 12
URLs / answer
132 → 61
Domain pool / prompt
18 → 10
Domains / answer
Read the Report

Executive Summary

A model change, not a geography rollout.

Since late December 2025 we've run the same 140 world-wide hotel discovery prompts on ChatGPT's UI every day, from 4 country locales (US, GB, DE, ES). On 2026-03-05 OpenAI quietly switched the UI from GPT-5.2 to GPT-5.3 (GPT-5.4 in the API). The same prompts, unchanged, now produce a very different answer: half as many URLs, half as many domains, and a radically different source mix.

The drop is uniform across all 4 locales (−47% to −53%), so this is a model/pipeline change — not a rollout that hit one country first. The breadth of the source pool per prompt shrank from 132 distinct domains to 61, and the source mix tilted away from UGC and chain brands toward a handful of programmatic- SEO hotel directories.

−49%
URLs per answer
24 → 12
−46%
Unique domains
18 → 10
−54%
Domain pool per prompt
132 → 61
−27pp
Wikipedia answer coverage
31% → 4%
Finding 1

The halving

On March 5, 2026 — the day OpenAI rolled GPT-5.3 out to the ChatGPT UI — the same 140 prompts started returning roughly half the sources overnight. Every headline metric moves together on that date, which is why we treat the UI model swap as the cause.

Source metrics — before vs after the Mar 5, 2026 cutover
MetricPre (GPT-5.2)Post (GPT-5.3)Δ
Captures6,4143,308
URLs per capture (mean)23.9912.21−49.1%
Unique domains per capture (mean)18.4310.03−45.6%
URLs per unique domain1.311.24−5%
Unique domains pooled per prompt13261−54%
The ratio URLs ÷ domains barely moved (1.31 → 1.24). ChatGPT didn't cite fewer pages per site — it cited fewer sites. The breadth of the source pool was halved.
Finding 2

Uniform across locales

If the drop were a geographic A/B rollout, we'd expect country deltas to differ. They don't.

URLs per answer — pre vs post cutover by country
By country: pre vs post URLs/answer and domains/answer
CountryCaptures (pre / post)URLs/answerDomains/answerΔ URLs
Germany1,249 / 73924.67 → 12.1218.75 → 9.82-50.9%
Spain1,232 / 94426.24 → 12.4419.25 → 10.03-52.6%
United Kingdom631 / 64123.11 → 12.1017.99 → 10.06-47.6%
United States3,302 / 98423.07 → 12.1318.09 → 10.17-47.4%
The drop lands everywhere at once — a model/pipeline shift, not a regional rollout. Country-level deltas range from −47% (US, GB) to −53% (ES), a narrow band given the sample-size differences.
Finding 3

The source mix tilted

Every category shrank in absolute terms, but some shrank far faster than others — and the tiny seo_directory bucket held flat, nearly tripling its share of URLs.

Category coverage — % of answers where category appears at least once
Category coverage — % of captures (answers) with ≥1 URL in that category, counted once per answer
CategoryPre %Post %Δ pp
OTA83.2%60.4%-22.8pp
Meta81.9%65.8%-16.1pp
Independent99.9%99.0%-0.9pp
Editorial47.4%35.9%-11.5pp
Chain34.2%11.5%-22.7pp
UGC21.1%2.4%-18.7pp
SEO Directory20.9%22.8%1.9pp
Google2.3%0.1%-2.2pp
UGC coverage (Reddit, Instagram, YouTube, Facebook) collapsed from 21% of answers to 2%. Chain-brand coverage fell by two-thirds (34% → 11%). Google Maps/Search citations effectively vanished (2.3% → 0.1%). Meanwhile, seo_directory was the only category whose coverage rose (21% → 23%) — in a market where every other source shrank.
Finding 4

Winners & losers

Ranked by change in answer coverage — the share of ChatGPT answers in which a brand appeared at least once (counted once per answer, no double-counting for repeated mentions).

Why two delta columns? Every answer now carries roughly half as many URLs(24 → 12), so every brand would lose coverage even if nothing else changed.

  • Δ % — relative change: (post − pre) ÷ pre. Wikipedia 31% → 4% reads as −88%.
  • Δ vs halving %(post − pre × 0.509) ÷ (pre × 0.509), where 0.509 is the URLs/answer ratio (12.21 / 23.99). This separates “squeezed by the halving” from “actively deprioritised”. Green = gained share net of the halving, red = lost share on top of it. E.g. Wikipedia went from 31% to 4% — a halving would have predicted 16%, so the relative shortfall is −76%.
Top losing brands — % of answers citing the brand, pre vs post cutover
Δ vs halving — who genuinely gained (green) or lost (red) once you factor out the halving

Chart filtered to brands with ≥2% pre-coverage. Emerging SEO directories (pre near zero) are shown in the dedicated SEO panel below where the % blow-ups reflect real 10–30× growth.

Top losers by Δ vs halving (actual post vs the halving baseline pre × 0.509)
BrandCategoryPre %Post %Expected post %Δ %Δ vs halving %
wikipedia.orgIndependent30.9%3.7%15.7%-88.0%-76%
reddit.comUGC14.2%2.1%7.2%-85.2%-71%
michelin.comEditorial14.4%4.5%7.3%-68.8%-39%
booking.comOTA44.3%20.3%22.6%-54.2%-10%
marriott.comChain10.0%2.9%5.1%-71.0%-43%
trivago.comMeta17.0%6.6%8.7%-61.2%-24%
Top gainers by Δ vs halving (brands that held more shelf than the halving predicted)
BrandCategoryPre %Post %Expected post %Δ %Δ vs halving %
tripadvisor.comMeta43.8%33.5%22.3%-23.5%+50%
theluxuryeditor.comEditorial17.1%15.1%8.7%-11.7%+73%
oyster.comIndependent21.5%16.1%11.0%-25.1%+47%
expedia.comOTA56.2%33.3%28.6%-40.7%+16%
travelmyth.comMeta12.5%10.3%6.4%-17.6%+62%
trip.comOTA16.5%11.9%8.4%-27.9%+42%
Adjusting for the halving flips the story for the biggest OTA/meta names. theluxuryeditor.com held +74% above halving, TripAdvisor +50%, Oyster +46%, Travelmyth +61%, Trip.com +42% and Expedia +16% — all genuinely won shelf. Even Booking.com (−10% vs halving) is close to neutral. The genuine losers are UGC and editorial: Wikipedia −76% below halvingis the clearest deprioritisation, Reddit −71%, Four Seasons −68%, Agoda −69% and Hilton −54% all collapsed far below a uniform halving. Every luxury chain brand landed in the −40% to −68% range. In the long tail, the yonder network (yonder-society.com + yonder.fr, combined pre 10.1% → post 6.1%) held +20% above halving. The other above-halving gainers are programmatic-SEO listicle sites — the subject of Finding 5.

Δ vs halving, broken down by source type

Same metric expressed as a relative percentage: (post − pre × 0.509) ÷ (pre × 0.509). Within each type the story is self-contained: OTAs and metas have a clear internal ranking, editorial is mixed, chains are universally down, UGC is a bloodbath, and the SEO directories have no losers at all.

OTAs

Booking, Expedia, Hotels.com, Agoda, Trip.com and peers

Expedia (+16%) and Trip.com (+42%) genuinely won shelf. Agoda (−69%) and Booking.com (−10%) landed below halving.

Metasearch & curated directories

TripAdvisor, Kayak, Trivago, The Hotel Guru, Travelmyth

TripAdvisor (+50%) is the single biggest winner of the whole cutover (in absolute pp). Kayak, Trivago and Mr & Mrs Smith all lost share on top of halving.

Editorial & publishers

Condé Nast, Michelin, Forbes, Time Out, The Luxury Editor

Editorial is mixed: theluxuryeditor.com (+74%) and Time Out (+46%) gained, while Forbes Travel Guide (−71%) and Guardian (−27%) were demoted.

Hotel chains

Marriott, Hilton, Four Seasons, Ritz-Carlton, IHG and peers

Every single chain brand is red. LHW −83%, Ritz-Carlton −69%, IHG −64%, Four Seasons −68%, Hilton −54%, Marriott −43%. GPT-5.3 appears to actively deprioritise brand .com sites in favour of meta/editorial.

User-generated (UGC)

Reddit, Facebook, Instagram, YouTube

Universal collapse. Reddit −71% below halving, and Facebook, Instagram, YouTube all lost 90-95% of their expected shelf (effectively zero post-cutover).

SEO directories (programmatic listicles)

luxuryhotel.guide, all-boutique-hotels, hotels-with-*, etc.

No losers. Every brand in this bucket gained share. Percentages are massive because most of these sites had near-zero pre-coverage — all-boutique-hotels.com +3,000%, luxuryhotel.guide +1,170%, boutiquehotel.guru +780% — i.e. 10-30× growth vs expected.

Finding 5

The Stockholm listicle network syndrome

The biggest surprise is who took ChatGPT's freed-up citation slots. Not Michelin, not Condé Nast, not Marriott — a cluster of thin-content, single-operator Booking-affiliate directories.

17
Confirmed network domains
14 primary + 3 secondary
1
Shared image host
images.luxuryhotel.guru
1
Shared registrar
Gandi SAS
2
Cloudflare NS pairs
ANDY/RITA, BRENDA/GRAHAM

Sample sites

All share a hero search backed by Booking.com's affiliate widget, a templated destination grid, a Scandinavian curator persona (Ted Valentin, Maja Holm, Elain Olsson, David Bachmann), and byte-identical footer boilerplate.

all-boutique-hotels.com
hotels-with-balcony.com
small-luxury-hotels.net
adults-only-hotels.net
spahotelbreak.com
hotels-with-pool.com
hotels-with-private-pool.com
hotels-with-tennis.com
hotels-with-sauna.com
hotels-with-great-views.com
hotels-with-rooftop.com
award-winning-hotels.com
amazing-hotels.com
romantichotels.com
luxuryhotel.guide
luxuryhotel.guru
boutiquehotel.guru
poolhotels.guide
5starhotels.guide
beachhotels.guide
couples-hotels.com
small-hotels-guide.com
romantic-getaway.me

Infrastructure fingerprint

WHOIS / DNS evidence
FingerprintValue
RegistrarGandi SAS
Cloudflare NS (primary cluster)ANDY.NS.CLOUDFLARE.COM + RITA.NS.CLOUDFLARE.COM
Cloudflare NS (secondary cluster)BRENDA.NS.CLOUDFLARE.COM + GRAHAM.NS.CLOUDFLARE.COM
Image host (shared across sites)images.luxuryhotel.guru
Bulk .guide registrations5starhotels / poolhotels / beachhotels / luxuryhotel — all 2014-05-08
About-page slug/about-us/ identical across inspected members

The operator

Ted Valentin, named as curator on all-boutique-hotels.com, is a Stockholm-based entrepreneur publicly known for directory sites (hitta.se among others). The other curator names — Elain Olsson, Maja Holm, David Bachmann — do not correspond to verifiable individuals and appear to be personas reused across the templated sites.

The absolute shelf these sites hold is still small: post-cutover, the biggest of them (luxuryhotel.guide) appears in 2.6% of answers. It's a spike, not a takeover — several of these domains had near-zero pre-cutover presence, so the percentage gains are large off a small base. What's notable is the direction: every single brand in this bucket gained, and the only category whose coverage rose while everything else shrank is this one. For a model update that otherwise tilted toward established meta/editorial names (TripAdvisor, Oyster, Expedia), the appearance of a Stockholm- operated affiliate network — not editorial authority — worth tracking.
Finding 6

Luxury queries hit hardest

We classified each prompt into an intent tier by keyword. Luxury and high-amenity queries — which previously leaned on OTAs, chains and editorial — took the biggest hit.

Δ URLs per answer by query tier
By query tier
TierPromptsPre URLsPost URLsΔ
Rooftop / Pool526.3211.30-57.1%
Luxury (5-star)3027.7412.64-54.5%
Beachfront727.9812.91-53.9%
Family1423.3411.90-49.0%
Boutique / Design2823.0211.89-48.3%
Other3023.5012.42-47.2%
Affordable1219.8311.69-41.0%
Romantic1420.3412.15-40.3%
Finding 7

City hotspots

St Barts and the Alpine ski markets were hit hardest — both previously pulled 30+ URLs per answer because the destinations are OTA-thin and required more sources to triangulate. Post-cutover every city lands in the same ~12-URL band.

URLs per answer — pre vs post cutover by city
By city
CityPre URLsPost URLsΔ
St Barts33.5912.28-63.5%
Courchevel26.6712.16-54.4%
Istanbul25.6712.02-53.2%
Barcelona25.2712.02-52.4%
Amsterdam23.7811.47-51.8%
Los Angeles23.6811.42-51.8%
Rome23.4611.67-50.2%
Saint-Tropez25.8113.27-48.6%
Berlin23.6512.28-48.1%
San Francisco22.6111.84-47.7%
Megeve23.8212.45-47.7%
Dubai23.3012.45-46.5%
New York23.4512.64-46.1%
Las Vegas23.2413.05-43.8%
Timeline

Weekly view of the cliff

The drop isn't drift — it's a step change. Weekly means sit in a ~23-URL band for all of January and February, then collapse to ~12 in the week of March 9 and stay there.

Weekly mean URLs and domains per ChatGPT hotel answer

Weekly means weighted by capture count across all 4 country locales. Cutover week is 2026-03-02. The tough week of 2026-03-23 (mean 8.95) is rollout noise — the following weeks return to a stable ~12-URL equilibrium.

Why this happened

Two plausible drivers

The data is what it is. The reason ChatGPT now consults half as many hotel sources isn't in the data — OpenAI didn't publish a rationale for the 5.3 rollout. But two hypotheses fit the shape of what we see.

Hypothesis 1

Cutting serving cost

Every extra URL pulled into context is a web-fetch request, extra tokens in the context window, and more reasoning time. Going from 24 URLs / answer to 12 roughly halves the web-search + context- assembly cost for every hotel query. ChatGPT handles millions of travel queries a day, so a straightforward 50% cost cut on a whole vertical is a meaningful margin lever.

Consistent with this: the halving is uniform across all 4 country locales and all query tiers. That looks more like a product/config knob being turned than an organic quality improvement.

Hypothesis 2

Reverting a GPT-5.2 spike

It's possible 5.2 was the anomaly, not 5.3. GPT-5.2 may have shipped with an unusually high retrieval budget (24 URLs / answer is quite a lot of citations for a 300-word answer), and 5.3 simply pulls that back toward a pre-5.2 baseline.

We don't have pre-5.2 data at the same resolution, so this is speculation — but if it's right, 5.3's 12 URLs isn't a cut, it's a return to normal. Implication for marketers: don't over-index on the pre-March 2026 citation counts; they may have been the outlier.

Both hypotheses predict the same data, and they aren't mutually exclusive — 5.3 could be both cheaper to serve and closer to a long-run equilibrium. What they have in common is that neither implies a quality reason for the cut. If OpenAI thought 24 URLs gave materially better answers, they'd have kept them.
Methodology

How we ran the study

Data collection

  • Same 140 world hotel discovery prompts, run daily from 4 country locales (US, GB, DE, ES)
  • Capture the ChatGPT UI (not the API) — including the full sources panel per response
  • Per URL: domain, position, cited=true/false (was it linked inside the answer text?)
  • Coverage window 2025-12-25 → 2026-04-17 (87 distinct capture days, 9,722 responses)
  • Residual gpt-5-2 rows after Mar 5 are dropped as rollout noise

Processing

  • Brand rollup. Localised TLDs (tripadvisor.es, expedia.de, hoteles.com, fr.wikipedia.org, maps.google.com) are canonicalised to a single brand.
  • Taxonomy. Every domain is bucketed into OTA, meta, chain, editorial, UGC, Google, independent, or the new seo_directory bucket for programmatic-SEO hotel listicles.
  • Two views. citation_count = every URL in the sources panel. cited=true = URLs also linked inline in the answer text.
  • Listicle forensics. WHOIS + Cloudflare nameserver lookups + image-host and template inspection, not just scrape signals.

Example prompts

The 140 prompts span 14 destinations × 10 intent tiers (luxury, beachfront, rooftop, family, romantic, affordable, design/boutique, ski, other). A representative sample:

“best luxury hotels in Dubai”
“5 star hotels in Rome with rooftop terrace”
“romantic hotels in Saint-Tropez for couples”
“family friendly hotels near Brandenburg Gate”
“boutique hotels near Place des Lices Saint-Tropez”
“best beachfront hotels in Los Angeles”
“design alpine hotels in Courchevel village”
“affordable hotels in Amsterdam with good reviews”
“5 star hotels in Las Vegas with spa”
“luxury hotels in Barcelona with pool”
“family friendly hotels near Pampelonne beach”
“affordable ski hotels in Megeve”

Full list of 140 prompts with tier + city classification: /data/chatgpt-hotel-source-shift-2026/prompts.csv. Aggregate headline stats: summary.csv.

Caveats

  • Post-cutover sample is lighter (3,308 vs 6,414) because of the shorter post window. Country-level deltas are consistent, which mitigates this.
  • The ChatGPT UI may have product-level changes stacked on the model change. We treat them as a single “March 5 intervention” because they shipped simultaneously.
  • WebFetch 403'd on several listicle homepages; WHOIS and DNS evidence confirmed the network irrespective of that.
  • 140 prompts is a finite set chosen for breadth of destinations and intent tiers, not statistical exhaustiveness.
FAQ

Frequently asked questions

Citations per answer dropped 49%. Before Mar 5, 2026 each hotel answer consulted 24 URLs from 18 distinct domains on average. After Mar 5 the same prompts returned just 12 URLs from 10 domains. The pool of unique domains across all 140 world prompts also halved: 132 → 61. The drop is consistent across all 4 country locales we tested (US −47%, GB −48%, DE −51%, ES −53%).

Summarize with AI

ChatGPTPerplexityClaudeGeminiGrok