When you let OTTO do your SEO, you’ll see improvements in website authority, content quality, technical performance, and user engagement across your websites. Leading to maximum organic traffic results for your brands of agency clients.
A first-of-its-kind AI SEO tool that will
revolutionize the way you do SEO.
Large language models (LLMs) reference external URLs to ground their answers. URL freshness measures how…
Large language models (LLMs) reference external URLs to ground their answers. URL freshness measures how recently a cited page was published relative to the moment the model produced the response. Freshness reveals whether the model draws from new material or from older content shaped by training data.
SEO and AI researchers debate how much web search access influences the age of cited sources. The missing piece is large-scale evidence that shows how search-enabled and search-disabled systems differ when selecting URLs.
This study examines 150,000 citations from OpenAI, Gemini, and Perplexity. The dataset contains 90,000 citations with search enabled and 60,000 citations with search disabled. Publication dates were extracted for 14,681 URLs to measure freshness with precision across both conditions.
The findings reveal clear patterns. Search-enabled models anchor their answers in recent webpages, often published within a few hundred days of the response. Search-disabled modes restrict the model to internal knowledge and cached references. These patterns show that retrieval access governs the freshness and structure of citations inside LLM responses.
This experiment measures how LLMs cite external webpages under different retrieval conditions. The analysis evaluates the age of the URLs that appear in LLM responses and examines how search access changes the freshness of those citations.
URL freshness was calculated by measuring the number of days between the publication date of a cited page and the timestamp of the LLM response that referenced it. This method shows how recently each source was published relative to the moment the model produced the answer.
The dataset integrates 2 primary components listed below.
Web Search Enabled Dataset. Contains 90,000 citations from OpenAI, Gemini, and Perplexity between October 28 and November 6, 2025. Each platform contributed 30,000 citations. Each record includes the cited URL, the response timestamp, and platform identifiers.
Web Search Disabled Dataset. Contains 60,000 citations from OpenAI and Gemini between October 1 and October 19, 2025. Each platform contributed 30,000 citations with search disabled. Perplexity does not offer a disabled-search mode, so it does not appear in this dataset.
These datasets provide 150,000 citations for freshness measurement across both conditions.
Publication dates were required for every evaluable URL. Homepage links were removed because they do not contain publication metadata. Each remaining URL was scanned for trusted fields (published_time, og:published_time, datePublished). Valid timestamps were recorded. This process produced 14,681 publication dates.
Freshness measures the time between the publication of a page and the LLM response that cited it. Freshness was calculated by subtracting the publication date from the response timestamp. The result reflects how recent the cited source was at the time of generation.
The analytical steps are listed below.
1. Review freshness distributions across all platforms.
2. Compare freshness between search-enabled and search-disabled responses.
3. Examine platform-level differences in citation age.
Only responses generated in 2025 were included to preserve temporal consistency. URLs published at any earlier time were eligible.
This framework creates a structured view of citation recency and shows how search access reshapes the age profile of URLs in LLM responses.
The analysis shows that web search access offers a direct impact on the freshness and structure of LLM citations. Enabled mode pulls recently published pages into answers, while disabled mode limits models to older material rooted in their training data. This contrast reveals how each system balances live information with internal knowledge.
Perplexity demonstrates the strongest retrieval behavior. Perplexity reliance on constant search produces a steady mix of fresh and mid-aged sources, which reflects continuous access to live webpages.
Gemini produces the freshest citations when search is enabled, but its performance changes sharply without retrieval. Disabled mode pushes Gemini toward homepage-level URLs and older content, which indicates heavy dependence on pretraining.
OpenAI shows the most resilience across both conditions. OpenAI continues to surface relatively recent, article-level URLs even without active search, which suggests broader internal coverage.
Freshness patterns remain consistent across conditions. Search-enabled systems cite pages published within a few hundred days of the response, while disabled systems shift toward older material.
These findings confirm that retrieval governs citation recency, specificity, and depth inside LLM responses. Search access increases freshness, while disabled modes reveal how each model defaults to internal knowledge. Freshness becomes a clear indicator of how LLMs source information and how their behavior changes once retrieval is present or removed.
I, Manick Bhan, together with the Search Atlas research team, analyzed 90,000 citations generated with web search enabled to measure how retrieval changes the structure and freshness of URLs in LLM responses.
The breakdown to show how OpenAI, Gemini, and Perplexity behave when search is enabled is listed below.
This analysis measures how often each model cites real article pages rather than top-level domains. Article pages matter because they contain publication metadata, which allows freshness to be calculated. Homepage URLs rarely expose timestamps, which limits freshness evaluation.
The headline results are shown below.
Gemini and OpenAI consistently cite specific content pages instead of general domain homepages, which enables stronger freshness analysis. Perplexity displays the most balanced split, which reflects a broader and more varied retrieval pattern.
This analysis measures how often the models cite the same domains for the same query. Domain overlap matters because it shows whether the systems retrieve similar sources when search is enabled or whether they diverge and construct distinct citation patterns.
The headline results are shown below.
All model pairs display low domain overlap, which confirms that they retrieve largely different sources even under identical retrieval conditions.
Extractable publication dates show how often each model surfaces article-level URLs that contain valid metadata. This determines how much freshness is evaluated for each platform.
This analysis measures how often each model returns article-level URLs with valid publication metadata. The headline results are shown below.
Perplexity is the most citation-dense model because it produces many article pages despite shorter responses. OpenAI retrieves broadly and returns many URLs with valid timestamps. Gemini generates longer outputs and fewer citations, which reduces the number of extractable dates even though it cites non-homepage URLs at a high rate.
The analysis requires equal sample sizes to compare freshness patterns fairly across platforms.
The headline result is shown below.
The smallest platform, Gemini, with 2,863 extractable timestamps, established the baseline. A random sample of 2,863 URLs was selected from OpenAI and Perplexity to match this count.
This normalization ensures fair cross-platform comparison. Each model contributes the same volume of publication-dated URLs, which prevents citation density differences from influencing freshness distributions.
This analysis measures how recently cited pages were published relative to the model response. The distribution covers the last 10 years of publication dates.
The headline results are shown below.
The distribution is strongly right-skewed. A large share of URLs have freshness values near zero, which shows that many pages were published only weeks or months before citation. The curve then tapers into a long tail, which confirms that while recent content dominates, older material still appears when relevant.
This analysis measures how recently each platform tends to cite published pages when search is enabled. The distribution spans the last 10 years of publication dates.
The headline results are shown below.
Gemini retrieves the freshest content overall. Perplexity maintains a balanced mix of recent and moderately old sources. OpenAI surfaces the broadest span of publication ages and the largest share of older URLs while still retrieving a meaningful volume of recent pages.
This analysis measures how well each cited URL matches the meaning of the user query. Strong semantic alignment means the model retrieves pages that directly answer the question. Weak alignment means the model retrieves pages that relate only loosely to the topic.
The headline results are shown below.
Gemini shows the strongest and most consistent match between the query and the cited page. OpenAI follows with stable and reliable alignment. Perplexity shows the widest spread, retrieving some highly relevant pages but many with a weaker connection to the query.
The average semantic similarity across all queries highlights how each model’s design shapes the quality of its citations. The headline results are shown below.
Gemini maintains the strongest overall alignment, which confirms that it retrieves content that closely reflects query intent. OpenAI performs well and stays consistent across queries. Perplexity shows the lowest averages, which indicates unpredictable relevance between the question and the cited content.
I, Manick Bhan, together with the Search Atlas research team, analyzed 60,000 citations generated with web search disabled to measure how the absence of retrieval changes URL structure and freshness. Perplexity does not offer a disabled-search mode, so the comparison focuses on OpenAI and Gemini.
This analysis measures how often each model cites real article pages when retrieval is unavailable. Article pages matter because they contain publication metadata. Homepage URLs do not provide this information.
The headline results are shown below.
Gemini moves almost entirely toward root domains in disabled mode, which shows heavy dependence on active retrieval. OpenAI maintains content-level citation behavior, which allows freshness to be evaluated even when search is unavailable.
This analysis measures how often each model references article-level URLs with valid publication metadata when retrieval is unavailable. Extractable timestamps matter because they determine how much freshness is measured in disabled-search conditions.
The headline results are shown below.
OpenAI continues to surface real content pages without retrieval, which generates thousands of URLs containing clean publication metadata.
Gemini shows a sharp drop in extractable timestamps. Only 5,710 of its 30,000 samples contain non-homepage URLs, and strict extraction rules identify valid dates for only 132 of them. This occurs because Gemini cites homepages in most responses, and those pages do not provide usable publication fields.
These patterns show that web search access increases the number of article-level URLs with verified timestamps. Gemini is the most constrained model when retrieval is disabled.
This distribution measures how recently pages were published relative to the response when search is disabled. The comparison shows whether each model continues to reference fresh material or defaults to older sources without retrieval.
The headline results are shown below.
OpenAI continues to reference relatively recent article-level content in disabled-search mode. Its citations cluster within the last 3 to 4 years, with many pages falling under 500 days.
Gemini shifts almost entirely toward long-aged sources without retrieval. Most cited pages are between 7 and 10 years old, which reflects heavy dependence on older information learned during pretraining.
This divergence aligns with homepage behavior.
OpenAI contributes 4,220 URLs with valid publication dates, while Gemini produces only 132 because most of its citations point to homepages. The absence of retrieval exposes clear differences in how each model sources information when fresh content is unavailable.
The 3 leading models (Perplexity, OpenAI, Gemini) were analyzed to determine which system cites the most recently published webpages. The comparison measures publication-date freshness, article-level URL behavior, and reliance on search access to show how each model sources information under search-enabled and search-disabled conditions.
The breakdown showing which model produces the freshest citations is outlined below.
Gemini produces the freshest citations when web search is enabled. Gemini responses concentrate near zero-day freshness, often citing pages published only weeks or months before the model generated the answer. This pattern confirms that Gemini leverages retrieval aggressively to surface recent, article-level content.
The Gemini freshness distribution drops sharply as publication age increases, which demonstrates a strong preference for new information. When search is enabled, Gemini behaves like a recency-optimized retrieval system designed to anchor answers in current material.
Perplexity cites a balanced mix of fresh and moderately aged URLs. Perplexity retrieves a large volume of article-level pages and maintains a strong recency signal, though not as concentrated as Gemini.
The Perplexity distribution extends into older material more frequently, which reflects its broad retrieval style. The pattern shows that Perplexity behaves like a high-coverage search model that blends fresh sources with deeper archival pages while maintaining consistent access to current information.
OpenAI demonstrates strong resilience in freshness behavior. With search enabled, OpenAI retrieves many recently published pages, though with a wider age range than Gemini or Perplexity.
When search is disabled, OpenAI continues to cite moderately recent articles at a high rate. It produces thousands of extractable publication dates, which confirms that GPT retrieves structured, content-rich URLs even without live-web assistance.
OpenAI pattern reveals a model anchored in article-level content regardless of retrieval mode, which maintains meaningful recency through internal knowledge and link-based reasoning.
SEO and AI teams need to treat URL freshness as a core visibility signal. Fresh citations inside LLM responses show whether a brand’s content remains present and relevant in AI-generated answers.
Teams need to optimize content for query-level relevance. Pages with clear topical focus, verified publication dates, and strong semantic alignment are cited more frequently across search-enabled systems. Structured metadata and recent updates increase the likelihood that models classify a page as current and contextually correct.
Tracking freshness across platforms reveals how often newer pages enter the citation stream and how they compete against older sources.
These patterns influence which brands appear in AI answers and how their content is positioned.
Strategic planning must account for these differences. Align content updates, metadata structure, and publication velocity with platform-level freshness trends to secure visibility across emerging AI discovery channels.
Brands that publish consistently, maintain clean article metadata, and reinforce topical clarity achieve the strongest presence in search-enabled LLM environments.
Every dataset and retrieval condition introduces constraints. The limitations of this study are listed below.
Despite these limits, the analysis establishes a clear baseline for understanding how retrieval access affects citation freshness. The findings reveal consistent behavioral differences across models and provide a foundation for broader longitudinal research on LLM citation dynamics.
Manick Bhan is a 3x INC 5000 Founder CEO/CTO of Search Atlas which is an AI SEO automation platform used by thousands of brands and agencies and awarded Best SEO Platform by the Global Search Awards, Shortlisted by Capterra, Front Runners by Software Advice, Category Leaders by GetApp, and best tool for customer satisfaction and usability by Gartner.
Manick Bhan founded LinkGraph, a digital marketing firm that helps enterprise brands and agencies scale through data-driven SEO with clients like Shutterfly and Samsung. LinkGraph is listed as one of the Fastest Growing Private Companies in the US by inc.5000, as one of the Best Workplaces in Advertising & Marketing by Fortune, as New York’s B2B Leaders by Clutch, won no.1 Spot in Nevada’s Top Workplaces, Best B2B SEO Campaign by The Drum Awards for Search, and named Best Start-Up Agency at U.S. Search Awards.
Manick Bhan is the owner for Signal Genesys, the leading platform for automated press release distribution and digital presence management, and LinkLaboratory, the largest online publisher catalog in the world.
With 10+ years of experience in SEO from the in-house and agency side, Manick Bhan has taught both startups and Fortune 500 companies how to scale their brands with a data-driven SEO strategy that can break into any market and outrank even the biggest of competitors. Bhan’s innovative approach to SEO has helped Search Atlas and LinkGraph scale to multiple 8 figures.
Manick's thought leadership has appeared in leading publications like Forbes, Search Engine Journal (SEJ), VentureBeat, G2, Digital Summit, Wordstream, Wix SEO Hub, Wordable, Inc. Masters, AllBusiness, SEO Blog, Jumpstory, Serpstat, Outbrain, Improvado, Unstack, Clickbank, Built in, Martechseries, Smartbrief, Marketingprofs, Readwrite, Honeybook, Content Marketing Institute, LocalIQ, CXL, Oncrawl, Venture Beat, Addicted2Success, Search Engine Watch, Business 2 Community, Digital Connect MAG, and VegasInc.
Manick Bhan is a speaker at events like TechCrunch Disrupt, Traffic & Conversion Summit, Ad World, HighLevel Summit, Chiang Mai SEO, Merchant Mastery, SEO Week, AI Bot Summit, SEO Spring Training, LeadSnap Mansion Mastermind, SEOROCKSTARS, LeadSnapEvents, DigiMarCon, brightonSEO, Affiliate Summit West, Traffic and Conversion Summit, Outranking Summit, TES Affiliate Conference, billo Summit, ContentTECH Summit, Content Marketing Conference, VEGPRENEUR Expert Hour, Ai4 Conference, SMX West, and Affiliate Summit West.
Manick Bhan is the Founder CEO/CTO of the SEOTheory community, a community designed for agency owners looking to increase their SEO results.
Manick Bhan enjoys writing and speaking on topics that range from digital marketing to artificial intelligence and machine learning to social impact in the animal welfare and environmental space.
Manick lives in Medellin, Colombia with his wife Sophia Deluz-Bhan, daughter Ruby, and a house full of animals including Voodoo the SEO cat.Identifying top competitors and their visibility…
Ask, create, search…
Dr. David McInnis Orthodontics struggled with low search visibility and inconsistent patient inquiries. Despite offering premium orthodontic services, their online presence failed to generate steady leads.
By implementing Search Atlas’s advanced SEO strategy, we restructured their website for search intent alignment, optimized local SEO, and enhanced technical performance to dominate Google rankings.
Now, Dr. David McInnis Orthodontics enjoys a steady stream of organic leads and a powerful online presence, making them the go-to orthodontic practice in their area.
Their mission is to provide clients with all the tools necessary to tackle addiction at its source. To do this, they needed to significantly increase their online presence and support their crucial mission.
The client utilized Search Atlas to identify and resolve technical flaws, including broken links, slow loading times, and navigation issues. With OTTO, they performed these fixes and optimizations in one day.
In Austin’s bustling legal market, standing out as a DUI law firm is challenging due to intense competition. Achieving local search visibility requires an innovative strategic SEO approach.
To improve search rankings for their keywords, we incorporated these terms into the website and Google Business Profile (GBP) over 4 weeks using OTTO. After OTTO implementation, 100% of the pins are ranking either in top 3 or top 5 local search positions.
OTTO’s automated SEO optimization process simplifies SEO efforts, reducing manual labor and allowing the team to focus on other crucial tasks.
This center is dedicated to providing essential resources and programs for children with special needs and their families. Despite their valuable mission, the center’s website traffic had stalled for months, preventing them from connecting with potential clients.
To drive more traffic to their site, the client implemented OTTO’s recommendations. This included enhancing content quality, optimizing technical aspects of the site, refining on-page SEO elements, and building authority through the publication of 2 press releases.
The results were astounding. The client transitioned from being relatively obscure online to becoming a go-to resource in local search results for families seeking support.
If Any of These Sound Familiar, It’s Time for an Enterprise SEO Solution:
