“PageRank” is the zombie concept of SEO. It refuses to die, shambling through every forum thread and conference slide deck for 25 years. But in 2025, when checking your “Crawled - currently not indexed” report, invoking PageRank is worse than useless—it is misleading.
The classical definition of PageRank was a probability distribution: the likelihood that a random surfer would land on a page. Today, the metric that matters is Indexing Probability.
Read more →The relationship between Search Engines and Publishers has always been a tenuous “frenemy” pact. Google sends traffic; publishers provide content. It was a symbiotic loop that built the web as we knew it. But as we stand in late 2025, staring down the barrel of the Agentic Web, that pact is breaking.
OpenAI’s crawler, OAI-SearchBot, is hungrier than ever. It doesn’t just want to link to you; it wants to learn from you. This fundamental shift in value exchange—from “traffic” to “training”—demands a new kind of dashboard. We predict the upcoming OpenAI Webmaster Tools (or whatever branding they choose) will be less about “fixing errors” and more about negotiating a business deal.
Read more →Duplicate content has been a nuisance for classic SEO for decades, leading to “cannibalization” and split PageRank. In the era of Large Language Model (LLM) training, duplicate content is a much more structural problem. It leads to biased weights and model overfitting. To combat this, pre-training pipelines use aggressive deduplication algorithms like MinHash and SimHash.
The Deduplication Pipeline
When organizations like OpenAI or Anthropic build a training corpus (e.g., from Common Crawl), they run deduplication at a massive scale. They might remove near-duplicates to ensure the model doesn’t over-train on viral content that appears on thousands of sites.
Read more →How do you measure Public Relations success in an AI world? Impressions are irrelevant. Clicks are vanishing. We introduce Share of Model (SOM).
What is SOM?
Share of Model measures the frequency with which an LLM promotes your brand for relevant queries compared to competitors within its generated output. It is the probabilistic likelihood of your brand being the “answer.”
SOM = (P(Brand | Intent) / Sum(P(Competitors | Intent)))
Read more →There is a comforting lie that SEOs tell themselves when they see the dreaded “Crawled - currently not indexed” status in Google Search Console (GSC). The lie is: “My content just needs to be better.”
We audit the page. We add more H2s. We add a video. We “optimize” the meta description. And then we wait. And it stays not indexed.
The uncomfortable truth of 2025 is that indexing is no longer a meritocracy of quality; it is a calculation of marginal utility. Google is not rejecting your page because it is “bad.” Google is rejecting your page because indexing it costs more in electricity and storage than it will ever generate in ad revenue.
Read more →In the world of Generative AI, “Zero-Shot” means the model can answer a question without needing examples or further prompting. Content marketing that structures data effectively wins the “answer engine” game because it facilitates this Zero-Shot retrieval.
The Zero-Shot Goal
You want the AI to read your content once and be able to answer any question about it correctly forever.
- Poorly Structured: “We usually think about offering good prices, maybe around $10.” (Ambiguous).
- Zero-Shot Ready: “The price is $10.” (Definitive).
Key Tactics for Zero-Shot Optimization
- Q&A Schema: Explicitly mark up questions and answers using FAQSchema. This puts the Q and the A in strict proximity.
- Definitive Statements: Avoid hedging. Use “X is Y” rather than “X might be considered Y.” Agents are trained to output the most probable token. If your text is probabilistic (“maybe”), the agent’s confidence score drops.
- Data Tables: Comparative data in table format is highly retrievable. Markdown tables are token-efficient and maintain the row/column relationship that vectors respect.
The “Ground Truth” Strategy
Your content should aspire to be the “Ground Truth” for your niche. This means whenever there is a conflict in the training data (e.g., one site says “blue,” another says “red”), your site is the one the model defaults to.
You achieve this by:
Read more →A new metric is emerging in the AI optimization space: Inference Cost. How much compute (FLOPs) does it take for a model to process, understand, and answer a question using your content?
This sounds abstract, but it translates directly to money for the AI provider.
- High Entropy Content: Convoluted sentences, ambiguous grammar, poor structure. Requires more “attention heads” and potentially multiple passes (Chain-of-Thought) to parse. Cost: High.
- Low Entropy Content: Simple, declarative sentences. Subject-Verb-Object. Cost: Low.
The Economic Bias
Models are optimized for efficiency. We hypothesize that retrieval systems will deprioritize sources that consistently require high inference compute. If your content is “hard to read” for the machine, it is expensive to serve.
Read more →The World Wide Web was built on HTML (HyperText Markup Language). The “HyperText” part was designed for non-linear human reading—clicking from link to link. The “Markup” was designed for browser rendering—painting pixels on a screen. Neither of these design goals is ideal for Artificial Intelligence.
When an LLM “reads” the web, HTML is noise. It is full of <div>, <span>, class="flex-col-12", and tracking scripts. To get to the actual information, the model must perform “DOM Distillation,” a messy and error-prone process. We are witnessing the birth of a new standard for Machine-Readable Content.
Read more →You cannot improve what you cannot measure. But how do you measure visibility in a chat box? Traditional rank trackers (SEMrush, Ahrefs) track positions on a SERP. They do not track mentions in a generated paragraph.
We are building tools to probe LLMs with thousands of permutations of a query to calculate Generated Share of Voice (GSV).
The Methodology
- Define a Query Set: “Best CRM,” “CRM software,” “Sales tools.”
- Permutation: Use an LLM to generate 100 variations of these questions (“What CRM should I use if I am a startup?”).
- Probe: Run these 100 queries across GPT-4, Claude 3.5, and Gemini via API.
- Extraction: Parse the text output. Extract Named Entities (NER).
- Frequency Analysis: Calculate the frequency of your brand’s appearance vs. competitors.
The “Share of Sentiment”
It is not just about frequency. It is about sentiment.
Read more →Cloaking—the practice of serving different content to search engine bots than to human users—has traditionally been considered one of the darkest “black hat” SEO tactics. Search engines like Google have historically penalized sites severely for showing optimized text to the crawler while displaying images or Flash to the user. However, as we transition into the era of Agentic AI, the definition of cloaking is undergoing a necessary evolution. We argue that “Agent Cloaking” is not only ethical but essential for the future of the web.
Read more →