In our previous analysis, Effect of Nofollow on LLM Training, we established a grim reality for the privacy-conscious webmaster: AI training bots do not respect the rel="nofollow" attribute.
For two decades, nofollow was the gentlemen’s agreement of the web. It was a digital “Do Not Enter” sign that search engines like Google and Bing respected to manage authority flow (PageRank) and combat spam. It was a protocol built for an era of retrieval, where the primary value of a link was the endorsement it carried. If you didn’t want to endorse a site, you added the tag, and the “juice” stopped flowing.
Read more →It is a common confusion in our industry: “GEO” often refers to “Generative Engine Optimization.” But for the scientific community, GEO means Geology. And interestingly, geological data provides one of the best case studies for how to ground Large Language Models in physical reality.
The Hallucination of Physical Space
Ask an ungrounded LLM “What is the soil composition of the specific plot at [Lat, Long]?” and it will likely hallucinate a generic answer based on the region. “It’s probably clay.” It averages the data.
Read more →“Near me” queries are changing. In the past, Google used your IP address to find businesses within a 5-mile radius. In the future, agents will use Inferred Intent and Capability Matching.
Agents don’t just look for proximity; they look for capability. “Find me a plumber who can fix a tankless heater today” is a query a standard search engine struggles with. But an agent will call the plumber or check their real-time booking API.
Read more →In traditional SEO, hreflang tags were the holy grail of internationalization. They told Google: “This page is for French speakers in Canada.” But in a world where AI models are inherently polyglot, does this tag still matter?
The Polyglot LLM
Models like GPT-4 and Gemini are trained on multilingual datasets. They can seamlessly translate between English, Japanese, and Swahili. If a user asks a question in Spanish, the model can retrieve an English source, translate the facts, and generate a Spanish answer.
Read more →Geological features are named entities. “Mount Everest” is an entity. “The San Andreas Fault” is an entity. “The Pierre Shale Formation” is an entity.
For researchers in the geospatial domain, linking your content to these distinct entities is the bedrock of MCP-SEO.
Disambiguation via Wikidata
“Paris” is a city in France. “Paris” is also a city in Texas. “Paris” is also a rock formation (hypothetically).
To ensure an AI understands you are talking about the rock formation, you must link to its Wikidata ID (e.g., Q12345).
Read more →Cross-lingual retrieval is the frontier of international SEO. With vector embeddings, the barrier of language is dissolving. A query in Spanish can match a document in English if the semantic vector is similar. This fundamental shift challenges everything we know about global site architecture.
How Vector Spaces Bridge Languages
In a high-dimensional vector space (like that created by text-embedding-ada-002 or cohere-multilingual), the concept of “Dog” (English), “Perro” (Spanish), and “Inu” (Japanese) cluster in the same geometric region. They are semantically identical, even if lexically distinct.
Read more →