Thomas H is a Sheffield Class Humber Keel Barge.

She was built in 1940 by Richard Dunston Ltd., Thorne, Yorkshire.

She was one of two sister ships commissioned by the Hodgsons Tannery at Beverley Beck on the Humber, where she worked for many years. Her sister ship was called Richard after the other Hodgson bother.

She was never under sail, at the time she was built the government was subsidising the building of motor driven barges.

She is extra wide beam at 15.5 feet and she is 62.5 feet long.

We bought her in early 2006 through Alan Pease in Goole and roped him into emptying the various tanks and debris she had in her at the time, decking over her open hold, replacing the unusable Lister engine and generally get her onto working order for the trip down from Goole around the coast to the Thames. Then, we got him to pilot her down too.

This is a belated attempt to diary the ups and downs of our journey so far.

Thursday, 26 February 2026

RAG in SEO Explained: The Engine Behind Google's AI Overviews

Retrieval-Augmented Generation (RAG) is the specific framework that allows Large Language Models (LLMs) to fetch external data before writing an answer. In my SEO consulting work, I define it as the bridge between a static AI model and a dynamic search index. This technology powers Google's AI Overviews and stops the model from hallucinating by grounding it in real facts. Unlike standard keyword-based crawling, retrieval in this context specifically refers to neural vector retrieval, which matches the semantic meaning of a query to a database of facts rather than simply matching text strings.

The process works by replacing simple keyword matching with Vector Search. When a user asks a complex question, the system does not just look for matching words. It scans a Vector Database to find conceptually related text chunks. The Retriever acts like a research assistant that pulls specific paragraphs from trusted sites and feeds them into the Generator. This means your content must be structured as clear facts that an AI can easily digest and cite. If your site contradicts the consensus found in the Knowledge Graph, the RAG system will likely ignore you.

Google uses this to create synthesized answers that often result in Zero-Click Searches. Consequently, you must optimize for entity salience and clear Subject-Predicate-Object syntax. This shift has birthed Generative Engine Optimization (GEO). My data shows that pages using valid Schema Markup are significantly more likely to be retrieved as grounding sources. You must treat your website less like a brochure and more like a structured database.

On the production side, smart SEOs use RAG to build Programmatic SEO workflows. We connect an LLM to a private database of brand facts, allowing us to generate thousands of accurate, compliant landing pages at scale without the risk of AI making things up. We are shifting from a search economy to an answer economy. To survive this shift, you must audit your data structure today. If your content is hard for a machine to parse, you will lose visibility in the AI-driven future. More on - https://www.linkedin.com/pulse/what-rag-seo-bridge-between-large-language-models-search-nicor-fdimc/

--
You received this message because you are subscribed to the Google Groups "Broadcaster" group.
To unsubscribe from this group and stop receiving emails from it, send an email to broadcaster-news+unsubscribe@googlegroups.com.
To view this discussion visit https://groups.google.com/d/msgid/broadcaster-news/a9249b8a-013a-4a96-beeb-53e7e6ba6984n%40googlegroups.com.

No comments: