Flagship Skill ยท Programmatic SEO

The programmatic SEO skill.

Quality at scale or not at all.

A senior SEO strategist's playbook for designing and running programmatic SEO programs that produce durable traffic instead of penalty-bait. Data source identification, template design, schema patterns, quality control at scale, internal linking architecture, crawl budget management, AEO/GEO for programmatic pages, refresh discipline, and the make-or-break question of whether pSEO is the right answer for the program at all.

Audience: SEO content strategists, content engineers, agencies running pSEO programs, in-house teams considering pSEO as a growth lever.

What this skill is for

The scaled-scope skill in a five-skill content workflow.

Five skills compose into the content discipline. Each has its own scope; the distinction is what keeps each skill sharp. This skill is the scaled scope: 100s to 100,000s of pages generated programmatically from structured data sources, each targeting a long-tail query.

  1. 01

    content-strategyPROGRAM scope

    Decides whether pSEO is a fit for the program at all.

  2. 02

    pillar-content-architectureHUB scope

    Editorial alternative: one topic with 10-15 intentional editorial pieces.

  3. 03

    content-brief-authoringPER-PIECE scope

    Briefs each editorial piece individually.

  4. 04

    content-and-copyEXECUTION scope

    Writes individual editorial pieces.

  5. 05

    programmatic-seoSCALED scope (this skill)

    100s to 100,000s of pages generated programmatically from structured data, each targeting a long-tail query.

The keystone distinction

Three data depths. Most pSEO programs ship the broken one.

The data source is the program. Field count per record is the simplest depth signal. Thin schemas produce thin pages no matter how clever the template; deep schemas produce pages that earn rankings and AI citations even under modest templates.

Penalty risk

Thin (5 fields)

5 fields per record. Pages render at 200 to 400 words of templated boilerplate. Algorithm updates detect the scale-without-substance pattern and penalize within 12 to 18 months. The bad-reputation pSEO that gives the technique its complicated name.

Workable

Competent (15 to 20 fields)

15 to 20 fields per record. Pages render at 800 to 1,500 words with substantive sections. Survives algorithm updates if the data is real and QC is budgeted. The minimum bar for a pSEO program that earns ongoing investment.

The discipline

Deep (30+ fields)

30+ fields per record, with computed fields and cross-record fields powering depth and linking. Pages render at 1,500+ words with distinctive sections per record. Compounds for years; defensible against competitors. The Zillow / Glassdoor / TripAdvisor end of the spectrum.

Field count is not the only signal that matters (queryable intent, defensible data source, refresh cadence, QC budget all factor in), but it is the fastest tell. A schema with 5 fields cannot produce substantive pages even with a perfect template; a schema with 30 fields produces substantive pages even with a modest template.

pSEO architecture

How the link graph compounds.

A hub at the center (parent category like "Homes in Denver"), child pages around it (specific neighborhoods or records), with hub-to-child links flowing top-down and sibling links connecting children to children. The sibling links are bidirectional and run between adjacent neighbors plus selected non-adjacent siblings (similar attributes, similar price ranges, similar features). A 10-page set is shown for legibility; real programs run 100s to 100,000s of records following the same architecture.

pSEO architecture: hub at center, child pages around it, sibling links between childrenA parent hub category at the center of the diagram is connected by solid lines to ten child page records arranged around it. Lighter dashed lines connect each child to its neighbors and to selected non-adjacent siblings, showing the bidirectional sibling-linking graph that distributes ranking signal across the set.ChildCapitol HillChildHighlandsChildRiNoChildCherry CreekChildWash ParkChildStapletonChildLoDoChildFive PointsChildPark HillChildSloan's LakeHubHomes in DenverParent categoryHub-to-child link (solid)Sibling link (dashed, bidirectional)
The link graph distributes ranking signal across the set. Hub-to-child links flow PageRank top-down; sibling links circulate it laterally; child-to-hub links (bottom-up, not shown for visual clarity) reinforce the hub. Without the sibling layer, child pages remain orphans no matter how many hub-to-child links exist.

The framework

Twelve considerations for programmatic SEO.

When designing or auditing a pSEO program, walk these 12 considerations. Skipping any of them produces one of the failure modes documented in the skill.

  1. 01Right answer for the program (5 criteria)
  2. 02Data source as moat
  3. 03Schema depth: 15+ minimum, 30+ for competitive
  4. 04Template variable density across sparse and dense records
  5. 05Above-the-fold answer in first 200 words
  6. 06Schema markup at scale on every page
  7. 07Internal linking: hub-and-spoke + sibling, no orphans
  8. 08Quality control sampling: 50 to 200 pages per cycle
  9. 09Failure thresholds: 5% halts new generation
  10. 10Crawl budget: segmented sitemaps, noindex thin pages
  11. 11Refresh cadence: quarterly minimum, daily for volatile
  12. 12Pruning lifecycle at 12 to 24 months

What is in the skill

Thirteen sections covered in the body.

The SKILL.md spans the full pSEO lifecycle from the keystone make-or-break decision through data sources, templates, schema design, quality control, internal linking, crawl budget, AEO/GEO patterns, refresh discipline, and the failure-mode catalog.

  1. 01

    What this skill is for

    Composition with five sister and adjacent skills. Methodology, not implementation: the skill teaches the shape of pSEO programs without prescribing a specific stack.

  2. 02

    When pSEO is the right answer

    The keystone section. Five criteria all required: real underlying data, long-tail volume, queryable intent, refresh-cadence alignment, QC budget. Default is no.

  3. 03

    Data source identification

    Five patterns: first-party, licensed, public-aggregated, expert-curated, synthesized. Defensibility analysis per pattern. The 'moat' question: would a competitor replicate this within 6 months?

  4. 04

    Template design

    Above-the-fold answer in first 200 words. Variable density across sparse and dense records. Heading hierarchy reflects data. Schema markup as part of template. Internal linking placeholders. Distinctive value per page.

  5. 05

    Schema design

    Field count signals depth (15+ minimum, 30+ for competitive depth). Required vs optional fields with graceful degradation. Computed fields. Cross-record fields powering sibling links. Update frequency tags. Schema-as-product principle.

  6. 06

    Quality control at scale

    Stratified sampling 50 to 200 pages per cycle. Automated checks (heading, schema, links, duplicates). Manual review checklist. 5% failure threshold halts new generation. Cohort tracking surfaces drift.

  7. 07

    Internal linking across the set

    Hub-and-spoke architecture. Sibling linking (5 to 15 per page) bidirectional. Bottom-up links from spoke to hub. Anchor text variation. Crawl-friendly architecture with no orphans.

  8. 08

    Crawl budget management

    Sitemap segmentation by category and recency. Noindex on thin pages below required-field threshold. Canonical handling for pivot duplicates (X vs Y inverse). Crawl rate monitoring. 24-month pruning checkpoint.

  9. 09

    AEO and GEO for programmatic pages

    Direct-answer extraction in first 200 words. Comprehensive structured data signals authority. AI engines cite programmatic pages for factual queries; editorial for analytical. Quality crackdown sensitivity is higher in AI surfaces than traditional search.

  10. 10

    Refresh and maintenance at scale

    Volatile-field refresh daily to weekly; slow-changing quarterly to annually. Cohort-by-cohort template version migration with rollback. Pruning lifecycle at 24 months. Set-level refreshes are 6-month projects.

  11. 11

    Common failure modes

    Eleven-plus patterns: penalized at scale, orphan child pages, crawl rate plateau, identical-looking pages, template versioning broken, drift in QC ownership, AI engines do not cite, engagement decay, replicable data, refresh overwhelm, traffic without conversions.

  12. 12

    The framework: 12 considerations

    Right answer for the program, data source as moat, schema depth, template variable density, above-the-fold answer, schema markup at scale, internal linking architecture, QC sampling, failure thresholds, crawl budget discipline, refresh cadence, pruning lifecycle.

  13. 13

    Quality at scale or not at all

    The only durable pSEO programs are the ones that hold quality at scale. The discipline is not optional and not free. Budget for it before generating the first page, or do not start at all.

Reference files

Ten references that go alongside the SKILL.md.

The references hold the decision framework, the data source patterns, the template and schema design principles, the QC discipline, the internal linking architecture, the crawl budget management, the AEO/GEO patterns, the refresh cadence, and the failure-mode catalog.

  • references/when-pseo-works-decision.md

    Five-criterion decision framework expanded with worked yes / no / maybe examples across real estate listings, SaaS comparisons, AI-generated city guides, and enterprise B2B pricing pages. The walk-away rule when two or more criteria fail.

  • references/data-source-identification-patterns.md

    Five data-source patterns (first-party, licensed, public-aggregated, expert-curated, synthesized) with defensibility analysis. The moat-replication test. Multi-pattern combinations and when they earn their complexity.

  • references/template-design-patterns.md

    Above-the-fold answer pattern, variable density handling, heading hierarchy, schema integration, internal linking placeholder slots, distinctive-value-per-page test. Methodology principles vs stack-specific implementation choices that stay internal.

  • references/schema-design-patterns.md

    Field count thresholds (5 thin, 15-20 competent, 30+ deep). Required vs optional fields. Computed fields, cross-record fields, update frequency tags. Schema-as-product principle (versioning, documentation, review, testing). When the schema is wrong.

  • references/quality-control-at-scale.md

    Stratified sampling strategy by sparse/dense, recent/old, popular/niche, cohort version. Automated check catalog. Manual review checklist. 5% failure threshold discipline. Cohort tracking dimensions. Team budget rule of thumb.

  • references/internal-linking-at-scale.md

    Hub-and-spoke architecture. Sibling linking (5 to 15 per page) bidirectional. Bottom-up cluster-to-pillar discipline. Anchor text variation patterns. Crawl-friendly architecture. The PageRank flow principle. Linking inventory.

  • references/crawl-budget-management.md

    Sitemap segmentation by category and recency. Noindex implementation patterns. Canonical handling for pivot duplicates. Crawl rate monitoring signals. Pruning options (noindex, 410, redirect). The 6/12/24 month checkpoints.

  • references/aeo-geo-for-programmatic-pages.md

    Direct-answer extraction. Structured data signals. Citation patterns (factual lane vs analytical lane). Quality crackdown sensitivity in AI surfaces. Two-engine optimization framing. Set-level reputation signals. Hub-level AEO/GEO metrics worth tracking.

  • references/refresh-at-scale.md

    Data refresh cadence by field volatility. Template version migration patterns (cohort-based, generation-period, data-density). Migration monitoring and rollback. Pruning lifecycle. Set-level refresh as a 6-month project. Refresh budget rule of thumb.

  • references/common-pseo-failures.md

    12 failure patterns with diagnoses and fixes. Cross-references to the other reference files. The pattern across all failures: most pSEO failures are designed-in at launch shape, not produced by operational drift.

Browse all reference files on GitHub

Pairs with these platforms

Nine platforms across four categories.

Programmatic SEO is genuinely the most platform-composable skill in the catalog. It touches the data stack (BigQuery, Snowflake, dbt for sources and pipelines), the publishing stack (Webflow, Contentful as destinations for generated pages), the content optimization stack (Frase for per-page optimization, AirOps for workflow), and the analytics stack (Profound for AEO measurement, Mixpanel for engagement) all at once. Most pSEO programs end up integrating across at least three of these categories.

Bridges to sister and adjacent skills

Five sister skills compose into the discipline.

This skill does not stand alone. Five skills connect to it: program companion above (decides whether pSEO fits), editorial alternative (the other end of the content spectrum), upstream research input (keyword space), QA companion forthcoming, and AEO/GEO strategy adjacent.

  • Program companion (above)

    content-strategy

    Decides whether pSEO fits the program at all. Editorial pillars, calendar, governance. The strategic decision precedes pSEO design; this skill takes the program-level go-ahead and designs the system.

  • Editorial alternative

    pillar-content-architecture

    Hub-level editorial architecture: one topic with 10 to 15 intentional editorial pieces. The two skills cover different ends of the content spectrum; programs often benefit from both, with editorial pillars handling analytical-query lane and pSEO handling factual-query lane.

  • Upstream research input

    seo-keyword

    Surfaces the long-tail keyword space pSEO would target. Volume, difficulty, intent classification across hundreds or thousands of candidate queries. Without this research, pSEO targets imagined queries instead of real ones.

  • QA companion (forthcoming)

    editorial-qa

    Sampled quality-control discipline across content sets. Forthcoming as the next content-suite skill. Will codify the QA practices this skill points to (5% failure threshold, cohort tracking, manual review checklist) at the catalog level rather than just per-program.

  • Strategy adjacent

    seo-aeo-geo

    AI search optimization at the program level. The two-engine optimization principles in this skill are applications of the broader AEO/GEO strategy work that lives there. Pair when designing pSEO templates that need to compete in both search rankings and AI citations.

Where this skill fits in the content suite

The third of five in the content suite.

Programmatic SEO is the third skill in a planned five-skill content suite. Skills 1 and 2 covered editorial discipline: content-brief-authoring for per-piece briefs, and pillar-content-architecture for hub-level architecture. This skill covers the scaled-content alternative for programs with real underlying data and queryable intent.

Skills 4 and 5 in the suite (forthcoming) cover editorial QA at the catalog level and AI content collaboration when some writers are agents. Both apply to pSEO programs as well as editorial programs; the QA discipline this skill points to (5% sample failure threshold, cohort tracking, manual review checklist) gets codified at the catalog level when the editorial-qa skill ships.

The suite ships sequentially. Skills 4 and 5 do not exist yet; cross-linking lands when those skill pages ship. For now, the pSEO discipline anchors the scaled-content lane and the nine platforms above cover the platform-specific tactics underneath.

Open source under MIT

Read the SKILL.md on GitHub.

The skill source lives in the rampstackco/claude-skills repository alongside dozens of other skills covering the full lifecycle of brand and product work. MIT licensed.

Frequently asked questions.

When is programmatic SEO actually the right answer?
When five criteria are all met: real underlying data (10+ fields per record, ideally 20+, with first-party, licensed, expert-curated, or synthesized sources), long-tail query volume that justifies the build, user intent that can be answered through structured data presented well, an update cadence that aligns with the data's volatility, and quality-control headcount budgeted in. If any one of these is missing, the program tends toward penalty-bait. The honest framing: most teams asking 'should we do pSEO?' should hear 'probably not, unless the underlying data is unique.' Detail in the when-pseo-works-decision reference.
How is this different from pillar-content-architecture?
Pillar content architecture is hub-scope: ONE topic with 10 to 15 intentional editorial pieces, each manually briefed and written. Programmatic SEO is scaled scope: 100s to 100,000s of pages generated from structured data sources, each targeting a long-tail query. Editorial pillars compound through topical depth and named-author expertise; pSEO sets compound through data depth and structural linking. The two are different disciplines with different operational shapes; teams sometimes need both, with editorial pillars handling the analytical-query lane and pSEO handling the factual-query lane.
Why do so many programmatic SEO programs get penalized?
Almost always because quality control was not budgeted at scale. The team designed for scale (10,000 pages generated efficiently) without designing for quality at scale (sample auditing, failure thresholds, cohort tracking, refresh discipline). The set looks fine on launch, drifts as data ages and templates ship untracked changes, and gets caught by the next algorithm update. The discipline is upstream: a team that budgets 0.5 to 1.0 FTE per 10,000 pages for ongoing QC tends to maintain quality; a team that ships and assumes QC will happen later tends not to.
How does pSEO work for AI search engines (AEO and GEO)?
AI engines treat programmatic pages well when the template leads with a self-contained answer in the first 200 words, when JSON-LD schema is comprehensive and validated, when FAQPage schema marks any FAQ sections, and when factual claims include source citations. AI engines tend to cite programmatic pages for factual queries (prices, statistics, comparisons) and editorial pages for analytical queries. The two-engine optimization framing applies: pSEO that earns search rankings is the same pSEO that earns AI citations. Designing for AEO does not trade off against SEO.
What does the team budget look like for a 50,000-page set?
Roughly 2 to 4 FTE of ongoing capacity: data pipeline maintenance for volatile fields, quality control sampling cycles (200 pages per cycle, monthly), automated check failure triage, manual review on sampled pages, cohort tracking analysis, refresh cycle execution, and template version migrations when they ship. Budgeted in headcount before generation starts. Teams that under-budget hope for the best; the program degrades within 6 to 12 months. Teams that budget honestly produce durable traffic for years.
What if our data source is not unique?
Then pSEO is probably the wrong answer. Replicable data has no defensibility; competitors will launch the same set with the same content and the program loses to whichever team has the larger SEO budget. The fix is to deepen the data: first-party accumulation (build the dataset by getting users or experts to populate it), licensing (acquire exclusive or hard-to-replicate feeds), expert curation (hire experts to populate), or synthesis (combine multiple sources into a unique view). If the data cannot be deepened to a moat, the program does not have a long-term future and editorial content is the right channel instead.