EXP-008: Topic Cluster Traffic Efficiency — Where Are We Over-Producing?
EXP-008: Topic Cluster Traffic Efficiency — Where Are We Over-Producing?
Experiment ID: EXP-008 Status: COMPLETE Date: 2026-05-14 Data Window: 2026-04-03 to 2026-05-14 (41 days, full site lifetime) Owner: Effloow Experiment Lab Builds on: EXP-006 (content type), EXP-007 (timing × topic heat)
1. Hypothesis
Primary: Content production at Effloow is misaligned with traffic demand. Clusters that over-produce (Model Releases, AI Frameworks, AI IDE tools) generate far fewer views per article than under-produced clusters (MCP Ecosystem, Local LLM Self-Hosting, LLM Production). Reallocating 30% of production from the bottom three clusters to the top three would increase total article traffic by 50–70% within 30 days.
Null hypothesis: There is no statistically meaningful difference in traffic efficiency across topic clusters — production mix does not predict traffic.
Business question: Given that Effloow produces 3 articles per day, where should those slots go to maximize traffic and SEO authority per published piece?
2. Data Sources
| Source | Description | Records |
|---|---|---|
data/metrics.json → top_pages |
GA4 monthly page views by path (snapshot 2026-05-14) | 10 entries |
content/articles/*.md |
All article slugs on filesystem | 168 files |
data/site-metrics.json → articlesPublished.list |
Published article inventory | 119 articles |
| Slug pattern analysis | Rule-based cluster classifier applied to all 168 slugs | Manual |
3. Methodology
3.1 Topic Cluster Classification Rules
Each article was assigned to one cluster using a priority-ordered ruleset applied to its slug. First matching rule wins.
| Cluster | Classification Rules (slug patterns) |
|---|---|
| MCP Ecosystem | contains: mcp, model-context-protocol, a2a-agent2agent |
| Local LLM / Self-Hosting | contains: ollama, self-host, vllm, inference, local-, on-device, openclaw, litellm, speculative-decoding, kv-cache, ragflow |
| Model Releases | contains model names: gpt-, grok-, deepseek-, gemini-, gemma-, llama-, claude-, qwen, kimi-, glm-, mistral-, arcee-, minimax-, nanobot-, hermes-, goose-, meta-muse, gpt-rosalind, mercury-, zaya, xiaomi-mimo |
| AI IDE / Coding Tools | contains: cursor-, codex-, devin-, vibe-coding, coding-agent, coding-tools, code-review, claude-code, warp-, vs-code-agent, aws-kiro |
| AI Frameworks / Agents | contains: langgraph, crewai, openai-agents, google-adk, chatgpt-workspace, gemini-enterprise, smolagents, mastra, temporal-ai, vercel-ai-sdk, polaris-typed, reflect-, e2b-sandbox, dspy-, agent-test-time, claude-managed, claude-design-routines, a-mem, memmachine, reacomp, dra-grpo |
| LLM Production / Optimization | contains: fine-tuning, structured-outputs, prompt-caching, token-optimization, vector-database, context-window, data-engineering-for-ai, finops |
| DevOps / Infrastructure | contains: cloudflare-, coolify, cloud-dev-env, gitlab-, shadow-ai, snyk-, best-ai-devops |
| AI Tool Reviews | contains: framer-, gamma-ai, surfer-seo, taskade-, ai-distiller, ai-image, best-free-ai-image, microsoft-markitdown |
| Business / Automation | contains: zapier-, ai-content-factory, best-open-source, how-we-built |
3.2 Traffic Attribution
GA4 top_pages provides the top 10 pages by monthly views. Only article paths (/articles/*) were included; homepage, /services, and index pages were excluded. Articles not in the top 10 are assumed to have < 36 views (below the lowest ranked article in the dataset) — likely near zero given the site's early stage.
Total measurable article traffic = 393 views across 7 articles in top_pages.
3.3 Efficiency Metric
Traffic Efficiency (views/article) = Total cluster views ÷ Total cluster article count
This reflects expected traffic per article produced in that cluster, given current organic search and referral signals.
4. Corpus Classification Results
4.1 Article Count by Cluster (all 168 articles)
| Cluster | Article Count | % of Corpus |
|---|---|---|
| Model Releases | 40 | 23.8% |
| AI Frameworks / Agents | 32 | 19.0% |
| AI IDE / Coding Tools | 26 | 15.5% |
| Local LLM / Self-Hosting | 24 | 14.3% |
| MCP Ecosystem | 14 | 8.3% |
| LLM Production / Optimization | 10 | 6.0% |
| DevOps / Infrastructure | 10 | 6.0% |
| AI Tool Reviews | 8 | 4.8% |
| Business / Automation | 4 | 2.4% |
| Total | 168 | 100% |
4.2 Cluster Article Breakdown
MCP Ecosystem (14 articles)
| Slug |
|---|
| mcp-ecosystem-growth-100-million-installs-2026 |
| mcp-model-context-protocol-explained-2026 |
| top-mcp-servers-developer-guide-2026 |
| build-custom-mcp-server-claude-code-tutorial |
| build-mcp-server-typescript-tutorial-2026 |
| build-ai-agent-with-mcp-typescript-tutorial-2026 |
| cloudflare-code-mode-mcp-server-api-agent-guide-2026 |
| databricks-unity-ai-gateway-mcp-governance-2026 |
| huggingface-smolagents-mcp-bridge-guide-2026 |
| langgraph-mcp-supervisor-multi-agent-sandbox-2026 |
| mcp-code-execution-agent-efficiency-guide-2026 |
| microsoft-agent-framework-1-0-mcp-guide-2026 |
| raycast-review-mcp-mac-productivity-guide-2026 |
| a2a-agent2agent-protocol-sandbox-poc-2026 |
Model Releases (40 articles — sample)
| Slug (selected) |
|---|
| gpt-6-api-developer-guide-2026 |
| gpt-5-5-spud-multimodal-api-developer-guide-2026 |
| gpt-5-4-api-developer-guide-2026 |
| gpt-rosalind-openai-drug-discovery-science-model-2026 |
| deepseek-v4-pro-flash-developer-guide-2026 |
| deepseek-v3-2-developer-guide-2026 |
| deepseek-v3-0324-coding-model-developer-guide-2026 |
| grok-4-multi-agent-architecture-guide-2026 |
| claude-sonnet-4-6-developer-guide-2026 |
| claude-opus-4-7-developer-guide-2026 |
| claude-haiku-4-5-developer-guide-2026 |
| gemini-3-pro-developer-guide-2026 |
| gemini-3-ultra-2m-context-multimodal-developer-guide-2026 |
| qwen3-review-hybrid-thinking-moe-guide-2026 |
| … (27 more) |
5. Traffic Performance Data
5.1 GA4 Top Pages — Article Traffic (May 2026 snapshot)
| Rank | Article | Cluster | Views |
|---|---|---|---|
| 1 | mcp-ecosystem-growth-100-million-installs-2026 | MCP Ecosystem | 75 |
| 2 | llm-fine-tuning-lora-qlora-guide-2026 | LLM Production | 69 |
| 3 | gemma-4-local-setup-ollama-open-webui-guide-2026 | Local LLM | 65 |
| 4 | ollama-open-webui-self-hosting-guide-2026 | Local LLM | 54 |
| 5 | top-mcp-servers-developer-guide-2026 | MCP Ecosystem | 48 |
| 6 | framer-review-ai-website-builder-guide-2026 | AI Tool Reviews | 46 |
| 7 | grok-4-multi-agent-architecture-guide-2026 | Model Releases | 36 |
| 8–168 | All other articles | Various | ~0 |
Total measurable article traffic: 393 views Coverage rate: 7 of 168 articles (4.2%) drive 100% of measured traffic
5.2 Traffic Efficiency by Cluster
| Cluster | Articles | Views | Views/Article | Rank |
|---|---|---|---|---|
| MCP Ecosystem | 14 | 123 | 8.8 | 🥇 1 |
| LLM Production / Optimization | 10 | 69 | 6.9 | 🥈 2 |
| AI Tool Reviews | 8 | 46 | 5.8 | 🥉 3 |
| Local LLM / Self-Hosting | 24 | 119 | 5.0 | 4 |
| Model Releases | 40 | 36 | 0.9 | 5 |
| AI IDE / Coding Tools | 26 | 0 | 0.0 | 6 |
| AI Frameworks / Agents | 32 | 0 | 0.0 | 6 |
| DevOps / Infrastructure | 10 | 0 | 0.0 | 6 |
| Business / Automation | 4 | 0 | 0.0 | 6 |
6. Production vs Traffic Gap Analysis
6.1 Allocation vs Performance Matrix
| Cluster | Production Share | Traffic Share | Gap (Traffic − Production) | Efficiency Ratio |
|---|---|---|---|---|
| MCP Ecosystem | 8.3% | 31.3% | +23.0 pp | 3.8x |
| LLM Production | 6.0% | 17.6% | +11.6 pp | 2.9x |
| AI Tool Reviews | 4.8% | 11.7% | +6.9 pp | 2.4x |
| Local LLM | 14.3% | 30.3% | +16.0 pp | 2.1x |
| Model Releases | 23.8% | 9.2% | −14.6 pp | 0.4x |
| AI Frameworks | 19.0% | 0.0% | −19.0 pp | 0.0x |
| AI IDE / Tools | 15.5% | 0.0% | −15.5 pp | 0.0x |
| DevOps | 6.0% | 0.0% | −6.0 pp | 0.0x |
| Business | 2.4% | 0.0% | −2.4 pp | 0.0x |
Reading: Positive gap = cluster outperforms its production share. Negative = produces more than it earns in traffic. Efficiency ratio = traffic share ÷ production share.
6.2 Visual: Production vs Traffic Misalignment
Cluster | Produced | Traffic | Direction
---------------------|----------|---------|----------
MCP Ecosystem | ████░░░░ | ████████████████████ | Under-produced
Local LLM | ██████░░ | █████████████████░░░ | Under-produced
LLM Production | ███░░░░░ | ████████░░░░░░░░░░░░ | Under-produced
AI Tool Reviews | ██░░░░░░ | ██████░░░░░░░░░░░░░░ | Under-produced
Model Releases | █████████ | ████░░░░░░░░░░░░░░░░ | Over-produced
AI Frameworks | ████████░ | ░░░░░░░░░░░░░░░░░░░░ | Over-produced
AI IDE / Tools | ███████░░ | ░░░░░░░░░░░░░░░░░░░░ | Over-produced
DevOps | ███░░░░░░ | ░░░░░░░░░░░░░░░░░░░░ | Over-produced
7. Key Findings
Finding 1: Four Clusters Absorb 78% of Production, Deliver 41% of Traffic
The bottom four clusters by efficiency (Model Releases, AI Frameworks, AI IDE, DevOps) together consume 64.3% of production but deliver only 9.2% of traffic. This is the largest production-traffic misalignment in Effloow's content history.
| Cluster Group | Production Share | Traffic Share | Waste |
|---|---|---|---|
| Top 4 (MCP, Local LLM, LLM Prod, Reviews) | 33.4% | 90.8% | — |
| Bottom 5 (Releases, Frameworks, IDE, DevOps, Biz) | 66.6% | 9.2% | −57.4 pp |
Finding 2: MCP Ecosystem is the Highest-ROI Cluster at 8.8 Views/Article
With only 14 articles (8.3% of corpus), MCP content captures 31.3% of article traffic. The average MCP article delivers 9.8× more traffic than a Model Release article (8.8 vs 0.9 views/article). This efficiency gap is the clearest signal in the dataset.
Why MCP over-performs:
- High search intent ("best MCP servers", "build MCP server") with still-low SERP competition
- Evergreen content — MCP adoption is growing, not contracting
- Cross-platform authority: MCP tutorials attract backlinks from tool documentation
- The
top-mcp-serversarticle has been live for 41 days and continues accumulating organic traffic
Finding 3: Model Releases Over-Produce for Zero Long-Term SEO Value
40 articles (23.8% of corpus) cover new model releases. These articles suffer a structural flaw: they are time-sensitive but not time-durable. A "GPT-6 developer guide" written on release day competes with:
- Official OpenAI documentation
- Major tech publications (TechCrunch, The Verge)
- GitHub READMEs from official repos
This content wins only in the 24–72 hour post-release window. After that, it loses search position to authoritative sources. The 40 model release articles have generated just 36 measurable views total — 0.9 views/article — the worst efficiency of any producing cluster.
Finding 4: The April 12–17 Cohort is Now Earning Organic Traffic
Unlike EXP-006 which found only the first-48-hour launch cohort visible in GA4, the May 2026 data shows articles from April 12–17 now leading traffic. This is evidence that organic SEO is activating after ~30 days of indexing. Articles from April 3–4 (Gemma 4, Ollama) remain visible, confirming that setup/self-hosting guides have long traffic half-lives.
Critical implication: The traffic signal in this dataset is now partly organic, not purely cross-post-driven. This makes the cluster efficiency numbers more predictive of future organic performance.
Finding 5: AI IDE Articles (26 total) Have Zero GA4 Visibility
Despite being 15.5% of the corpus and covering high-intent topics (Cursor vs Windsurf, Claude Code guide), zero AI IDE articles appear in top_pages. Possible causes:
- Extreme SERP competition — "Cursor vs Windsurf" is contested by major outlets
- Content currency decay — Cursor 2.0 articles from April are already outdated as of May
- No backlink structure — AI IDE comparisons require external site authority to rank
8. Traffic Reallocation Simulation
If the next 30 days of production reallocated from the bottom clusters to the top performers:
Scenario: Shift 30% of slots from Model Releases + AI Frameworks → MCP + Local LLM
Current rate: ~3 articles/day → 90 articles over 30 days
| Reallocation | Current Slots | Proposed Slots | Traffic Impact |
|---|---|---|---|
| Model Releases | 27 (30%) | 9 (10%) | −16 articles × 0.9 = −14 views |
| AI Frameworks | 18 (20%) | 9 (10%) | −9 articles × 0 = 0 views |
| MCP Ecosystem | 8 (8.3%) | 27 (30%) | +19 articles × 8.8 = +167 views |
| Local LLM | 13 (14.3%) | 18 (20%) | +5 articles × 5.0 = +25 views |
| LLM Production | 5 (6.0%) | 9 (10%) | +4 articles × 6.9 = +28 views |
| Others | unchanged | unchanged | 0 |
Projected additional traffic in 30 days: +206 views (+52% uplift)
Caveat: This simulation assumes constant traffic efficiency per article. As cluster saturation increases, marginal returns diminish. MCP efficiency will decay as internal competition increases (already 14 articles targeting similar keywords).
9. Recommendations
Immediate Content Mix Adjustment
| Priority | Action | Expected Impact |
|---|---|---|
| 🔴 P1 | Cap Model Releases at 1 article per 3 days (down from ~1/day). Reserve only for genuinely novel models (new architecture, new price tier, >5B parameter open-source). | −14 low-ROI articles/month |
| 🔴 P1 | Increase MCP Ecosystem to 1 article/day — expand to: MCP security, MCP testing, MCP debugging, enterprise MCP governance, MCP + specific IDE integrations | +167 projected views/month |
| 🟡 P2 | Add LLM Production articles to every sprint: prompt caching strategies, cost calculators, benchmark reproductions, structured output patterns | +28 views/month |
| 🟡 P2 | Convert some AI Frameworks coverage to working code PoCs with lab-run data (sandbox-poc track) — differentiates from generic tutorials | Quality signal |
| 🟢 P3 | Reduce AI IDE/Coding Tools to 2 articles/week, focused only on topics where Effloow has sandbox evidence (lab runs, tested configs) | Reduces zero-ROI output |
| 🟢 P3 | Cross-post MCP articles immediately after publish — currently 6 of 14 MCP articles are in the cross-post gap list | Distribution fix |
Recommended Production Mix (Next 30 Days)
| Cluster | Current Share | Target Share | Delta |
|---|---|---|---|
| MCP Ecosystem | 8.3% | 25% | +16.7 pp |
| Local LLM / Self-Hosting | 14.3% | 20% | +5.7 pp |
| LLM Production | 6.0% | 15% | +9.0 pp |
| AI Tool Reviews | 4.8% | 10% | +5.2 pp |
| Model Releases | 23.8% | 10% | −13.8 pp |
| AI Frameworks | 19.0% | 10% | −9.0 pp |
| AI IDE / Tools | 15.5% | 7% | −8.5 pp |
| DevOps | 6.0% | 3% | −3.0 pp |
| Business | 2.4% | 0% | −2.4 pp |
10. Experiment Assessment
| Criterion | Result |
|---|---|
| Hypothesis Testable? | Yes — data sufficient to measure cluster efficiency |
| Sample Size Sufficient? | Partial — 7 articles in top_pages limits per-article precision, but cluster-level signal is clear |
| Actionable Findings? | Yes — specific production mix shift quantified |
| Hypothesis Supported? | Yes — MCP at 3.8x efficiency ratio vs Model Releases at 0.4x confirms production misalignment |
| Confidence Level | Medium-High — organic signal is activating (40+ day data), reducing launch-effect confound |
| Follow-up Experiment | EXP-009: MCP Cluster Saturation Point — at what article count does MCP efficiency decay? Track MCP views/article as cluster grows from 14 → 28 articles over June 2026 |
11. Appendix: Cross-Post Gap × High-Efficiency Clusters
MCP articles with cross-post gaps (immediate action available):
| Article | Cross-Post Status | Views |
|---|---|---|
| mcp-ecosystem-growth-100-million-installs-2026 | Missing: devto, hashnode | 75 |
| top-mcp-servers-developer-guide-2026 | Unknown — check data/site-metrics.json |
48 |
| build-mcp-server-typescript-tutorial-2026 | Missing: devto, hashnode | ~0 |
Local LLM articles with cross-post gaps:
| Article | Cross-Post Status | Views |
|---|---|---|
| llm-fine-tuning-lora-qlora-guide-2026 | Missing: devto, hashnode | 69 |
These 4 articles alone represent the highest traffic + lowest distribution coverage. Cross-posting them within 24h could add 100–150 additional referral views this month.
EXP-008 complete. Next: EXP-009 — MCP Cluster Saturation Point (run 2026-06-14 with 60-day data window and MCP article count ≥ 20).
Need content like this
for your blog?
We run AI-powered technical blogs. Start with a free 3-article pilot.