Skip to main content
AI ProductivitySEO

How I automated 80% of SEO work with N8N and LLMs

Building AI-powered workflows that transformed our organic acquisition capabilities at Picsart.

Niels KaspersNiels Kaspers
January 10, 2025
10 min read
How I automated 80% of SEO work with N8N and LLMs

TL;DR

I built automated pipelines with N8N and LLMs to handle content optimization, keyword research, and technical SEO monitoring for 50,000+ pages. Manual SEO doesn't scale—automation does.

SEO at scale is a nightmare. When you're managing thousands of pages across dozens of tools, manual optimization becomes impossible. So I built a system that does 80% of the work automatically.

This isn't a theory post. I'm going to walk through the actual workflows, the specific APIs, and the real numbers.

The Problem

At Picsart, we had:

  • 50,000+ live pages
  • 40+ languages
  • Constantly changing search trends
  • A team of... me

Traditional SEO workflows don't scale. You can't manually optimize thousands of pages, track ranking changes, and respond to algorithm updates with one person. Even with a team of five, you'd be drowning. Something had to change.

I'd already seen what small teams with good systems could do at Quicktools. The principle was the same here: don't hire more people, build better infrastructure.

Why N8N

I evaluated several workflow automation tools before landing on N8N. Zapier and Make are fine for simple triggers, but they get expensive fast when you're processing thousands of records. N8N is open-source, self-hosted, and doesn't charge per execution. When you're running workflows that touch 50,000 pages, that matters.

I host it on a basic VPS. Total cost: about $20/month. The equivalent Zapier setup would run $500+/month at the volumes I needed.

N8N also has native HTTP request nodes, JavaScript code nodes, and a built-in scheduler. That combination covers 90% of what you need for SEO automation.

The Solution: N8N + LLMs

I built an automated pipeline using N8N and various LLMs to handle the three biggest time sinks in SEO: content optimization, keyword research, and technical monitoring.

1. Content Optimization Pipeline

This was the first workflow I built, and it had the biggest impact.

The workflow:

  1. Trigger: Scheduled daily at 6 AM EST
  2. Google Search Console API pull: Fetch all queries where our pages rank positions 4-20 (the "striking distance" keywords)
  3. Ahrefs API call: For each keyword, pull the top 5 ranking pages and their content structure
  4. LLM analysis: Send the competitor content + our current page to GPT-4 with a prompt that identifies specific content gaps—missing sections, thin answers, missing entity coverage
  5. Generate suggestions: The LLM outputs structured JSON with recommended title changes, meta description updates, new H2 sections to add, and internal links to include
  6. Strapi CMS update: For straightforward changes (meta titles, descriptions), the workflow pushes updates directly through our CMS API. For structural changes, it creates a ticket in JIRA with the recommendations

The key insight: I don't let the LLM write final copy. It identifies gaps and suggests structure. A human reviews structural changes before they go live. But meta tag optimization? That runs on autopilot. I built a dedicated SEO meta tag generator prompt that handles the formatting rules, character limits, and brand voice consistently.

Numbers: This workflow processes about 2,000 pages per week. Before automation, I could manually optimize maybe 20 pages per week. That's a 100x throughput increase.

2. Keyword Research at Scale

Manual keyword research is one of those tasks that feels productive but scales terribly. You open Ahrefs, search a seed keyword, export to a spreadsheet, cluster manually, repeat. I was spending 6-8 hours per week on this.

The workflow:

  1. Trigger: Weekly, every Monday morning
  2. Ahrefs API: Pull trending keywords in our category (image editing, AI tools, design) with volume > 1,000 and keyword difficulty < 40
  3. Google Trends API: Cross-reference with rising search trends to catch emerging topics early
  4. LLM clustering: Feed the keyword list to Claude with instructions to cluster by user intent (informational, transactional, navigational) and topic group
  5. Scoring: Custom JavaScript node that scores each cluster based on: search volume, competition level, relevance to our product, and whether we already have a page targeting it
  6. Output: A prioritized list of keyword opportunities pushed to a Google Sheet, with recommended page types (tool page, blog post, landing page) for each cluster

This is where it gets interesting. The workflow doesn't just find keywords—it suggests what to build. When it identifies a cluster around "AI background remover for product photos," it knows we have a background remover tool and recommends a landing page targeting that specific use case. That recommendation feeds directly into our programmatic SEO engine and landing page generator.

Numbers: The system identifies 50-100 new keyword opportunities per week. Before automation, I found maybe 10-15. More importantly, it catches trends 2-3 weeks earlier than manual research would.

3. Technical SEO Monitoring

Technical SEO is the least glamorous part of the job, but broken pages lose you traffic fast. With 50,000 pages across 40 languages, something is always broken.

The workflow:

  1. Trigger: Every 6 hours
  2. Sitemap crawler: Custom Python script that parses all sitemaps and checks HTTP status codes
  3. Google Search Console API: Pull indexing errors, coverage issues, and Core Web Vitals alerts
  4. Page speed check: Lighthouse API calls on a rotating sample of 100 pages per run
  5. Schema validation: Check structured data markup against Google's requirements
  6. Alert routing: Critical issues (5xx errors, deindexed pages) go to Slack immediately. Non-critical issues (slow pages, missing schema) get batched into a weekly report

The 6-hour cadence matters. When a deployment breaks something, I know within hours instead of discovering it in the next weekly review when traffic has already dropped.

Numbers: This catches an average of 3-4 critical issues per month that would otherwise go unnoticed for days. At our traffic levels, a broken page can cost 10,000+ visits per day.

4. Internal Linking Automation

This one was a game-changer I didn't expect. I built an internal linking system that maps every page on the site, identifies orphan pages (pages with zero internal links pointing to them), and suggests link placements.

The workflow:

  1. Crawl all pages and extract existing internal links
  2. Build a link graph
  3. Identify orphan pages and pages with thin link profiles
  4. Use an LLM to match pages by topical relevance
  5. Generate specific anchor text and placement suggestions
  6. Push link additions through the CMS API

We found 127 orphan pages on the first run. Pages that existed, had content, ranked for nothing—because no other page on the site linked to them. After adding internal links, 40% of those pages started ranking within 6 weeks.

The Localization Layer

Here's where it gets really complex. Everything I described above needs to work across 40+ languages. You can't just translate and publish—each language market has different search patterns, different competitors, different content expectations.

I built a localization system that:

  • Adapts keyword research per market (German users search differently than Spanish users)
  • Localizes meta tags with market-specific conventions
  • Monitors ranking performance per language independently
  • Flags when a localized page is underperforming vs. its English equivalent

The LLM handles translation quality checks, but the real value is in the market-specific optimization. A page that ranks #1 in English might need a completely different title in Japanese.

The Stack

Here's everything running in production:

  • N8N: Workflow orchestration (self-hosted, $20/month VPS)
  • GPT-4 + Claude: Content analysis and generation (about $200/month in API costs)
  • Google Search Console API: Performance data (free)
  • Ahrefs API: Competitor analysis and keyword data ($99/month plan)
  • Strapi CMS: Content management and API updates
  • Python scripts: Data processing, sitemap crawling, custom analysis
  • Google Sheets: Output dashboards for team visibility
  • Slack: Alert routing

Total automation infrastructure cost: roughly $320/month. The equivalent in human hours would be 2-3 full-time SEO specialists at $60-80K/year each.

Results

After 6 months of running these systems:

  • 3X organic traffic growth (from 3.1M to 9.3M monthly organic sessions)
  • 90% reduction in manual SEO tasks (from 40+ hours/week to about 4 hours of review)
  • 2-day response time to algorithm updates (down from 2-3 weeks)
  • 500+ new keyword rankings in top 10
  • 127 orphan pages fixed, 40% now ranking

The 3X traffic growth wasn't just from automation—it was from speed. When you can identify opportunities and execute on them in days instead of weeks, you compound faster than competitors who are still doing manual keyword research in spreadsheets.

What Didn't Work

Not everything was smooth. A few failures worth sharing:

Auto-publishing blog content: I tried having the LLM write and publish blog posts automatically. The quality was inconsistent enough that I rolled it back within a month. LLMs are great for structured content (meta tags, product descriptions) but unreliable for long-form editorial. I now use a blog post outline prompt to generate structure, then write or heavily edit the actual content.

Real-time algorithm monitoring: I built a workflow that scraped SEO Twitter and Google's blog for algorithm update signals. Too noisy. Most "algorithm updates" that people tweet about are normal volatility. I simplified this to a weekly digest instead.

Fully automated link building: Tried automating outreach emails. Response rates were terrible and it felt spammy. Killed it after two weeks. Some things still need a human touch.

Key Learnings

  1. LLMs are great for first drafts, not final copy. Always have human review for important pages. Let automation handle the 80% that's formulaic—meta tags, structured data, internal links. Keep humans on the 20% that requires judgment.

  2. Start with the most repetitive tasks. The biggest ROI comes from automating what you do most often. For me, that was meta tag optimization and keyword clustering. Not the sexy stuff, but it freed up 20+ hours per week.

  3. Build in quality checks. Automated doesn't mean unmonitored. Every workflow has validation steps and anomaly detection. If a meta title comes back with weird characters or exceeds 60 characters, it gets flagged instead of published.

  4. Document everything. When you automate, you need to understand what's happening to debug issues. Every workflow has inline comments explaining the logic. My Claude Code skill library follows the same principle—every automation is documented well enough that someone else could maintain it.

  5. The compound effect is real. Each automation frees up time to build the next one. After 6 months, I had a system that was generating insights I never would have found manually—because I had the bandwidth to look at the data instead of doing the grunt work.

What's Next

I'm now building a GTM content orchestrator that connects all these SEO workflows with content distribution—so when we identify a keyword opportunity, the system doesn't just create the page, it also generates social posts, internal links, and distribution plans.

The end state is a system where I spend my time on strategy and the machines handle execution. We're about 80% of the way there.

If you're an SEO team of one managing a large site, you don't need more headcount. You need better pipes. Start with one workflow, prove it works, and build from there. The tools are all available—N8N is free, the APIs exist, and LLMs are cheap enough to experiment with.


Want to learn more about AI-powered SEO? Check out how we built Quicktools to 10M users or read about being an early AI adopter.

Niels Kaspers

Written by Niels Kaspers

Principal PM, Growth at Picsart

More articles

Get in touch

Have questions or want to discuss this topic? Let me know.