u/Tricky_Literature397

▲ 3 r/n8n

Help: How to scrape dynamic websites using n8n

Hi everyone,

I’m working with n8n and trying to scrape data from dynamic websites (JavaScript-rendered pages), but I’m running into some limitations.

For example, I’m trying to extract content from pages like this:
http://www.iort.gov.tn/WD120AWP/WD120Awp.exe/CTX_9648-63-ijOAVgefuu/CodesJuridiques/SYNC_218121892

The issue is that:

  • The page content is loaded dynamically (not fully available in the initial HTML)
  • The URL changes randomly every time (session-based or generated links), so it’s not stable
  • Using the HTTP Request node in n8n doesn’t return the actual rendered content
  • I suspect it relies on JavaScript execution or internal requests

What I’ve tried so far:

  • Basic HTTP Request node → only returns partial/empty HTML
  • Comparing page source vs inspected DOM → content mismatch

My questions:

  1. What’s the best way to scrape this kind of dynamic website using n8n?
  2. Is there a way to integrate a headless browser (like Puppeteer or Playwright) with n8n?
  3. How do you handle scraping when URLs are dynamic/session-based like this?
  4. Should I try to replicate the underlying API calls from the Network tab instead?
  5. Any recommended workflow architecture for handling this reliably?

I’d really appreciate any tips, best practices, or examples 🙏

Thanks!

reddit.com
u/Tricky_Literature397 — 22 hours ago
▲ 21 r/n8n_ai_agents+1 crossposts

Best Ollama model for n8n workflows (RAG, file handling, reasoning) + hardware requirements?

Hi everyone,

I’m currently building automation workflows using n8n with local LLMs via Ollama, and I’m trying to choose the most suitable model for production/company use.

My main use cases:

  • RAG (retrieval-augmented generation) with documents (PDFs, text, etc.)
  • File handling & structured data extraction
  • Reasoning tasks (not just simple chat)
  • Reliable JSON outputs for automation

Constraints:

  • Running locally on a physical server (not cloud)
  • Looking for a good balance between performance, speed, and accuracy

Questions:

  1. Which Ollama models would you recommend for these use cases? (e.g., LLaMA 3, Mistral, Mixtral, DeepSeek, etc.)
  2. Which models handle RAG + structured outputs best?
  3. What are the minimum and recommended hardware specs (RAM, GPU/CPU) for smooth performance in production?
  4. Any tips for optimizing n8n + Ollama workflows (latency, batching, etc.)?

I’d really appreciate feedback from anyone using this setup in real-world scenarios.

Thanks!

reddit.com
u/Tricky_Literature397 — 4 days ago