u/NefariousnessSharp61

▲ 3 r/n8n

Free GPT-4.1 API access for ~12hrs — works directly with n8n's OpenAI node

Hey n8n folks,

Stress testing my OpenAI-compatible reverse proxy gateway. Since it's fully OpenAI-compatible, it just drops into n8n's OpenAI node with zero config changes — just swap the base URL.

Available models:

  • gpt-4.1 — Latest, 1M context
  • gpt-4.1-mini / gpt-4.1-nano
  • o4-mini — reasoning
  • gpt-4o-mini-tts — TTS node compatible

Comment your workflow type and I'll DM the endpoint + key.
(Non-commercial side project, no paid tier)

reddit.com
u/NefariousnessSharp61 — 4 hours ago

Free GPT-4.1 + o4-mini access for ~12hrs — testing my reverse proxy under agent workloads

Hey,

I've been building an OpenAI-compatible reverse proxy for routing agent traffic and want to stress test it with real agentic workloads before open-sourcing.

Available for ~12 hours:

  • gpt-4.1 — 1M context, great for long agent chains
  • gpt-4.1-mini / gpt-4.1-nano — fast tool calling
  • o4-mini — reasoning tasks
  • gpt-4o-mini-tts — TTS

Works with LangChain, LangGraph, AutoGen, CrewAI — any OpenAI-compatible framework.

Comment your use case in 1 line and I'll DM the key. Keeping it comment-gated to avoid bot flooding.

Will share latency + error stats in a follow-up.
(Personal project, non-commercial, no paid tier)

reddit.com
u/NefariousnessSharp61 — 4 hours ago

Built an OpenAI-compatible API reverse proxy — opening for community stress testing for ~12hrs (GPT-4.1, o4-mini, TTS)

Hey Devs,

I've been building a personal, non-commercial OpenAI-compatible reverse proxy gateway that handles request routing, retry logic, token counting, and latency tracking across multiple upstream endpoints.

Before I finalize the architecture, I want to stress test it under real-world concurrent load — synthetic benchmarks don't catch the edge cases that real developer usage does.

Available models:

  • gpt-4.1 — Latest flagship, 1M context
  • gpt-4.1-mini — Fast, great for agents
  • gpt-4.1-nano — Ultra-low latency
  • gpt-4o — Multimodal capable
  • gpt-4o-mini — High throughput
  • gpt-5.2-chat — Azure-preview, limited availability
  • o4-mini — Reasoning model
  • gpt-4o-mini-tts — TTS endpoint

Works with any OpenAI-compatible client — LiteLLM, OpenWebUI, Cursor, Continue dev, or raw curl.

To get access:

Drop a comment with your use case in 1 line — for example: "running LangChain agents", "testing streaming latency", "multi-agent with LangGraph"

I'll reply with creds. Keeping it comment-gated to avoid bot flooding during the stress test window.

What I'm measuring: p95 latency, error rates under concurrency, retry behavior, streaming reliability.

If something breaks or feels slow — drop it in the comments. That's exactly the data I need.

Will post a follow-up with full load stats once the test window closes.

(Personal project — no paid tier, no product, no affiliate links.)

reddit.com
u/NefariousnessSharp61 — 9 hours ago
▲ 3 r/LLMDevs+1 crossposts

Built an OpenAI-compatible API reverse proxy — opening for community stress testing for ~12hrs (GPT-4.1, o4-mini, TTS)

Hey Devs,

I've been building a personal, non-commercial OpenAI-compatible reverse proxy gateway that handles request routing, retry logic, token counting, and latency tracking across multiple upstream endpoints.

Before I finalize the architecture, I want to stress test it under real-world concurrent load — synthetic benchmarks don't catch the edge cases that real developer usage does.

Available models:

  • gpt-4.1 — Latest flagship, 1M context
  • gpt-4.1-mini — Fast, great for agents
  • gpt-4.1-nano — Ultra-low latency
  • gpt-4o — Multimodal capable
  • gpt-4o-mini — High throughput
  • gpt-5.2-chat — Azure-preview, limited availability
  • o4-mini — Reasoning model
  • gpt-4o-mini-tts — TTS endpoint

Works with any OpenAI-compatible client — LiteLLM, OpenWebUI, Cursor, Continue dev, or raw curl.

To get access:

Drop a comment with your use case in 1 line — for example: "running LangChain agents", "testing streaming latency", "multi-agent with LangGraph"

I'll reply with creds. Keeping it comment-gated to avoid bot flooding during the stress test window.

What I'm measuring: p95 latency, error rates under concurrency, retry behavior, streaming reliability.

If something breaks or feels slow — drop it in the comments. That's exactly the data I need.

Will post a follow-up with full load stats once the test window closes.

(Personal project — no paid tier, no product, no affiliate links.)

reddit.com
u/NefariousnessSharp61 — 9 hours ago