u/M-eisen

▲ 0

Naxios AI is a starting up LLM proxy, with access to text, image, and video models using OpenAI-compatible endpoints. 

https://preview.redd.it/idiyc1sbcdzg1.png?width=1255&format=png&auto=webp&s=269ab670dc3665c2bf46b0451e9cb3fce1668478

What do I mean with generous free tier?

The free tier is a permanent part of Naxious AI. You get 100 chat messages per day and 100,000 API tokens a day, no credit card required. Upgrade only when you need more. 

The keys are attained via Tier 0 slots on Discord, granted to you through a gateway, and act as your password to the site's dashboard. 

No, you don't need to use your credit card to start!

RP or code?

Both are allowed! you can use your credits in whatever you want to, and hook it to agents for code, there's a no log policy in place to protect your privacy!

What models do you have access to?

It operates via a tiered subscription system (Like Nanogpt does) and you get access to more rpm and tokens the higher in tiers you go

┇━───────┙𖡹┕───────━┇

Open Source / Open-Weight - free plans + all plans (Tier  0 + all)

  Kimi-K2.5 | kimi-k2.6 | MiniMax-M2.1 | MiniMax-M2.5 | MiniMax-M2.6

  minimax-m2.7 | DeepSeek-V3.2 | DeepSeek-V3.2-Speciale | DeepSeek-V4-Flash

  gemma-4-e2b | gemma-4-e4b | gemma-4-26b-a4b | gemma-4-31b

  gpt-oss-20b | gpt-oss-120b | **glm-5.1** | ministral-3b

  ollama/llama3.2 | ollama/qwen2.5:0.5b | together_ai/meta-llama/Meta-Llama-3-8B-Instruct-Lite

  Agens-sft-v7-vba-8k-15000 | OfficeCLI-Coder | kaxie-c1 | kaxie-r1-exp

  gemini-2.5-flash | gemini-2.5-pro | gemini-3-flash-preview | gemini-3-pro-preview

  ollama/qwen2.5:0.5b | qwen3.6-max | qwen3.6-plus

┇━───────┙𖡹┕───────━┇

┇━───────┙𖡹┕───────━┇

Frontier / Proprietary - Paid plans (Tier 1 and above)

  gpt-4.1 | gpt-4.1-mini | gpt-4o | gpt-4o-mini

  gpt-5 | gpt-5-chat | gpt-5-mini | gpt-5-nano

  gpt-5.1 | gpt-5.2 | gpt-5.4 | gpt-5.5

  gpt-5-codex | gpt-5.1-codex | gpt-5.2-codex | gpt-5.3-codex

  claude-haiku | claude-haiku-3.5 | claude-haiku-4.5 | claude-haiku-4.6

  claude-sonnet | claude-sonnet-3.5 | claude-sonnet-4.5 | claude-sonnet-4.6

  claude-opus-4.0 | claude-opus-4.5 | claude-opus-4.6 | claude-opus-4.7

  gemini-2.5-flash | gemini-2.5-pro | gemini-3-flash-preview | gemini-3-pro-preview

  gemini-3.1-pro-preview | gemini-pro-latest | gemini-flash-latest | gemini-flash-lite-latest

  grok-4 | gpt-image-2 | text-embedding-3-large | imagen-4.0-generate-001

┇━───────┙𖡹┕───────━┇

Can I get Free Access?

The free access is now handled on site. No, you don't need to farm invites, you don't need to share info, you don't even need a credit card for initial access.

Right now, all models are operational!

Links: 

Discord: https://discord.gg/QA8J6x22Hm

Main site: https://naxios.tech/

reddit.com
u/M-eisen — 9 days ago
▲ 32

New proxy provider with a generous free tier!

Naxios AI is a starting up LLM proxy, with access to text, image, and video models using OpenAI-compatible endpoints. 

What do I mean with generous free tier?

The free tier is a permanent part of Naxious AI. You get 100 chat messages per day and 100,000 API tokens a day, no credit card required. Upgrade only when you need more. 

The keys are attained via Tier 0 slots on Discord, granted to you through a gateway, and act as your password to the site's dashboard. 

No, you don't need to use your credit card to start!

RP or code?

Both are allowed! you can use your credits in whatever you want to, and hook it to agents for code, there's a no log policy in place to protect your privacy!

What models do you have access to?

It operates via a tiered subscription system (Like Nanogpt does) and you get access to more rpm and tokens the higher in tiers you go

┇━───────┙𖡹┕───────━┇

Open Source / Open-Weight - free plans + all plans (Tier  0 + all)

  Kimi-K2.5 | kimi-k2.6 | MiniMax-M2.1 | MiniMax-M2.5 | MiniMax-M2.6

  minimax-m2.7 | DeepSeek-V3.2 | DeepSeek-V3.2-Speciale | DeepSeek-V4-Flash

  gemma-4-e2b | gemma-4-e4b | gemma-4-26b-a4b | gemma-4-31b

  gpt-oss-20b | gpt-oss-120b | **glm-5.1** | ministral-3b

  ollama/llama3.2 | ollama/qwen2.5:0.5b | together_ai/meta-llama/Meta-Llama-3-8B-Instruct-Lite

  Agens-sft-v7-vba-8k-15000 | OfficeCLI-Coder | kaxie-c1 | kaxie-r1-exp

  gemini-2.5-flash | gemini-2.5-pro | gemini-3-flash-preview | gemini-3-pro-preview

  ollama/qwen2.5:0.5b | qwen3.6-max | qwen3.6-plus

┇━───────┙𖡹┕───────━┇

┇━───────┙𖡹┕───────━┇

Frontier / Proprietary - Paid plans (Tier 1 and above)

  gpt-4.1 | gpt-4.1-mini | gpt-4o | gpt-4o-mini

  gpt-5 | gpt-5-chat | gpt-5-mini | gpt-5-nano

  gpt-5.1 | gpt-5.2 | gpt-5.4 | gpt-5.5

  gpt-5-codex | gpt-5.1-codex | gpt-5.2-codex | gpt-5.3-codex

  claude-haiku | claude-haiku-3.5 | claude-haiku-4.5 | claude-haiku-4.6

  claude-sonnet | claude-sonnet-3.5 | claude-sonnet-4.5 | claude-sonnet-4.6

  claude-opus-4.0 | claude-opus-4.5 | claude-opus-4.6 | claude-opus-4.7

  gemini-2.5-flash | gemini-2.5-pro | gemini-3-flash-preview | gemini-3-pro-preview

  gemini-3.1-pro-preview | gemini-pro-latest | gemini-flash-latest | gemini-flash-lite-latest

  grok-4 | gpt-image-2 | text-embedding-3-large | imagen-4.0-generate-001

┇━───────┙𖡹┕───────━┇

Can I get Free Access?

The free access is now handled on site. No, you don't need to farm invites, you don't need to share info, you don't even need a credit card for initial access.

Right now, all models are operational!

Links: 

Discord: https://discord.gg/QA8J6x22Hm

Main site: https://naxios.tech/

u/M-eisen — 10 days ago
▲ 92

★・・・・・・★・・・・・・★・・・・・・★

┌─ ✦ So I’ve been messing with Lumiverse ━━,

First of all, no I wasn't paid, I'm just shilling hard work and an actual usable frontend.

Lumiverse is basically a newer self-hosted AI chat/RP frontend in the same general ecosystem as SillyTavern. The big reason I think ST users may care is that it isn’t just “ST but reskinned.” A lot of the stuff I used to bolt onto ST with extensions that are not really maintained or just QoL in general, are already there on Lumiverse from the get-go

Fully open source!

https://preview.redd.it/zy43ix2fktxg1.png?width=824&format=png&auto=webp&s=d6cab6046edb33dc7acc101dac156abb09fc905b

★・・・・・・★・・・・・・★・・・・・・★

┌─ ✦ Fun things they’ve been working on and implemented ━━,

*ૢ✧ Vector memory / embedding-based recall,

Instead of relying only on stale summaries or hoping your lorebooks still line up 400 messages later, Lumiverse has long-term memory built around embeddings. It chunks chat history, vectorizes it, and retrieves relevant older moments during generation.

That means the AI can pull back specific past moments based on meaning, not just whatever happened to still be inside context.

There’s also semantic world book activation, so lorebook entries can be found by meaning instead of only exact keyword matching.

Memory Cortex demo - Smart Vectorization based memory building demo

*ૢ✧ Lorebooks/world books,

World books are still there, but retrieval has more going on than “keyword appeared, dump entry.” Semantic search can be used, entries are deduplicated, and there are great tools around prompt assembly and activation.

Lorebook retrieval was also optimized, runs algorithmically on a sorting system that works quite a bit faster.

*ૢ✧ Dry run is built in,

You can assemble the full prompt without actually calling the model.

This is one of those features that make preset tinkering a lot more enjoyable, I am a preset maker myself and ST, even with prompt inspector, felt fairly clunky. This is great if you’re debugging a preset and trying to figure out why your character suddenly forgot their species, the plot, and basic object permanence.

Dry run lets you see what the AI is actually getting out of your prompt, including resolved macros and injected world info.

*ૢ✧ The macro system is much stronger,

They support arguments, variable shorthand, conditions, scoped variables, chat-persisted variables, global variables, math, logic, memory macros, pipeline state, council/Lumia content, Loom content, etc.

Also, macro evaluation is AST-parsed and single-pass. So, it’s more predictable, but also less forgiving if your old ST macros relied on weird post-processing behavior.

*ૢ✧ Better prompt processing / prompt structure,

Prompt blocks are very explicit. You can set role, position, depth, injection triggers, groups, ordering, enabled/disabled states, etc. Lumiverse also has a specific “Assistant/User append” prompt types that aid on long term behavior and stick out more than “at depth” prompts.

It also has context filters built into outgoing prompts, so older messages can have HTML, details blocks, or Loom tags stripped while recent messages stay untouched. That’s very nice for people running stylized HTML replies.

*ૢ✧ A lot of ST-extension-type stuff is native or better integrated,

Stuff like image generation, regex scripts, macros, push notifications, prompt viewing/dry run, world books, alternate fields, character expressions, theme customization, and sidecar/council-style tooling are treated as first-class systems that, in my experience, are a lot sturdier than their ST counterparts.

★・・・・・・★・・・・・・★・・・・・・★

┌─ ✦ Databanks / notes-style knowledge bases ━━,

Lumiverse has support for vectorized chat documents / databank-style notes, meaning you can attach larger chunks of information to a chat and have them indexed for retrieval instead of manually stuffing everything into the prompt forever.

Wiki pages, setting notes, relationship docs, faction info, timelines, canon references, custom mechanics, ability lists, city guides, campaign bibles, or whatever enormous lore creature you’ve been feeding in a folder somewhere.

*ૢ✧ Best use cases,

ִֶָ໑ Canon wiki reference material

ִֶָ໑ Character relationship notes

ִֶָ໑ Timelines and plot summaries

ִֶָ໑ Worldbuilding documents

ִֶָ໑ Ability systems / RPG mechanics

ִֶָ໑ Faction, location, and organization info

Works by either importing the information, or scraping Wiki URLs directly in the frontend.

★・・・・・・★・・・・・・★・・・・・・★

┌─ ✦ Dreamweaver ━━,

It is an LLM-assisted character creator. Not just “generate me a description and call it a day,” but a fuller card-building workflow where the AI can help produce the character, organize their fields, generate supporting lore, build associated world book material, and even help with image-gen configuration.

So instead of manually assembling:

ִֶָ໑ Description

ִֶָ໑ Personality

ִֶָ໑ Scenario

ִֶָ໑ First message

ִֶָ໑ Alternate greetings

ִֶָ໑ Example messages

ִֶָ໑ Lorebook entries

ִֶָ໑ Regexes

ִֶָ໑ Image-gen through ComfyUI, SwarmUI, or Img Gen profiles

DreamWeaver demo

★・・・・・・★・・・・・・★・・・・・・★

┌─ ✦ RP / character-card quality-of-life stuff ━━,

expressions demo

*ૢ✧ CHARX support,

Lumiverse supports .charx, which means cards can come bundled with things like avatar assets, expressions, alternate fields, and other modules.

Risu cards are importable with assets, and there’s an extension in the making to even utilize full on systems made for Risu cards. This means custom fields, UI, and actions.

*ૢ✧ In the talk about presets,

Preset assembly and usage is far more comfortable. Presets get categories, and slider macros that are user configurable without editing the prompts in their own Config UI.

Preset categories

*ૢ✧ Built-in imports,

It supports PNG cards, JSON cards, and CHARX bundles. It can import from Chub, CharacterHub, JanitorAI, and direct links to card files.

Embedded lorebooks can also be extracted and linked properly during import, which is very nice if you use lore-heavy cards.

*ૢ✧ SillyTavern migration,

Lumiverse has an interactive migration tool for importing ST characters, chats, world books, and personas.

*ૢ✧ Alternate fields,

You can make alternate descriptions, personalities, and scenarios for the same character without duplicating the whole card.

So you can have “default,” “post-timeskip,” “AU,” “romcom,” “bad ending,” whatever, and select them per chat. The active variant is what gets resolved into the prompt.

*ૢ✧ Dynamic character stuff,

Expressions are supported, avatars can be handled more cleanly, and CHARX can carry expression mappings. If you like VN-style character presentation or sprite-ish RP, this is one of those “why wasn’t this always normal” things.

Characters can carry their own regex scripts too.

*ૢ✧ Persona configs,

Your personas can carry their own lorebooks, be bound per chat, retroactively activated and switched, and you can even run a different persona for one turn and then go back to your base one.

*ૢ✧ Group chats,

Group chats are far more enjoyable to run, you can have chats with a ludicrous amount of characters, assign group chat specific scenarios, mute and force gen without issues.

*ૢ✧ Impersonation,

You can gen your own message based on full preset assembly or a contextual nudge for narrative continuation

★・・・・・・★・・・・・・★・・・・・・★

┌─ ✦ Image gen / immersion stuff ━━,

*ૢ✧ Generated scene backgrounds,

Lumiverse has image generation connections separate from normal LLM connections. It supports providers like Gemini, NanoGPT, Pollinations and NovelAI.

Generated images can be displayed as chat backgrounds, and automatic generation can trigger when the scene changes enough. You can also tune opacity and fade transitions.

There’s also an extension in the making for scene and character inline generation that will support embedded LoRas on local

★・・・・・・★・・・・・・★・・・・・・★

┌─ ✦ Mobile / interface stuff ━━,

*ૢ✧ It is actually designed with mobile in mind,

Panels collapse and slide in as drawers on mobile instead of the interface, Mobile has a priority mode where the side panel becomes a full view. Font scaling for accessibility is also supported.

Performance is also far better on mobile devices.

That alone is a QoL jump if you RP on phone/tablet or remote into your setup.

*ૢ✧ It does not look like a crime at launch,

Subjective, yes, but the default UI is much nicer out of the box. The theme system has proper controls for colors, fonts, radius, glass effects, preset themes, character-aware accents, and CSS variables.

There’s also a proper theme system where extension overrides are scoped instead of throwing500 lines into one vague CSS box and may god forgive you.

★・・・・・・★・・・・・・★・・・・・・★

┌─ ✦ Extensions / sidecar / council stuff ━━,

Lumiverse's Groupchat with the Hone, Chatroom and Spotify extensions (and song aware theming!)

This part is still early compared to ST’s giant extension graveyard/library, but there are already some very useful Lumiverse-native extensions floating around.

*ૢ✧ SimTracker,

This is the Lumiverse port of the classic SimTracker idea: the model can output structured JSON/YAML stat blocks, and the extension turns them into actual visual tracker cards.

Good for dating sim stats, RPG meters, relationship points, health, trust, desire, contempt, party state, pregnancy tracking, internal thoughts, scene state, whatever cursed little spreadsheet your RP goblin brain wants.

*ૢ✧ Spotify Controls,

This lets you control Spotify from inside Lumiverse.

Playback controls, now-playing info, search, queueing, lyrics support, album-art/theme stuff, command palette actions, and prompt-facing macros like current track / album art / lyrics.

The funniest part is that it can also expose Spotify tools for LLM/council use, so the current agent (you can create council members or download them!) can help pick music based on mood or scene vibe depending on your setup.

*ૢ✧ LumiScript,

It is basically a scripting platform for Lumiverse. You can write scripts that react to chat events, automate behavior, inject prompt context, call LLM generation, store variables, manipulate chat messages, show UI elements, register macros, and build custom interactions without directly touching Lumiverse’s internal code.

It has per-chat, global, per-character, and temporary variable scopes. It also has a built-in Monaco editor, script bindings, library scripts, event triggers, and permission controls. If Lumiverse itself doesn’t do the weird hyper-specific thing you want, LumiScript is probably where you start building it.

*ૢ✧ Mode Toggle,

This is a Lumiverse/Spindle port of the ST mode toggles concept.

You get a bunch of non-diegetic modifier modes that can be injected into the prompt, grouped by category, searched, toggled per chat, scheduled, imported/exported, and managed through a quick popover.

It has 210+ built-in modes, covering things like visual/aesthetic, genre/cinematic, social/power, temporal/physics, etc.

*ૢ✧ CharacterNudges,

This one lets characters send push notifications after you have been away from chat.

Not just a generic “come back” ping either. It uses the recent conversation, the character card, and its own nudge history to generate short in-character messages.

You can configure timing, which chat to use, how many recent messages it sees, generation settings, global defaults, and per-character configs. It is either really cute or deeply dangerous depending on how parasocial your setup already is. Probably both.

*ૢ✧ Story Weather,

Story Weather adds a draggable weather HUD and animated ambience effects to the chat.

It is not meant to be live real-world forecast data. It is for story-driven weather and scene atmosphere. The model can emit a hidden <weather-state> tag, and the extension turns that into a HUD update plus visual ambience.

It supports conditions like clear, cloudy, rain, storm, snow, and fog; palettes like dawn, day, dusk, night, storm, mist, and snow; and layers that can render behind the chat, in front of it, or both.

There is also manual lock mode if you want to override the scene yourself instead of letting the model drive it.

*ૢ✧ Prompt Viewer,

It lets you inspect the fully assembled prompt after it has been sent to the LLM, similar to ST’s prompt inspector workflow.

You can view prompts in formatted mode, raw JSON mode, or rendered readable text. It tracks prompt history per chat, shows model/generation metadata, links prompts to the message they produced, estimates tokens, separates dry runs, and lets you copy the prompt out.

Dry Run shows you the assembled pipeline, but Prompt Viewer shows you what got sent.

*ૢ✧ Hone,

Hone is an LLM-powered message refinement system.

The idea is that your main writing model can generate the message normally, then Hone can run a second pass with a shorter, more targeted context to refine it.

That can be used for prose polishing, translation, anti-slop cleanup, lore consistency, formatting passes, UI elements, or other quality-control workflows.

This is interesting because it stops trying to make one giant prompt do everything at once. Instead, the model can write first, then a smaller controlled pass can edit the output afterward.

You can use it for a bunch of different things, it has it’s own preset/prompt directives.

*ૢ✧ Live Shitposting Chatroom,

There’s an extension that gives you a floating chatroom where your council members can comment on the ongoing story. Basically a live peanut gallery / council group chat for your RP.

It can use council members, react to the active story context, preserve a chatroom history, and run on its own generation connection.

*ૢ✧ LoreRecall,

Tree-aware retrieval, per-character managed books, tree workspaces, collapsed/traversal retrieval, reranking, live retrieval feed, diagnostics, import/export snapshots, and safer book permissions.

*ૢ✧ Sidecar LLM support,

Lumiverse has council/sidecar tooling built around using a smaller/cheaper model for background analysis tasks instead of making your main RP model do everything.

So you can have one model writing and another doing support work like analysis/tool calling/deliberation, depending on your setup.

Council members are entirely customizable and there are packs ready for download in the frontend itself.

★・・・・・・★・・・・・・★・・・・・・★

┌─ ✦ Other nice stuff I’d mention ━━,

Settings modal and Operator Tab

ִֶָ໑ Push notifications exist, useful if you’re using it as a PWA or multitasking.

ִֶָ໑ Regen feedback exists, so when you regenerate you can give a reason like “too short” or “stay in character” instead of just rolling the same prompts again. This feedback is stripped from context and acts as either a system prompt or an OOC message.

ִֶָ໑ Image generation configs are stored separately from normal LLM connections.

ִֶָ໑ API keys/secrets are stored encrypted.

ִֶָ໑ It has multi-user auth and per-user data isolation, which is nice if you run an instance for more than one person.

ִֶָ໑ The launcher can do first-run setup, build/start, and the frontend has an Operator tab which serves as an update install (and trust me, the dev is a freak and updates multiple times per day) or a remote access whitelist on-site so you can run custom domains or services like Tailscale.

ִֶָ໑ The docs are already pretty good. Like, really nice. And installs come with a dev doc so people who develop extensions can reference them easily.

ִֶָ໑ There’s LumiHub support for installing characters/world books, including Chub imports through LumiHub.

┌─ ✦ TLDR ━━,

Lumiverse is a self-hosted AI chat/RP frontend that feels very interesting for ST power users.

Big draws: vector memory, semantic lorebook retrieval, stronger macro/preset processing, dry run, better mobile layout, prettier UI, dynamic image backgrounds, CHARX support, ST migration, sidecar/council tooling, and a more modern (and easy to develop) extension system.

★・・・・・・★・・・・・・★・・・・・・★

┌─ ✦ Links ━━,
Discord server:
https://discord.gg/28rBWVFfCu

Lumiverse repo:
https://github.com/prolix-oc/Lumiverse

Lumiverse guides:
https://lumiverse.chat/guides/

reddit.com
u/M-eisen — 16 days ago