r/FluxAI

Flux.2 Klein prompt help - cannot get rid of studio camera flash
▲ 3 r/FluxAI+1 crossposts

Flux.2 Klein prompt help - cannot get rid of studio camera flash

Would appreciate any help or insights. I'm trying to have the light in this image be 99% red post-sunset light. But there is the overwhelming daylight / flash source to the upper right.

I do see those little lamps mounted on the eaves but I'm not sure they are at fault. I have tried eliminating just about everything in this prompt to get rid of this. I've tried calling the structures "a hut," "a gazebo," "a pavilion" etc. to no avail. I added a photo below of generally what I'm aiming for.

  • Model: Flux.2 4B Klein Distilled
  • Steps: 8
  • Seed: 841190067
  • Prompt:

​

It is night. Natural light only. The scene is lit only by the red sky behind the pavilion. Dramatic lighting from the sunset. The sky is dark red. Subject: A wooden chaise longue. The chaise has a soft beige woven cover with finely detailed bali designs in a range of orange, brown, and black threads.  The chaise has a large soft emerald green embroidered velvet pillow. A wet beach towel with the name of the resort is carelessly draped over the back of the chaise. Environment: The setting is a balinese pavilion on the island of Bali. The pavilion is open-air and made of dark extensively weathered teak wood with many detailed thin wood slats. The pavilion is only raised two steps above the ground-floor garden. Leafy vines grow along most slats. The pavilion is in the center of a lush emerald-green tropical garden with many small and medium water features, stones, and flowering tropical plants. Paths made of many dark rounded stones and some larger circulare slate stones lead to other areas and to the private polynesian beach, which is visible beyond the garden. The architecture and design is modern but native to the culture of the island. Camera: The camera is positioned so that we are looking directly at the chaise from the front. We can see the whole length of the chaise. The camera stands back from it. Film style: Shot on Canon 5D Mark IV, RAW photography, 8K. 28mm lens at f/5.6

From https://pngtree.com/freebackground/a-beauty-of-sunset-on-tanah-lot-temple-in-bali_15494252.html

reddit.com
u/OrdinaryAward4498 — 5 hours ago
▲ 5 r/FluxAI

Whats the best photorealistic Flux model for local use right now?

I'm new to local AI world and I have a pretty beefy PC, so I want the best of the best.

reddit.com
u/Interesting_Air3283 — 9 hours ago
▲ 3 r/FluxAI

Having a problem using AI-Toolkit to train a lora

I have AI-Toolkit installed inside Stability Matrix. When I open it, everything looks fine. I set up how I want the training, but when I click to start training, I get "No Checkpoints Available". I've entered and saved my Hugging Face API, and the models dropdown points to the default Hugging Face page for Flux1.dev. Alternatively, I put a copy of the model in /AI-Toolkit/Models/Checkpoints (this is what CoPilot told me to do and I had to create these folders) and then pointed AI-Toolkit to the location. Neither of these work for me.

Unfortunately, I don't feel competent enough in technical matters to attempt to use ComfyUI, which ironically might make this process easier. pinokio does not work on this computer because its installations don't take into account differences in the 50xx Nvidia GPUs. I'm very close to just giving up. I have literally been trying to get different lora training programs to work for a full year now, and I have yet to train a single lora, so any help you can provide will be greatly appreciated. If you need more info, just let me know. I wasn't sure exactly what to provide. My GPU is a 5070 ti.

reddit.com
u/ZedClampet — 11 hours ago
🔥 Hot ▲ 532 r/FluxAI+2 crossposts

Coming up Tomorrow! Flux2Klein Identity transfer

I successfully found a way to transfer the character from the reference latent into the generation process without losing features; meaning I give full freedom to flux2klein to generate whatever it wants. My previous approach was a bit rigid as I scaled the k/v layers, which worked but was tough to move at times. Instead, this new approach uses attention output steering. The reference latent stays in the image stream, but after every attention layer, the model finds where the generation's features are similar to the reference and pulls them closer. Because it is similarity-gated, features that are completely different like new backgrounds or different poses are left entirely alone. This lets us lock in the identity of the full character deep in the blocks while allowing the model to change poses and follow the prompt without restraints. I am preparing the documentation and preparing the release!

Examples are in order, first vanilla and second is with node

u/Capitan01R- — 2 days ago
▲ 47 r/FluxAI+2 crossposts

ZPix, an open-source local image generator, now supports image editing via FLUX.2 [klein] 4B, has a bigger output gallery and a prompts history.

To add a reference image, just drag an image directly from output gallery or any location. On my RTX 3070M (8GB VRAM), once warmed, ZPix takes around 10s to generate a 720p image based on a 720p reference.

Output images are now automatically saved in your Pictures folder, ZPix subfolder, one sub-subfolder per LoRA.

Prompts are stored in a local database file, they are instantly searchable and selectable. You can also retrieve a prompt by dropping in prompt zone an image generated by ZPix, including from output gallery.

FLUX.2 [klein] 4B LoRAs are supported.

More aspect ratios are available.

FlashAttention is now used instead of SageAttention for better compatibility.

Download at: https://github.com/SamuelTallet/ZPix

As always, your feedback is welcome!

u/SamuelTallet — 1 day ago
▲ 18 r/FluxAI+13 crossposts

Tera Byte - Never Gonna Last Official Video

Tera Byte - Never Gonna Last Official Video

u/TeraByteOfficial — 2 days ago
▲ 1 r/FluxAI+1 crossposts

Beginner Needing T2I and I2I Workflow Help with Flux Klein Model on Colab

Hi everyone,

I’m new to ComfyUI. Could someone please share a workflow for text-to-image (T2I) and image-to-image (I2I) using the Flux Klein model and Guff?

I’m running ComfyUI on Google Colab, so I can’t load heavier files. I’ve been frustrated for the past couple of days due to coding issues and errors, and most of my time ends up getting wasted on troubleshooting rather than actually creating.

Any help or shared workflow would be greatly appreciated. Thanks in advance!

reddit.com
u/comfy_1922 — 17 hours ago
▲ 7 r/FluxAI

AI for video face swap?

Tried going from photo face swaps to video recently and didnt expect the gap to be this big. With images, results are almost perfect now but with video keeping the same face across frames with motion, angles and lighting is way harder than it looks

reddit.com
u/snckr_bar — 2 days ago
▲ 3 r/FluxAI

VAE and text encoder for FLUX.2-klein-4B

Hey! I have been using FLUX.2-klein-4B on my comfy setup lately with qwen_3_4b_fp4_mixed.safetensors and flux2-vae.
I was wondering if inference providers like fal, replicate etc, use the same or different.

reddit.com
u/Sensitive_Cat6439 — 15 hours ago
🔥 Hot ▲ 229 r/FluxAI+1 crossposts

I built a free 90-node All-in-One FLUX.2 Klein 9B ComfyUI workflow — Face Swap, Inpainting, Auto-Masking, NAG, Refiner, Upscaler — runs on 8GB VRAM

UPDATED TO 2.1 tutorial post
Hey everyone,

I've been working on this for a while and wanted to share it with the community. This is a 6-in-1 ComfyUI workflow for FLUX.2 Klein 9B that handles everything in a single workspace — no more switching between different workflow files.

What's inside:

  • 🎨 Text → Image — standard txt2img with optimized settings
  • 🖼️ Single-Reference KV Edit — load an image + describe what to change, the model preserves everything else
  • 🚀 Face + Pose Swap — extract a face from one image, a pose from another, combine them realistically
  • 🎭 Inpainting — manual mask OR Florence2 AI auto-masking (describe what to mask in text)
  • 🔀 Image Merge — blend two images with adjustable ratio
  • Refiner — enhance any image with detail injection, lighting correction, skin texture improvement

Technical features:

  • 🧭 NAG (Normalized Attention Guidance) — restores negative prompting that normal CFG breaks in distilled Flux models
  • 🤖 Florence2 auto-masking — type "Segment the shirt" and it generates a pixel-perfect mask automatically
  • ⬆️ 4x UltraSharp upscaler built in
  • 🔷 All VAE decodes are Tiled — prevents OOM on 8GB VRAM
  • 🔗 2-slot LoRA chain — enhancer LoRA always last, add your own LoRAs in the first slot

Hardware tested on: RTX 4060 Mobile (8GB VRAM), 16GB RAM, i7-13620H. Works with FP8 or GGUF Q4 models.

update 2.1: added groupe bypasser, notes for new people to comphyui.
Each pipeline is in its own color-coded group. Only the Refiner is active by default — right-click any group to enable/disable it or use groupe bypassers. The workflow includes built-in guide notes with download links and prompting tips.

Free download on Civitai: https://civitai.com/models/2543188?modelVersionId=2860464

Includes a full guide with all model download links, prompting tips, and troubleshooting. Let me know if you run into any issues — happy to help.

How to Use

  1. Load the JSON in ComfyUI
  2. use comphyui manager to install any missing node. ( critical step )
  3. Only the Refiner is active by default — everything else is bypassed
  4. To activate a pipeline: right-click its group header → Set Nodes Mode → Always Execute
  5. To deactivate: right-click → Set Nodes Mode → Bypass ( or bypass groupe nodes )
  6. Read the built-in Note nodes for prompting tips and download links
u/official_geoahmed — 6 days ago
🔥 Hot ▲ 525 r/FluxAI+2 crossposts

[Release] LongExposureFX COMP | An experimental temporal ghosting / long-exposure toolkit for TouchDesigner

An experimental temporal ghosting / long-exposure toolkit for TouchDesigner, built for turning prerecorded and real-time footage into smeared, split-exposure, echo-like motion.

The system layers delayed frames, masks the active subject region, and adds optional feedback persistence to generate distorted portrait, face, and full-body trails that sit somewhere between long exposure, temporal rupture, and spectral motion blur.

This release also includes:

a custom FLUX-2 LoRA trained on experimental photography [the one used in this demonstration]
the pertinent ComfyUI workflow for FLUX-2.dev + LoRA text-to-image generation

Available now through my Patreon profile.

Both music and visuals by myself, deeply inspired by the recent BoC-related events.

u/uisato — 8 days ago
▲ 39 r/FluxAI+1 crossposts

Tested the new FLUX.2 Small Decoder — faster and lower VRAM, with basically no quality hit

Black Forest Labs just released the FLUX.2 Small Decoder, which is supposed to be faster, use less VRAM, and keep image quality almost unchanged. I tested it on my end and that pretty much checks out. I put the full comparison and results in the video if anyone wants to see the actual workflow and timings.

FLUX.2-small-decoder

https://huggingface.co/black-forest-labs/FLUX.2-small-decoder

youtu.be
u/EmilyRendered — 6 days ago
▲ 0 r/FluxAI

Stack For AI Avatar

Hello everyone.

I've been trying to generate an AI avatar to be used in YouTube videos. However, I couldn't get consistent images.

What I've done so far:
- I used Flux Dev on Runpod. Generated an image. Found my character and fixed the seed.

- Everytime the scene and the light changed, even with the same seed, face changed a little bit.

- I got help with my prompts from ChatGPT and Gemini. Results didn't change.

- One way or another, I gathered similar face, different angles and different shots. 50 images in total. Similar to my original image.

- Trained a Lora by captioning images with Joycaption & ai-toolkit. Some sample images were acceptable, others are not.

- Used the Lora over Flux Dev with fixed and random seeds. No luck. My avatar works well in dark light but when it comes to daylight and different poses, facial expressions etc, I completely lose the original face.

- Tried everything above with a few realism checkpoints for Flux. Same result.

- I've added Flux Dev, PulID. With or without the Lora, I either kept losing the face or getting anatomically wrong results.

What I want to achieve is:

- Keep the same character, including face, in different scenes and shots

- If I could get consistent images, then I'll move forward with the video generation. I need that base images for different videos like sports events, concerts, public transportation etc.

What am I missing? What would be your suggested stack?

reddit.com
u/buyukadam — 4 days ago
▲ 5 r/FluxAI+4 crossposts

I built an agent that runs a creative studio on an infinite canvas. it's working way too well

I built a creative director agent and an infinite canvas and now I'm a little freaked out

Gave it a one-line brief and a product photo. that's it. it opened the canvas, planned the shot, built a multi-layer node workflow, chained the models, wrote the prompts and delivered a finished campaign. I didn't click anything. I didn't write a single prompt after the brief.

The thing that broke my brain is that it made taste decisions. it chose the lighting. It chose the angle. it chose the environment. and they were good. not "ai good." actually good.

I think this can help teams a lot in creating visuals pretty fast, what do you guys think?

u/Luffy-writes — 3 days ago
▲ 11 r/FluxAI

Inpaint, outpaint and erase objects with Flux2.Klein

Took inspiration from some comfyui workflows and made a more user-friendly approach. A bit of hacking and it might struggle in some cases, but in general it works quite well.

more details here

u/Living_Gap_4753 — 8 days ago
▲ 8 r/FluxAI

Do We Still Need ControlNet with Flux 2?

It looks like in img2img inference, adding “keep depth” in Flux 2 9B can produce results that are somewhat similar to using ControlNet.

Does this mean we can drop ControlNet entirely when working with Flux 2 models?

For those of you doing tasks that require strong structural consistency (depth, line art, pose, etc.), what are you currently using?

Curious to hear real-world experiences.

reddit.com
u/Due_Caramel3612 — 9 days ago
▲ 0 r/FluxAI

1000 unique clones! And News

The GIMP plugin is good shit.

I'm very proud to say that the debug marathon went great. By now, Spellcaster has become what I use to operate ComfyUI over 90% of the time. Generating layers (including transparent layers) over an image within GIMP is by far the most efficient and effective way to alter any image through AI. Blending generations is a walk in the part (right click add alpha channel, then delete at whatever opacity that suits your purposes).

Based on my testing, which could be misleading, Spellcaster now has a very robust, single-source-of-truth for all generative methods, making it far more modular and easier to modify moving forward. The single source of truth is located within a custom node of ComfyUI, which makes perfect sense.

Spellcaster now comes with a stand-alone interface called the Wizard Guild. Although it needs more work and debugging, and will never have the precision of advanced and fine image editing tools that GIMP possesses, it remains a very exciting, fun, beginners-oriented, chat-based interface for Spellcaster. This also makes Spellcaster natively equipped to work with chatbot applications, like Silly Tavern.

I'm working on an entirely chat-based and fully scaffolded installation and management of Spellcaster for a couple of reasons:

  1. very small LLMs can now be loaded and unloaded between generations, providing novice users with guidance as they use generative AI for the first time;

  2. properly scaffolded, the LLM will be able to guide the user through the installation process step-by-step, from pressing install to generating their first images, and initiate them to far more complex generative methods - for instance, the LLM will offer to produce multi-model generative workflows with guided LoRA setup and calibration, where each workflow is embodied by a Wizard, automatically named and populated with its own scaffolded and robust generative methods.

  3. through natural speech or through a highly simplified LLM scaffold editor that doubles as a workflow manager (check it out in a submenu of Wizard Guild called the travelling Wizard) the user will be able to build highly refined generative workflows;

  4. Spellcaster's own core Wizards / chatbot characters oversee the installation of Spellcaster, either during initial installation or later-on upon request. In order to avoid frustration and dysfunctional setups, the Wizards both populare and strip Spellcaster (and all of its associated front ends, like the GIMP or Darktable plugins, or the Silly Tavern enhancing characters) from non-working generative methods. If it is initiated through Spellcaster, it either works or the user won't see it.

  5. Wizard Guild also comes with a Signal bridge, allowing users to communicate with their local ComfyUI server remotely, through the Signal app

As always, thanks for your feedback and bug reports. Please report all bugs here or directly in the GitHub repos. If you want to lend a hand and help me out with dev, LMK

https://github.com/laboratoiresonore/spellcaster

https://github.com/laboratoiresonore/ComfyUI-Spellcaster

u/ActionInUganda — 8 days ago
▲ 0 r/FluxAI

Misa - gothic lolita

This is Misa, part of a twin sister character project I'm developing with Flux + PuLID.

The twin setup is intentional - by using the same face reference for both characters through PuLID, I can maintain identical "twin" faces while differentiating them completely through hair, styling, and aesthetic direction. PuLID handles face consistency way better than I expected across very different outfit contexts.

Misa is the gothic lolita twin (blonde twin side-up, dark aesthetic, playful personality). Her sister Asuka will be the mature redhead I'll share next.

Happy to discuss the PuLID approach if anyone's interested.

u/misa-asuka-fans — 7 days ago