Model Description
Mugen is a continuation of our SDXL to Flux 2 VAE conversion, renamed to signify a substantial divergence from the original NoobAI models.
It has been trained for 7 additional epochs, totaling under 8000$ for a full latent space conversion, while preserving and improving upon model anime knowledge.
In particular, we have paid attention to characters in this iteration, and developed in-house approach for benchmarking their performance, about which you can read below.
Overall, model performs particularly well with textures and patterns that were previously simply impossible due to SDXL VAE. We prioritized keeping our training as standard-friendly as possible, so local community can easily train on it like on a new Base Model, which it practically is.
We provide 4 models:
Mugen: A base model.
Mugen - Aesthetic: Slightly tuned on a limited dataset model for better quality output.
Mugen - Aesthetic - Anzhc/Selph: Further tune on opinionated dataset selection.
- Developed by: Cabal Research (Bluvoll, Anzhc)
- Funded by: Community
- License: fair-ai-public-license-1.0-sd
- Resumed from: NoobAI Flux2 VAE v0.3
Character Knowledge Benchmark
This benchmark measures character similarity across 1815 characters in this iteration of it. For convenience, we've gathered few major categories: gachas and vtubers.
We utilize reference(non-generated) set of images, and measure character features against ai-generated data - this is the similarity score. Our custom in-house model for character discrimination trained on ~1.2kk images is used. Results are sorted indiscriminately, by score, treat it as general character knowledge index, not as any specific characters in particular. Same point on graphs might, or might not be corresponding to the same character.
Due to compute constraint, we selected only single model to compare against - not yet released latest version of Chenkin model, which currently is the most trained SDXL-based anime model.
Future benchmark iterations might include different arches, more models and more characters.
Bias and Limitations
General data biases from Danbooru might apply.
Flux 2 VAE seem to have brown bias overall, which can be alleviated by adding sepia or brown theme to negative.
Model Output Examples
You can download most of those images from Here for reference.
Recommendations
Characters
While in benchmark we test characters purely with their own trigger with no helper tags, it is advised to utilize series/game for better adherence. Characters that might appear not working initially could start working with appearance tags.
Inference
Comfy
We will provide a Node, and hope it will be adapted natively in main repo eventually:
https://github.com/Anzhc/SDXL-Flux2VAE-ComfyUI-Node
Just install it, and it will patch the model config, no node changes required.
Apparently works in SwarmUI as is.
Same as your normal inference, but with addition of SD3 sampling node, as this model is Flow-based.
Recommended Parameters:
Sampler: Euler A, Euler, DPM++ SDE, etc.
Steps: 20-28
CFG: 4-7
Shift: 8-12
Schedule: Normal/Simple/SGM Uniform
Positive Quality Tags: masterpiece, best quality
Negative Tags: worst quality, normal quality, bad anatomy, sepia
Alternative Extended Negative: (worst quality:1.1), normal quality, (bad anatomy:1.1), (blurry:1.1), watermark, sepia, (adversarial noise:1.1), jpeg artifacts
(Some of our testers pointed out that they prefer longer negative)
A1111 WebUI
Recommended WebUI: ReForge - has native support for Flow models, and we've PR'd our native support for Flux2vae-based SDXL modification.
How to use in ReForge:
Support for RF in ReForge is being implemented through a built-in extension:
IMPORTANT
Set your preview method to this, if you use it.
Flux2VAE does not currently have an appropriate high quality preview method, please use Approx Cheap option, which would allow you to see simple PCA projection(ReForge).
Recommended Parameters:
Sampler: Euler A Comfy RF, Euler A2, Euler, DPM++ SDE Comfy, etc. ALL VARIANTS MUST BE RF OR COMFY, IF AVAILABLE. In ComfyUI routing is automatic, but not in the case of WebUI.
Steps: 20-28
CFG: 4-7(or 7-15, if it appears to be weak/bugged)
Shift: 8-12
Schedule: Normal/Simple/SGM Uniform
Positive Quality Tags: masterpiece, best quality
Negative Tags: worst quality, normal quality, bad anatomy, sepia
Alternative Extended Negative: (worst quality:1.1), normal quality, (bad anatomy:1.1), (blurry:1.1), watermark, sepia, (adversarial noise:1.1), jpeg artifacts
(Some of our testers pointed out that they prefer longer negative)
ADETAILER FIX FOR RF: By default, Adetailer discards Advanced Model Sampling extension, which breaks RF. You need to add AMS to this part of settings:
Add: advanced_model_sampling_script,advanced_model_sampling_script_backported to there.
If that does not work, go into adetailer extension, find args.py, open it, replace _builtin_scripts like this:
Here is a copypaste for easy copy:
_builtin_script = (
"advanced_model_sampling_script",
"advanced_model_sampling_script_backported",
"hypertile_script",
"soft_inpainting",
)
Or use my fork of Adetailer - https://github.com/Anzhc/aadetailer-reforge
LoRA Training
You can directly reference config with all parameters: Download
Hardware
Model was trained on cloud 8xH100 node.
Software
Custom fork of SD-Scripts(maintained by Bluvoll)
Acknowledgements
Sponsors
To a special supporter who singlehandidly sponsored whole run and preferred to stay anonymous
Testers
- ComradeAnanas
- Daruda
- Drac
- itterative
- kagame
- Remix
- Ryusho
- edf
- Epic
- Ly
- Panchovix
- Rakosz
- Sab
- Silvelter
- Talan
- Void
- Why ping
Support
If you wish to support our continuous effort of making waifus 0.2% better, you can do it here:
https://ko-fi.com/bluvoll (Blu, donate here to support training)
https://ko-fi.com/anzhc (Anzhc, non-training, just survival)
BTC: 37fLcfxX5ewhJXnb3T9Qzu9jiSLjVtoUJX
ETH: 0xfdF54655796bf2F5bf75192AeB562F8656c1C39E
Send DM to Blu if you want to donate on another network.
- Downloads last month
- 350
Model tree for CabalResearch/Mugen
Base model
Laxhar/noobai-XL-Vpred-0.75










