Thread #8834450
File: OP19022026.png (3 MB)
3 MB PNG
Kissu Edition
Previous Thread: >>>>8827652
>LOCAL UI
reForge: https://github.com/Panchovix/stable-diffusion-webui-reForge | Stable commit: 6964ceb
Comfy: https://github.com/comfyanonymous/ComfyUI | https://comfyanonymous.github.io/Co mfyUI_examples
>NOOBAI-XL
https://civitai.com/models/833294/noobai-xl-nai-xl
https://huggingface.co/Laxhar
>NOOBAI SHITMIXES
102d custom: https://civitai.com/models/1201815?modelVersionId=1491533
291h: https://civitai.com/models/1301670/291h
>RESOURCES
4chanX Catbox/NAI prompt userscript: https://rentry.org/hdgcb
Easy booru tag export userscript: https://github.com/Takenoko3333/Danbooru-Tags-Sort-Exporter
Tags: https://danbooru.donmai.us/wiki_pages/tag_groups | https://danbooru.donmai.us/related_ tag | https://tagexplorer.github.io/#/
Inpaint: https://files.catbox.moe/fbzsxb.jpg| https://huggingface.co/Wenaka/NoobA I_XL_Inpainting_ControlNet_Full
Upscalers: https://openmodeldb.info
Booru: https://aibooru.online
ControlNet: https://rentry.org/dummycontrolnet (OLD) | https://civitai.com/models/136070 (look at the green links in the description)
>TRAINING
Guide: https://rentry.org/yahbgtr (WIP)
Anon's scripts: https://mega.nz/folder/VxYFhAYb#FQZn8iz_SxWV3x1BBaJGbw
Trainers: https://github.com/derrian-distro/LoRA_Easy_Training_Scripts | https://github.com/bmaltais/kohya_s s | https://github.com/Nerogar/OneTrain er
>NEWS
https://huggingface.co/circlestone-labs/Anima
https://huggingface.co/ChenkinRF/ChenkinNoob-XL-v0.2-Rectified-Flow
OP Template/Logo: https://rentry.org/hggop/edit | https://files.catbox.moe/om5a99.png
1221 RepliesView Thread
>>
>>
File: 00755.jpg (871.8 KB)
871.8 KB JPG
>Why are you doing [thing]
Why not, I am just that bored so anything goes I guess
>Why is so [thing]
Because it's me
>Are you going to do [thing]
Maybe I will maybe I won't
>Can you [thing] more [thing]
No, usually I determine the degree of [thing] while working on the gen with no plans on adding more of [thing]
>>
>>
>>
>>
>>
>prompt @koku on anima
>generic 2000s slop completely unrecognizable as him
>add game cg
>still slop
>add some of the more recent titles he worked on with a fairly consistent style that's more distinct than his earlier works
>STILL slop
What the fuck is this model's major malfunction? Fucking WAI has more of an idea what the guy's style looks like.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>8834507
Kobuichi and Muririn (they almost always draw together so it's one of the few cases where artist mixing works, just poorly), Shida Kazuhiro, Izumi Tsubasu, Amazuyu Tatsuki, Harada Takehito, Pochi (pochi-goya), Ishikei, Yaegashi Nan. All of these either look nothing like themselves or just terirbly bad.
>>
>>
>>
>>8834523
No, it's way too bothersome.
>need to collect a dataset
>need to curate a dataset
>need to label a dataset
>need to use the right training settings (what settings are "right"? fuck knows!)
>need to wait however long it takes
>need to compare the different checkpoints to see which one is the best
Meanwhile if the arch wasn't stuck in 2023 and just had an edit mode, I could upload one singular image, say "do my prompt in this style" and be on my merry way.
>>
>>8834520
It's a theory that homosexual and other degenerate sexual behaviors happen because of mind altering worms that use these behaviors as a way to propagate themselves
It's been shown in multiple studies that some worms will make people more promiscuous, much like the flu will make you more sociable before visible symptoms show up
>>
>>
>>
>>8834526
>Meanwhile if the arch wasn't stuck in 2023 and just had an edit mode, I could upload one singular image, say "do my prompt in this style" and be on my merry way.
there is not a single model that can actually accomplish this. No, NBP, grok and dall4 cannot actually do this, they're all too corposlopped. Even nai struggles to do this consistently because vibetransfer is garbage and so is their style reproduction/consistency.
>>
>>
>>
>>
>>
>>8834537
So there are three ways to censor a model. You remove all loli, bestiality, etc from the dataset. Or you make it so everything captioned with these produces a blank screen or a "safety" image, like what Auraflow did unintentionally. Or the SAAS way, you run recognition models on the prompt text and on the output image, then block the user from seeing the output if it's tagged with "bad" stuff.
Third doesn't apply to local. Second I've never heard of anyone doing on purpose, it's unreliable and probably has negative side effects. And in the first case, the encoder still produces an output when encountering something it doesn't recognize, and you can align the Unet to whatever that output is.
He mentioned training with many different caption styles, including pure tags, pure nl, tags first, nl first and more. I don't remember the exact wording, check the HF readme.
>>
>>8834522
I tested only a few of these in according to my own tastes
Harada Takehito worked great, Ishikei worked, although it looks more like his older stuff than his newer stuff, Pochi and Yaegashi Nan gave me generic slop
>>
File: 20260219143505-221129379.png (1.3 MB)
1.3 MB PNG
>>
>>
>>
>>
>>
>>8834556
Yeah I've noticed that as well. I managed to get the old fate artstyle by using date tags and downright ABUSING the weight but I never managed to get the newer styles
>>8834558
Autotagger? Here: https://github.com/picobyte/stable-diffusion-webui-wd14-tagger
Be advised though, I couldn't get it to work with Forge Neo.
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 2026-02-19180338_stealthmeta.png (2.2 MB)
2.2 MB PNG
>>
>>8834435
I've been keeping my noob mixes down to 3 artists (honestly, most of my mixes with 4+/5+ could have been shaved to 3 and barely affected the gen but cope) and been having success using 1 at full strength and scheduling the other 2 at 0.55
>>8834565
>Be advised though, I couldn't get it to work with Forge Neo.
Same. Crashed the entire thing for me so I personally switched over to >>8834574. I miss just having a tab to auto tag for me but I kinda like the ui on this program so I'll manage.
>>
>>8834552
>>8834556
I've personally swapped from 'newest' to 'year 2025'. Another thing to note, experiment with samplers/schedulers. I've had styles look off with one that look almost artist perfect with another. Really depends on the artist too. Softer, more shading, not so flat works well on Euler A. The opposite does better with dpm++ 2m sde or er sde.
>>
>>
File: 00257-196365535.png (1.1 MB)
1.1 MB PNG
>>
>>
>>
>>
>>
File: 00221-3541027986.jpg (982.7 KB)
982.7 KB JPG
>>8834786
Aside from inpainting, not really, I assume that you are already using all the tags you should be using
>>
>>
>>
>>8834790
I use it for gens where the image isn't from a booru, you know already tagged up, so maybe something I run into a trailer, twitter, or artist concepts of old games, etc. Run it there and have it do all the work for me and then just plug it into a ui to gen with.
I'm sure most use it for compiling datasets for loras.
>>
>>8834786
Are you talking anima or illu/noob? The latter clearly is a vae limitation. You're stretching that shit thin the more you pile into the image. Hence why when you have /hgg/'s bread n butter, 1girl solo looking at viewer close-up, it'll effortlessly look crisp. As for anima, it does a lot better due to the larger vae but you'll still have an issue here and there which you can chalk up to the preview model and the fact it's only been trained on 512x512. Or at least that's the going narrative.
>>
how do you find inspiration for poses and then figuring out how to type it out in a way that anima understands?
I'm struggling with coming up with creative stuff so I thought I'd try looking at some references but I have no idea how I'd even begin to explain how these pose are set up
I love natural language but I think it's harder to gauge what works and doesn't, unlike tags
>>
>>
>>8834845
>I love natural language but I think it's harder to gauge what works and doesn't, unlike tags
A little yeah. With noob, I'd throw in a shit ton of tags that I think would accomplish the image I saw in my mind's eye and when it worked, great. When it didn't, I'd mald and keep at it until I got something close or something completely unique.
Anima, on the other hand, you can do the same but the nl support gives you even more control so any failings are truly a skill issue or lazyness. Describe to it like the lobotomized model it literally is what you want in one sentence, throwing in actual tags in the process, and you should get what you want mo sweat. I think camera angles is the most difficult thing about it since that's always influenced by the resolution.
>>8834847
Nope. Those normally make it worse, especially with inpaint. The vae is only 4-channel so there's no getting around that limitation.
>>
>>
>>
>>
>>
>>
>>
>>
>>8834857
Is voice cloning, particularly music lyrics replacement, still where it was in 2023? I looked it up a bit and while normal voice cloning/TTS did make some progress with stuff like Qwen3-TTS, I couldn't find anything beyond RVC for lyrics replacement.
>>
>>8834875
there's a thread on /g/ >>>/g/108164777
the model came out a few days ago. The UI and training portion of all of this is basically vibecoded trash.
But it's in general really lightweight. A 12gb 3060 can generate two 90 second tracks in like 35 seconds(read: you can generate faster than you can listen to it on a $200 (used) GPU). It can do lyrics if you want, but captions are all LLM generated wordy garbage, lyric sheets require really specific syntax and it's all NL garbage.
The trainers are bad. Really fucking bad. It recognizes that there's a u-net and a TE(and the ability to effectively train unet only) but it only gives you the option to variable a single, nondescript LR that I have no idea if it's actually for said u-net/TE or if it's something retarded like the GLOBAL LR that gets set for the optimizer.
>>8834877
honestly haven't played with it that much in that respect. Right now I'm just dipping in for the experience and to learn how it all functions. Someone posted something like a lyric replacement in the /g/ thread but it still sounded pretty TTS robotic tier.
>>
>>
>>
>>
>>
>>
>>8834882
there's a gradio UI, you can completely ignore cumsoft. The better trainer is commandline only.
>>8834883
it gets worse.
here's a very brief list of my complaints >>>/g/108194641
oh also it has two different architectures for training. bogstandard LoRa and lyco-type! But it's. Uh. um. It's LoKr.
>>8834884
just be like me and completely ignore anima until the full res model is out. Maybe by then it'll have significantly improved training support, too.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 1763391141708900.png (785.1 KB)
785.1 KB PNG
>>
File: 20260220060730-511072677_cleanup.png (1.2 MB)
1.2 MB PNG
>>
File: 1771592838.jpg (1.3 MB)
1.3 MB JPG
10 minutes on a gtx 970, being poor is no excuse
>>
>>
>>8834945
i did some testing a while ago and imo it should be pos/neg
masterpiece,
(ye-pop deviantart:1.2), worst quality,
personally i also add (@hero neisan:0.2) to negs because it's intensely sepia and fat two things i do not want
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: no comma.png (772.7 KB)
772.7 KB PNG
>>8834965
Fucking catbox.
>>
>>
File: no comma.png (798.5 KB)
798.5 KB PNG
>>8834967
I'm retarded, ignore the filename in the previous post, it IS with a comma. This is without.
>>
>>
>>
>>
>>
File: ComfyUI_temp_sbmiq_00033_.png (1.3 MB)
1.3 MB PNG
>>
>>
>>
>>
>>
>>
>>
I'm finding it pretty hard to control breast size on Anima
I would love to see an example of how body shapes are tagged in NL because I feel like with NL it should be easier to control the shape and size of breasts
If they've even tagged breast sizes, because I seriously hope it's not just tagged "breasts"
>>
>>
>>
>>
>>
>>
>>
File: 1771604395.jpg (441.3 KB)
441.3 KB JPG
>>8834974
old and busted but it better last me another decade because i'm not buying a new gpu any time soon
it can still spit out something small in like 3 minutes, though
>>
>>8835007
You don't need it. You can just say "the girl on the left / the girl on the right" or any other image composition thing you want. Camera angles are kind of annoying but relative direction works completely fine.
>>
>>8835029
It works the same as normal. flat chest > small breasts > medium breasts > large breasts > huge breasts > gigantic breasts > hyper breasts
And you can use all the modifiers like sagging, bouncing, hanging, breasts apart, breasts together. It's also based on artist, obviously, and just use negs if you hit some weird place in the latent space.
skill issue
>>
>>
File: 1758298585627276.png (924.7 KB)
924.7 KB PNG
>>
>>
>>
>>
File: 20260220115325-221129383.png (1.2 MB)
1.2 MB PNG
>>
>>
>>
>>
>>
>>
>>
>>
>>8835110
LoRAs have their entire file size budget dedicated wholly to just one style/concept; checkpoints have to somehow fit the ENTIERTY of Danbooru inside like 4 gigs, which usually means that any tag without at least a good several dozen examples will simply not be learned. And something tells me that dividing 4 gigs by the number of unique tags on Danbooru will land you at a FAR smaller size than a LoRA's typical file size of anywhere between a few dozen and several hundred megs.
>>
>>
>>
>>
>>8835118
4 182 218 360 bytes (size of the Anima checkpoint) / 313 235 (number of lines in the 2025/09/19 Danbooru .csv) lands you at just 13.35 kilobytes per tag on average. That's the ballpark of embedding (remember those?) file sizes, not LoRAs. If we make the fairly safe assumption that the vast majority of the tags are below the knowledge threshold, meaning that a small fraction of the top tags takes the majority of the file size, that leaves even less space, and much less at that, than 13.35 kilobytes for any one of the far more numerous smaller tags.
>>
>>
>>
>>
>>
File: {60B48FEA-3A0C-41A1-B3F5-876A195880B0}.png (45.2 KB)
45.2 KB PNG
I think my finetune config might be chopped
>>
I am trying to prompt two views of the same scene as I want a view from the front to also see the girl's face during sex from behind but it's messing up whether it's turning the girl's head 180 degrees on the back view or completely failing at giving me a front view at all
>there are two views defined by an even split in the middle of the image. the view on the right is a pov from the man's perspective as he grab's the girl's ass with his hands sinking deep into the skin of her ass cheeks while fucking her. the view on the left is the same scene but from the front of the girl featuring her expression and reaction to the sex with the man standing behind her as he is fucking her pussy from behind with his hands on her ass, sex from behind, the man's head is head out of frame on the left view, the setting is the same in both instances in a dimly lit hotel room and the girl is leaning over a desk with a desk lamp on it her breasts resting on the desk
This is my shitty attempt, maybe someone else can come up with something better?
>>
>>
>>
>>
>>
>>
>>
>>8835145
The approach I've been having is to use first an NL description of general aspects of the composition and then do a separate line for booru tags to support it and emphasize individual features (even in a multiple views scenario) and changing one or the other up depending on how the gens do/don't reflect what I'm aiming for.
So in this case I would suggest rephrasing most of it into 1-2 sentences and all the small stuff relegated to tags as you normally would do.
>>
>>
>>
>>
>>
I condensed it to
>sex from behind, vaginal, grabbing another's ass, deep skin,
>the man is fucking the girl in the ass penetrating her pussy with his girthy penis while he grabs her ass with his hands sinking deep into her soft ass cheeks,
>the image is evenly split in two views of the same scene by a thin black line going vertically through the middle with the view on the right being the pov of the man and the view on the left is from the front
It just wont work. With this prompt it's giving me the pov sex from behind on the right but the left side is just the girl standing, facing the viewer
Why would the danbooru tags not apply universally
>>
>>
>>
>>
>>
>>
>>
>>8835194
I am obviously aware of the danbooru tags. The ones that would help me would be multiple views and maybe 2koma/comic but neither do
So unless someone else can successfully prompt for what I'm trying to do I think I'll just have to write it off as not possible
>>
File: 00028-2420658026.png (3.5 MB)
3.5 MB PNG
>>
>>
>>
>>
>>
File: 00022-4170550367.png (965.8 KB)
965.8 KB PNG
>>8835198
nta but i felt like trying it, something like this?
>>
>>
>>
File: 20260220172452-3379832193.png (1.5 MB)
1.5 MB PNG
>>
Been out of the loop for a little under 2 years now. What's the best way to prompt two different characters with their own lora? Last I remember the best way was to use regional prompts, is that still the way to go or is there a better way now?
>>
>>
File: 00016-2726385220.jpg (958 KB)
958 KB JPG
>>8835241
it depends on which 2 characters and what type of interaction you want, it can be as easy a just pure prompt or using several hacks like controlnet and inpaint to hell and back
>>
>>
>>
>>
File: 1753495008725952.png (1.1 MB)
1.1 MB PNG
>>8835185
you're still over-complicating it. remember that it's a 0.6b model and try something like this instead:
instant loss, 2koma, multiple views,
top panel:\(rosemi lovelock is looking smugly at the viewer, 1girl, solo, rosemi lovelock, nijisanji en, grin, looking at viewer, hand to mouth, mesugaki, outdoors, portrait\)
bottom panel:\(1girl, 1boy, mating press, plap, sound effects, boy on top, fat male, deep penetration, vaginal, anus, uncensored, leg lock, impregnation, cum in pussy, hetero, folded, on bed, indoors, bedroom, full body\)
>>
>>
>>8835308
This is exactly how I tag 2koma and have no prob.
>quality, artists, # of players, 2koma, general tags not easily conveyed in nl
>top panel is yada yada yada
>bottom panel is yada yada yada
It's a lobotomized model so speak to it like a fellow indian anon when describing things. Don't be too verbose.
>>
>got a new GPU
>this completely broke reforge because of course it did because lmao cuda problems
>had a reforged neo install I didn't bother with the other day because it was requiring cuda 13.0 and I wasn't updating to that and breaking anything, but since I have to use it now anyway whatever
>get it installed and running for all of about 10 minutes until I start trying to install extensions
>restart UI, EVERYTHING COMPLETELY FUCKING BREAKS
>can't figure out how to revert it, deleting extensions does nothing
>try and just do a clean pull and restart
>now it's throwing an error requiring python 3.13
I want to die
>>
>>8835322
kekerina of time. I ran into the same shit trying to initially get on Neo. Just a straight up git clone and install with requirements was throwing up errors related to python 3.13 as well. I'm sure I had a separate install of it which I verified so I changed the env paths to reflect it and then it started throwing shit about numpy, some audio engine (I imagine due to the wan support) and some other minor shit. Raged and deleted everything. Do a bit more research and see some shit about Stability Matrix basically doing everything for you and the option to swap ui's on the fly. Cool. Try it. Some python error again. Fuck you. Uninstall everything go to the Neo page and actually read this time. Choose the uv install route. Everything turned out better than expected.
Try the uv route, anon.
>>
>>8835326
thanks, that actually worked.
also, neat, gens take about half as long as they did before. wasn't quite the goal/purpose of the GPU upgrade but still nice I guess.
maybe now I'll actually fuck around with anima tomorrow.
>>
>>
File: 00140-474227639.png (1.4 MB)
1.4 MB PNG
>>
>>
>>
>>
>try out anima
>try to upscale to see if it actually is fucked or if it was just people being retarded and/or a fucked cumrag upscale implementation
>upscaling works perfectly fine
wow it really just needs controlnet tile and for actual trainers that matter to implement code for it in addition to it actually being trained the full way through since this feels pretty rough at the moment.
also I'm guessing upscaling issues people were complaining about is just another failed cumrag implementation(who could have possibly seen that coming), since 1.5~x upscaling with a decent amount of denoise works perfectly fine, it just has the issue of not having a controlnet option to reinforce the composition.
>>
>>
>>
>>
>>
>>8835381
I am quite literally upscaling images that are taller than they are wide.
it's a bit rough around the edges and it has some really obvious resolution driven anatomy issues but it's functional and isn't spazzing out.
https://files.catbox.moe/6ubki7.png
>>
>>
>>
>>
>>
>>
https://huggingface.co/nightknocker/cosmos-bert
>BERT for Anima/Cosmos.
>This is not an adapter model, but rather an early replacement for the T5/Qwen model.
>This means that the T5, Qwen, and LLM adapter files are about to say goodbye.
>>
>>
>>
>>
>>8835411
see >>8834948 and replies (i do think no comma is better)
prompt structure is the same as before except you can use NL. so think about camera angle, composition, and body shape first, then do whatever.
note that the advice on the anima page https://huggingface.co/circlestone-labs/Anima about what order to put stuff before the main prompt is mandatory, shit will break if you don't put things in order. also note that unlike CLIP typos can cause tags to just be ignored
>>
>>
>>
>>
>>
>>
>>
>>
>highres fix upscale 2x, 0.3 denoise, no upscaler "none" option
>send to img2img
>denoise 0.32, multidiffusion tilesize at base resolution, mix of diffusers option, overlap of 512 for sdxl, rescale 2x, random seed, click orange button
>open result in krita, downscale 4x image, erase any mistakes using lower resolution image, save new 2x resolution image that looks perfect with no double bellybuttons or weird shit happening in it with more detail that took 2 minutes more to make
>post
highres fix works great in a1111 too have been using it forever, i dont get why anyone would use an upscaler alone when they all just take longer for the same or worse result as "none" in sdxl
>>
File: ComfyUI_temp_qquqa_00015_.jpg (178.9 KB)
178.9 KB JPG
Thanks for all the help, bros
Anima truly is wondrous
>>
>>
https://files.catbox.moe/sgd8nb.png
>>8835395
only a retard would bother trying to upscale that much though. it doesn't matter what you do it's always going to look like trash unless you either look at it on a phone or you have it zoomed out at half scale. Maybe if the models were trained at 2048px it would be functionally fine but we've seen how that sort of shit goes with illustrious v2.
>>
>>
>>
>>
>>
>>8835436
>only a retard would bother trying to upscale that much though
It's easier to upscale to absurdres and then downscale it back compared to inpainting little details that get destroyed by 4ch vae. Honestly anima didn't feel like much of an improvement when it comes to intricate outfit details.
>>
>>8835451
>Honestly anima didn't feel like much of an improvement when it comes to intricate outfit details.
well the positive effect of the 16ch vae while genning is destroyed by the shit upscaling and the positive effect while training was destroyed by the 512x dataset.
>>
>>8835451
i disagree, besides the obvious degradation that training at 512x made, i've done a bit of testing on highly specific concept loras, and the ability to pick up details is greatly improved, better than sdxl after months of testing and dataset curation. now some of this could just be simply that anima's dataset is superior but i don't think so.
>>
>>8835451
what details. to upscale to resolutions that large you have to nerf denoise to miniscule levels and all you get is nearly exactly what you started with at low resolution but fuzzier and with a lot of halo/glowing between linework. You're not introducing or improving any details over what you had at the start.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>8835501
Upscaling isn't free - it costs time. On SDXL that wasn't much of a problem since it's so fast AND upscaling was a massive improvement over no upscaling because of the compressed VAE, but on Anima that's both slower and much better at details out of the box, its relatively smaller benefits are much harder to justify when juxtaposed against its much harsher downsides. You're basically paying more to get less benefit.
>>
>>
>>
>>8835509
That's only temporary, it's ought to get much better once it's trained on 1024p. Still, even the way that it is, it simply artifacts less and produces much more coherent hands when they're far away from the camera.
>>
>>
>>
>>
>>
>>
File: 1770908519487510.png (924.7 KB)
924.7 KB PNG
>>
>>
>>
>>
File: 1765960094985900.png (1.2 MB)
1.2 MB PNG
>>
>>
>>
>>
>>
>>
>>
>>
File: 00069-21997391.png (3.7 MB)
3.7 MB PNG
>>
>>
>>
>>
>>
>>
>>8835670
actually on that note my first two bakes worked fine it's just in an "meh, ok" state, which took ~2hours per full complete bake at 3500 steps, but will changing from 1024x1024 to 512x512 help with the quality or does nothing?
>>
>>8835672
i only got worse results when i tried 512x512 but maybe that's just me. it's also much faster so you might as well try it out if you're curious
i've also only gotten results at the "meh, ok" level so far, although there are definitely things it does much better than sdxl models
>>
>>
>>
>>
>>
>>
>>
File: 20260221205651-1619422409.png (1.2 MB)
1.2 MB PNG
>>8835680
>>
>>
>>
>>
File: 20260221212420-2426971709.png (1.1 MB)
1.1 MB PNG
>>8835699
i respect women irl
it's only in the world of imagination that i rape and enslave them
>>
>>
>>
File: ComfyUI_temp_zajzh_00003_.png (1.7 MB)
1.7 MB PNG
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 1746368146400697.png (2.1 MB)
2.1 MB PNG
/hgg/ - pedophile mossad billionaires general
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: ComfyUI_temp_zajzh_00020_.png (1.1 MB)
1.1 MB PNG
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: ComfyUI_temp_zajzh_00034_.png (1.3 MB)
1.3 MB PNG
penis in pussy
>>
>>
File: 1754625965731883.png (1.1 MB)
1.1 MB PNG
>>
>>
>>
>>
>>
File: ComfyUI_temp_huojp_00009_.jpg (489.2 KB)
489.2 KB JPG
Asuka doesn't really care about Shinji's dick
>>
>>
>>
>>8836012
Unlike with SDXL models and their CLIP TE that simply binds image patterns to trigger words and does nothing else, shitmixes of any model that uses an actual LLM as a TE actively destroy its knowledge and NL comprehension by diluting them with uncaptioned image slop. Your only real option is to finetune on a dataset that is as well-tagged and described as the training data.
>>
>>
>>
>>
>>8836017
That's still infinitely better than blindly merging with existing checkpoints and LoRAs that were made on, at best, images captioned with just tags alone without a hint of NL if not straight up completely uncaptioned images.
>>
>>
>>
>>
>>
>>8836033
shuffle is good because you could unintentionally be creating a bias based on the order of tags. same reason to do caption dropout, you could be unintentionally be associating training tags with each other in an undesirable way. i like very high dropout, 0.5
>>
File: 20260222142450-3286071325.png (1.4 MB)
1.4 MB PNG
>>
>>
>>
>>
>>
>>
>>
File: HByiN9Gb0AAiDVr.jpg (615.2 KB)
615.2 KB JPG
pussy pussy face lora when
>>
File: file.png (82.1 KB)
82.1 KB PNG
updated arena
https://arena.sharkgone.com/
>>
>>
>>
>>
>>
>>
>>
>>8836089
Illu and Noob can't do Rance either. There's a grand total of 9 "rance, solo" images on danbooru. This character is simply too horny, almost every image with him you can barely see him because he's mostly offscreen fucking some girl.
>>
>>
>>
>>
>>
>>8836158
The eyes shape is kinda there. Will try mixing with something.
Chink artists have sometime quite interesting styles, but like with the linked one, they are very inconsistent and low post count, so making lora from like 10 pics would be shit.
>>
>>8836163
I don't think they're very inconsistent but it's true that they have low post count
I've come to find that chiggers completely mog the japanese when it comes to art and then they use it to make gay shit
Something about their use of colors is fantastic
>>
>>8836166
Yeah, I agree. When I see some interesting style on pixiv it's usually chink or gook with 10 images.
Japs seem stuck and westerners either have no self awareness of their low skills or are "original" by drawing lips and noses.
>>
>>8836169
I like lips and noses, westerners just make it look fucking awful because they generally have a grotesque sense of "muh realism" which isn't even realistic, it's just ugly. Chinks will oftentimes do very realistic styles and still make them look great. So when it mixes with anime it also looks fantastic, yuming li comes to mind
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 2026-02-23101512_stealthmeta.png (2.5 MB)
2.5 MB PNG
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>8836432
Zero, unless it's a character and you're too lazy to prompt all her clothes.
The model was trained long enough with dropout to understand it can add things to the image without explicitly being prompted to do it. Like how a girl has nipples even if you just imply she's topless. With a Lora you just worry about the one thing you're training it for.
>>8836449
iirc 0.5 dropout means every time the image is processed it will be missing half of its tags, selected at random. Seems extremely high, but I have not tried it above 0.1.
>>
>>
>>
>>
>>
>>
>>
>>
>>
what's the anima trainer you use?
https://github.com/Moeblack/AnimaLoraToolkit
https://github.com/FHfanshu/Anima_Trainer
ive had enough of sd scripts, maybe i just roll one with diffusers or smth
>>
>>
>>
>>
File: 1769842970535128.png (3.9 MB)
3.9 MB PNG
>try to upscale using anima in img2img
>get weird artifacts
The fuck? SD upscale works without artifacts but it makes way more mistakes.
>>
>>
File: 00075-3531864708.jpg (724.8 KB)
724.8 KB JPG
>>8836507
Very sexo outfit
>>
>>
>>8836507
Because it's genning in small tiles but still sees the full prompt, without tile controlnet. You can somewhat negate it by making your tiles larger, default size is 512 from SD1.5 days but the artifacts only start ways past 1024. Also helps to keep denoise lower.
>>
>>
>>
File: 1764344396077014.jpg (509 KB)
509 KB JPG
>>8836533
1024 is my default tile size and I already tried the other tricks but even with them and at low denoise it just sucks and tries to shove every part of the prompt into every tile.
>>
>>8836531
I was referring to those mistakes, you probably have your tile size set very low. Then the model is working on a 512 square area of her neck too small to see the overall context, and it still has "penis" in the prompt.
>>
>>8836534
That's just how it is. Best thing we have, for now.
In theory you could do the tiling yourself, and update the prompt manually to only what's currently visible. It's just a lot of manual work, especially for the seam fix pass.
>>
>>
>>
>>
>>8836539
Yeah, 1.5x seems a lot better. Btw what's the deal with MultiDiffusion Integrated? Isn't it a similar sort of thing? Doesn't seem to get any progress when I try it, does one step and then just hangs. Doesn't crash but never makes any progress either.
>>
>>
pretrained_model_name_or_path = "E:/sd-webui-forge-neo21/models/Stable-diffusion/anima-preview.safeten sors"
qwen3 = "E:/sd-webui-forge-neo21/models/text_encoder/qwen_3_06b_base.safetenso rs"
vae = "E:/sd-webui-forge-neo21/models/VAE/qwen_image_vae.safetensors"
output_dir = "E:/sd-scripts/output"
output_name = "anima_finetune"
save_model_as = "safetensors"
save_precision = "bf16"
optimizer_type = "adafactor"
optimizer_args = ["scale_parameter=False", "relative_step=False", "warmup_init=False"]
learning_rate = 3e-6
lr_scheduler = "cosine"
lr_warmup_ratio = 0.1
loss_type = "l2"
timestep_sampling = "sigmoid"
sigmoid_scale = 0.35
discrete_flow_shift = 1.0
min_snr_gamma = 5.0
max_train_epochs = 20
train_batch_size = 1
gradient_accumulation_steps = 2
seed = 1337
cache_latents = false
resolution = "1024,1024"
enable_bucket = true
min_bucket_reso = 512
max_bucket_reso = 1536
bucket_reso_steps = 64
random_crop = false
max_data_loader_n_workers = 2
persistent_data_loader_workers = true
llm_adapter_lr = 0
mixed_precision = "bf16"
gradient_checkpointing = true
save_every_n_epochs = 1
save_last_n_epochs = 5
sample_every_n_epochs = 1
sample_sampler = "euler"
sample_prompts = "E:/sd-scripts/config/prompt.txt"
attn_mode = "sdpa"
my current finetune config. It's getting there imho
>>
>>
>>8836549
t bh I've only tried it once but it worked, for a 2x. Didn't see much need to upscale that high on anima, most I just inpaint eyes. I'll give it a few more tries with a more complex scene. And maybe use Multidiffusion instead of ultimate SD, if it's been updated.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>8836568
i use it almost every day anon. you have to actively fight it so it doesn't look like shit. it's very good with getting concepts, styles and characters from a very low amount of tags but its visual quality is just straight up bad by default. I still wish there was updated v3, it's still the best-looking nsfw-capable anime model in my opinion (noob with good loras is a close second though).
>>
>>
>>
>>
>>
>>
>>8836581
i read the 1st 3 sections and the part about concerns at the end. looks pretty good. but this is mainly applicable for proper finetuning it seems. as long as it's using the same basic logic as adam (which it is) then i agree. the other takeaway is just don't use adafactor because it's like a fuckin turbo lora that rapes your accuracy for speed
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>8836712
because the only reason model mixing worked nicely on 1.5 and sdxl models was because of CLIP tag bleeding. that wasn't intended but it was a happy accident for our purposes
T5 and LLMs are much less retarded and don't bleed tags, but they have no consistent way of handling multiple style prompts
>>
>>8836714
So, in essence, it was good because it was shit?
Man, that's a blackpill. I love anima a lot and it's way better but having to use a shit ton of loras or being stuck with single artist tags is really shitty.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>8836807
Oh yeah, I agree. I'm just waiting to see if the full version evens out that particular problem. Are there any NL description taggers even worth it as a starting point? I think there's going to be a huge skill check just because most of us have forgotten how to write a sentence (even a run-on one, lol)
>>
>>
>>8836813
Have you tried it? It's free.
They started training at 1536 res which destabilized anatomy, then they added cosplay images to try and fix it. Plus it only has Danbooru knowledge on the level of early noob eps, and no e621.
>>
>>
>>
https://files.catbox.moe/zfpmyg.png
also this is a really stupid fucking meme and I hate that there are tons of retards posting this shit to xitter without marking it as sensitive/explicit. That they're also typing in "sfw" on their tweet is even more retarded, though that part doesn't really matter.
>>
>>
>>
>>8836861
np be sure to yike and crabscrub for more content.
also if anyone actually wants it here's a shitty 5.6mb lowstep lora for the general composition trained on vpred
https://files.catbox.moe/z4vjmy.safetensors
you'll need to scribble the cat face on yourself but you can do that in 5 seconds in mspaint so who the fuck cares. it'll mostly carry through if you scribble over the lowres before upscaling.
for context of use here's the metadata for >>8836858
<https://files.catbox.moe/pq4fcj.png>
generally you'll want to tag out
>pussy peek, clitoris, table
it was tagged with jz235 in training but since it didn't want to actually catch the drawn on face to that tag it's probably not actually necessary to use in practice and you can just ignore it.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: ComfyUI_temp_dlxyo_00010_.png (1.2 MB)
1.2 MB PNG
>>8836925
I just use animal_print with tribal_dress, which isn't even a tag but works fine. You can also split it up into loincloth+bandeau, because the "tribal" tag adds bodypaint usually so you might not want it.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 1760623234084642.png (940.7 KB)
940.7 KB PNG
>>
>>
>>8836732
I think you'd need a lora for that, or find some niche artist who draws noses like this often (which I doubt you'd fine). I guess you can also just gen a normal pic and then use flux klein or qwen edit to alter the nose with a pic of Annie as the 2nd input
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 1747250000964833.png (2.7 MB)
2.7 MB PNG
Genitals situation on anima is pretty bad
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>find some decent-ish artists that anima can imitate
>anatomy goes out of the window
>have to spend hours upon hours mixing them up
>or I'd have to mix in another artist that looks like utter sd1.5 slop but miraculously fixes all the issues
Surely the final version won't require some "stabilizer" artists/loras for good outputs, right?
>>
>>
>>
>>
>>
>>
>>
>>
File: Comfyui 30804 .webm (438.6 KB)
438.6 KB WEBM
>>8837370
it's kinda bad, I haven't made anything worth posting
>>
>>
>>8837361
>qrd
just the difference between the first and the last volume
for better or worse he's doing this deliberately, because for example these are from the same period https://danbooru.donmai.us/posts/7894022 https://danbooru.donmai.us/posts/77 46095
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 1746334024531980.png (104.1 KB)
104.1 KB PNG
>>8837494
this is all you need for a basic gen, i really dont understand whats so fucking complicated about it
>>
>>
>>
>>
File: 12 (1).jpg (1.3 MB)
1.3 MB JPG
Anima status? I am not messing with that shit until the actual full model is out.
>>
>>
File: 1771804777636.jpg (1.1 MB)
1.1 MB JPG
>>
>>
>>
>>
File: Screenshot 2026-02-25 113704.png (64.7 KB)
64.7 KB PNG
>>8837518
whats she doing to him?
>>
>>
File: 1758172021521665.png (2.6 MB)
2.6 MB PNG
>>8837469
They just don't have the same appeal as good old 'churls
>>
>>
File: 1753109221808529.png (1.2 MB)
1.2 MB PNG
>>
>>
>>
>>
>>
im going to tinker with anima style training at 1536px, low LR, batch size 1 & no gradient accumulation
the model already knows artists pretty reasonably, even if they have only a couple dozen samples on danbooru. high res training and getting more high frequency detail might be more value added
>>
>>
>>8837575
that's not entirely correct
```
The preview model isn't that good at higher resolutions yet.
It is a medium-resolution intermediate checkpoint, trained on a small amount of high-res images.
The final version will have been trained on a dedicated high-res phase. Details and overall image composition will improve.
```
besides, if you dont fuck around youll never find out
>>
>>
>>
>>8837575
it's possible to boost a model ability to handle higher resolutions with trained loras
e.g : https://huggingface.co/nblight/noob-ft/blob/main/noob-ft-1536x-extract .safetensors for noob
I don't know how that guy did it, but it makes noob work at 1536x much better than Illustrious 2, the other sdxl model claiming to be able to handle this
>>
>>
>>
>>
>>
>>
what is this...
https://www.reddit.com/r/StableDiffusion/comments/1revwgq/clip_is_back _on_anima_because_clip_is_eternal/
i fucking hate clip as much as the next guy, but hear me out, maybe this shit will allow us to use artist mixing properly.
>>
>>
>>
>>
>>
>>8837718
>Unfortunately it won't let you use weighting as on SDXL, but from what i tested that also was a bit better at least.
It doesn't seem to be even working as clip should, I don't even understand what the benefit of this would be
All of his examples feel either like cherry picking, or confirmation bias as the difference is not big enough to warrant losing natural language prompting
I've no idea why you'd want to restrict yourself with tags again
>>
File: more_results[1].jpg (732.8 KB)
732.8 KB JPG
>>8837788
from the github
https://github.com/Anzhc/Anima-Mod-Guidance-ComfyUI-Node
>Seems to weaken styles, but needs more testing.
uh oh stinky
and what it's based on
https://github.com/quickjkee/modulation-guidance
it just makes things look more like sloppa
>>
>>8837788
>restrict yourself with tags again
Not necessarily 'restrict', the idea is to use CLIP in addition to qwen. If you've used Flux 1 with its own text encoder you might remember that there were separate prompt fields for CLIP and T5 encoder, it's the same thing here.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>masterpiece, best quality, highres, absurdres, safe, Dramatic digital illustration of a long-haired, handsome mafia boss seated with legs crossed, viewed from a low-angle upward perspective. He exudes authority and danger, wearing a sharp, dark tailored suit with a subtle pinstripe pattern, the jacket slightly open to reveal a dark vest and crisp collared shirt beneath. His long, flowing black hair cascades over his shoulders, some strands catching the ambient light. In one hand, he casually holds a lit cigar between his fingers, with a faint trail of smoke curling upward. His gaze looks down from above, brows slightly furrowed, lips set in a firm line, conveying displeasure or irritation. The low-angle viewpoint emphasizes his dominance, power, and intimidating presence. His other arm rests confidently on the armrest or his knee. The background is dim and atmospheric, possibly the interior of a lavish private lounge or office, with hints of gold trim, leather furniture, and shadowy lighting. The digital art style is refined and cinematic, using sharp contrasts, detailed facial expression, and dramatic lighting to heighten tension and elegance.
/hgg/ never had the makings of a varsity prompter.
>>
>>
>>
>>
>>
>>
File: 00023-1868971004.png (2.1 MB)
2.1 MB PNG
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
reading that thread i wonder if people are doing block weighting to remove style influence, also someone reported they are using 32dim which is way overkill for a style imo...though catastrophic forgetting does seem a bit worrisome
>>
>>
>>
>>
>>
>>
>>
>>
>>8837942
nta but "preferable" is a matter of preference. You might want to just fap-gen things en masse, without having to look for or create a reference for every single image.
Admittedly there are few things noob doesn't know when prompted correctly. Weird clothing, niche /d/ stuff, real-world objects like specific cars, guns. Various symbol-shaped pupils, because the VAE ruins all training data that's not a close-up. etc.
Anima has worse tag knowledge and no e621, so some tags are gone entirely. But it works better with concepts you can explain in simple words.
>>
>>
>>
>>
>>8837956
Yeah, CN helps for poses that don't have a tag or a name, or are just too niche to be recognized. But even with fixed linear and shading it's hard to push the model to produce something it doesn't understand at all. Objects, textures, backgrounds.
>>
>>
>>8837718
anima is pretty decent at mixing artists. much better than 4.5 at least. you have to step back from your expectations on how it "should" work with SDXL and experiment afresh. also adding natural language description on subjective aspects of how it should look (these also work in the negatives) -- what was 1.5 schizo prompting can actually add value now.
i think the DiT architecture vs SDXL's u-net has more to do with artist mixing than the text encoder.
>>
>>
>>
>>
>>
>>
>>
>>8837973
NAI CEO and employees fud Anima because fully trained Anima, with finetunes and loras, is the first model that has any real chance of competing with NAI4.5.
But it's not just them. /ldg/'s resident faildev openly admitted to fudding Anima any chance he gets, because he's a jealous spitefag who has suddenly put on his model training hat and is now trying to build a better version of Anima where he can take all the credit. A bunch of furries also hate Anima and are shitting on it because they're mad it's not being trained on e621.
I've never seen a model with this much hostility directed toward it. Not even Chroma was this bad.
>>
>>
>>
>>
>>
mixing isn't even that bad on 4.5. its main issues are atrocious default aesthetic and mediocre contrast/lighting, pretty much the opposite to how nai v3 was. mixing stability is pretty bad but it doesn't feel significantly worse than unshitmixed clip models
>>
>>
File: 1766180728591545.png (3.2 MB)
3.2 MB PNG
.
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 1767330492238690.png (978.6 KB)
978.6 KB PNG
found a new snakeoil for anima. try removing linebreaks
linebreaks:
https://files.catbox.moe/ykx44b.png
https://files.catbox.moe/2um3gu.png
https://files.catbox.moe/bi7afc.png
no linebreaks:
https://files.catbox.moe/uh0wr7.png
https://files.catbox.moe/8zpv53.png
https://files.catbox.moe/aj6rmt.png
>>
>>8838095
Also for the nostalgia, schizoneg. Wouldn't recommend, it kills style.
https://files.catbox.moe/vi49kw.png
https://files.catbox.moe/a4nffk.png
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 20260227175407-2392990160_cleanup.png (1.7 MB)
1.7 MB PNG
>>
>>
>>8838594
the anima model page tells you were to put tags, which is super important for anima. also be careful not to misspell things, it actually matters. https://huggingface.co/circlestone-labs/Anima
you don't need ratings tags, it will infer that from context and make near-zero difference in x/ys. you don't want score tags since they are shiny ponyslop
>>
>>
I always find hard to make new characters with my models in local, what checkpoint should I use if I want to make stuff with Aria (ZZZ)?
every checkpoint I know doesnt know her tag
what checkpoints use new datasets of images from boorus?
>>
>>
>>
>>
>>8838607
>>8838609
im fucked and cant buy a good GPU, mine is a 4060ti (8GB)
When I try to train lora I'm always buttfucked and showed that if I want I will have to wait 17 days
Maybe my configs are wrong cus im dumb, or maybe Im locked cus my pc config
all I have is good will to make porn that no one is making so I can get anons happy
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>8838602
>the anima model page tells you were to put tags, which is super important for anima
Wait so is there any actual difference in where I put the artist tag?
It says to use after 1girl/character/series but it's such an odd placement
>>
>>
>>
>>
>>
File: 00009-2820939457.jpg (666.5 KB)
666.5 KB JPG
>>
>>
>>
>>
File: 20260228064022-1983959415.png (1.4 MB)
1.4 MB PNG
>>
File: 20260228070341-1440115749.png (1.5 MB)
1.5 MB PNG
>>
File: 20260228065211-1440115749.png (1.1 MB)
1.1 MB PNG
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
I did some permutations on the recommended tag order,
https://files.catbox.moe/sp3obx.jpg
https://files.catbox.moe/przq2h.jpg
https://files.catbox.moe/sxf7lr.jpg
Maybe I am blind or picked bad examples, but I don't see much of a difference
>>
>>
>>
>>
>>
>>
>>
>>
>>8838865
>Uniqueness Rank: My alternative to "global favorites." Since this is a serverless tool, I’ve used CLIP embeddings and KNN to rank artists by their stylistic impact. It’s the fastest way to find "hidden gems" that truly stand out.
>>
>>8838863
>>8838849
https://litter.catbox.moe/qb6n5kvdxdcm36fo.jpg
https://litter.catbox.moe/himfkzxvzip8m5me.jpg
https://litter.catbox.moe/w9bo6g0v5tey1pgu.jpg
What has happened to catbox that it stops working every five minutes?
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 1769334411597905.png (495.6 KB)
495.6 KB PNG
>>
>>
>>
>>
>>
>>
>>
Anima bros, artists are coming for us
https://files.catbox.moe/vhss15.png
https://files.catbox.moe/cfklvh.png
>>
>>8838933
>2 followers
probably a random troll account from a bored zoomer (most likely than not a pro ai one if they're even aware of anima's existence)
in reality anima is still basically unknown by the general public
>>
>>
>>
>>
>>8838933
>>8838936
Paid for and/or pushed by Ani, the /ldg/ faildev who by his own admission is going on a crusade against Anima because he thinks he should have been the one to get the ComfyOrg grant to train a model.
The Anti-Anima twitter account and website was posted on /ldg/ less than 24 hours after that tweet, when it had gotten zero engagement. Absolutely no chance a random ldg anon somehow saw it that quickly.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>8838975
>oekakischizo
splotch/shadow nose
>noodleschizo
dot nose
there are a lot of other very distinct differences but oekakischizo nose type is unique(doesn't always apply when he's posting with a different style thoever).
>>
>>
>>
>>
>>
>>
>>
>>
File: 00037-2542501734.jpg (732.5 KB)
732.5 KB JPG
>>8839005
I don't remember if I posted this one before
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>8839069
>Anima rocks until you need to upscale
I just did some additional finetuning. 1.3k steps @ 3e-6 and it does 1.5x -1.75x upscales just fine. tradeoff is that I'm locked into 1 style, but since i never gen much in other styles anyway, it's not much of a loss and I'll manage until the full release.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: ComfyUI_31556_.png (917 KB)
917 KB PNG
>>8839145
doesn't have much meme potential
>>
>>
>>
>>
>>
>>
>>
>>
File: ComfyUI_31564_.png (3.4 MB)
3.4 MB PNG
>>8839151
>>8839159
haven't had much trouble so far, but I'm mostly using non-hentai artists
>>
File: 1752978873262084.png (1.1 KB)
1.1 KB PNG
>been going through psd posts on kemono for the last 2 hours
>actually got lucky a couple of times
>now cant stop
just fuck my shit up
>>
>>8839164
it breaks if you do too many at once, but two is fine. unironically if you were really obsessed with a show or whatever you'd be better off with a full finetune or some giga dim lora just so it can contain everything it needs to
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: imagem_2026-02-28_222434044.png (14.6 KB)
14.6 KB PNG
explain this then.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 20260301042911-3428111898.png (1.4 MB)
1.4 MB PNG
>>8839001
https://litter.catbox.moe/qzr37rtu5iyxskr9.safetensors
here is a v1 (anima), it kinda sucks. as usual these retarded fucking gacha designs are so complex the ai can barely handle them. i had to be really specific in prompting. it's probably fine for your porpoises. i'll give it another few shots for curiosity's sake, and will bake another for illu just in case. i also have absolutely no idea if NL captions are helping or hurting or don't matter. could just be undertrained?
you will likely need to neg pointy ears since most wings are intrinsically tied to them in anima, a quirk i've noticed in other gens.
mainly the problem is that she is wearing a dress/leotard/shorts and the thigh clamp and clothes layering and patterning are weird, if you really want it to work properly you need to include a bunch of closeups of every part of the outfit.
cute idol whore tho she exudes a strong rapeable aura.
>>
File: 20260301044914-3950968671.png (1.5 MB)
1.5 MB PNG
>>
>>
>>
>>
Is it a skill issue on my part or does inpainting with Anima suck for everyone else? It's like I'm back in SD1.5 era of endless gacha rolls. And the denoise settings don't make sense at all, at 0.4 I get almost identical results while on 0.45-0.5 suddenly it looks completely off and doesn't even match the original image anymore
>>
>>8839365
I haven't messed around with it much to thoroughly test it but it does seem pretty inconsistent even using settings that have worked totally fine with other models. I'm assuming it's mostly an issue with Anima itself in its current state.
>>
>>
>>
>>
>>
>>
>>
>>
>>8839378
>local sdxl inpaint
Noobai just works. For non-anime I guess inpainting with Z would work best
>>8839380
>>8839381
nta but the main problem for both Flux Klein and Qwen Edit is that they still alter the entire image, sometimes ever so slightly and sometimes quite significantly. You can minimize the change by writing a more detailed prompt, but not eliminate it completely
>>
>>
>>
>>
>>8839380
I was able to make this while only providing a pair of images at a time so no, not really. Character/concept LoRAs really are just obsolete at this point, the fact that local will have to continue to put up with them for the foreseeable future is really shitty news.
https://litter.catbox.moe/h09rhe71hpfi1ixh.png
>>
>>
>>
>>
>>
>>
>>
>>
>>8839400
Unless 1girl standing is enough for you it's gonna be years before edit models are gonna be viable. Flux klein and qwen have very shitty porn loras which of course are trained either on real porn or cancerous synthslop.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: test.jpg (1.7 MB)
1.7 MB JPG
>>8839419
You still haven't tried it? It's a cool gimmick, fun to mess around with. It basically turns your reference image into a prompt, though you don't get to read it cause it's already encoded.
Problem is it's very hard to separate character, style, composition, etc so you get a bit of everything, and none of them are accurate enough to replace lora. It can help if you are using a lora and it's too weak, or to get some interesting backgrounds, etc.
picrel is just "asanagi, 1girl, standing, large breasts" with a fixed seed and different references
>>
>>8839435
>You still haven't tried it?
No lol, I saw a few of posts such as your own over the years and concluded it's unusably bad for any real practical purposes. If people somehow figured out how to bring to at least NAI level of fidelity for character references I'd be using the living shit out of it.
>>
More hecking epic doggo tests.
https://litter.catbox.moe/fy9lp86hbkogiwy3.jpg
https://litter.catbox.moe/tm5b35csc6sgte4c.jpg
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>8838939
No trigger word. Didn't test much but q50-75 is probably the best? Anima loras seem to withstand resizing weirdly well.
https://pixeldrain com/l/JkMAVdQZ
>>
>>
File: 00005-4028944822.jpg (993.8 KB)
993.8 KB JPG
>>8839475
The more the merrier
We are (mostly) fine, chat just need to remember this is not /hdg/ so you need to re-adapt and remember how to properly express yourself to have proper human-to-human interactions
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>8839446
For the advertised purposes, yeah. It's more of a bias than a proper reference. But sometimes I want a bias toward a difficult pose without straight up controlnet which is too rigid. Or a bluish tint in my starry sky background, or make an OC more consistent without baking a lora. etc.
>>
>>
>>
>>
>>
>>
>>
File: 1767692711661710.png (1.2 MB)
1.2 MB PNG
>>
File: ComfyUI_temp_euqho_00018_.png (1.7 MB)
1.7 MB PNG
This is the power of Anima?
>>
>>
File: ComfyUI_00004_.png (953.4 KB)
953.4 KB PNG
anima is fun
>>
File: ComfyUI_00010_.png (815.8 KB)
815.8 KB PNG
anima is cool, but back to nai...
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 1748558028563759.png (1.5 MB)
1.5 MB PNG
>>8839654
Wo wir sind da geht's immer vorwärts
>>
>>
>>
>>
>>
>>8839695
>They love drama there
True
>Plenty of autists
You're on 4chan nigga
>Plus ERPfags
Eh it's not that much if you're in it for the fag stuff
I get the pain though. Most of the niggers there are barely literate as well, it's literally called slop lmao
>>
>>
>>
>>
>>
>>
>>8839727
The only way to figure out what you are doing is by baking. I'm getting personally what I consider to be mediocre results with my loras on Anima, but at least it will be a starting point for when the full model comes out.
I've been baking loras for years at this point and I never really knew what I was doing. I just trained stuff over and over again until I was personally happy with the results.
>>
>>
>>
>>
>>8837242
couldn't tell you. I'm not touching anima training until full release. Or at least until it's fully implemented in a non vibecoded fork of a random trainer.
there's a pretty good chance that anything learned trying to train anima now is going to mostly go out the window once full release hits, anyway.
>>
>>
>>
File: 00035-nais_bibicipredv5hggedition-3634857350.png (1022.8 KB)
1022.8 KB PNG
Hello sirs, thanks for inviting me to this thread, very thoughtful ;^)
>>
>>
>>
>>
File: mathsnstuff.png (959.1 KB)
959.1 KB PNG
Ok, I have a question that is hard as fuck to solve
some character loras only work when the trigger word is in prompt, like pic example
if I'm downloading a lora, what should I seek (in descriptions)?
but if I'm training a lora, what should I do (in configs)?
I have lots and lots of characters loras, and some have strong triggers, others do not, the ones that have strong triggers always work better
pls frens help me with this
btw any good guide in how to train a lora will help a lot, I've never found a good guide
>>
https://civitai.com/articles/26762
>We tried higher dropout and more aggressive repeat strategies to strengthen character quality, but the model kept degrading. After fixing dropout and repeat, quality was still getting worse, so we spent a long time debugging. We only recently confirmed the main cause: the e621 dataset. Once we removed it, quality improved clearly and we’re back on track.
holy kek chenkin dropped e621 because it was so shit, furries BTFO
>>
>>
>>8839829
no, the model degraded because they're fucking idiots fucking up the dropout and introducing nonsensical tags. it has nothing to do with e621 and has everything to do with how they handle dataset and training.
keep coping about muh e621 poisoning the model, refusing to handle it properly, and getting the safest model, only capable of pov 1girl 1boy cowgirl like animagine xl
also
>sdxl
lmao
>>
>>
>>
>>8839828
>some character loras only work when the trigger word is in prompt
That is the standard for character loras. Without it the lora will apply to all characters in the scene, if you have 2girls or more.
>if I'm downloading a lora, what should I seek (in descriptions)?
Civitai has a field for it on the right side under Details, called "Trigger Words".
If you are training, add the trigger word as the first caption of every image in your dataset. And make sure you set keep_tokens to 1 if you are shuffling captions, this ensures it will remain in first place.
There is a training guide in the OP, it's old but mostly still applies.
>>
>>
>>
>>
>>8839864
anima full can't come soon enough.
unrelated, i found that you could mix natural language descriptors to artist tags. like if an artist is known for realistic looking rough look you can write like "realistic style of @artist with messy brushstrokes" which is pretty cool.
>>
>>
>>
>>
>>
>>
>>
>>
>>8839875
to be honest: i find that anima to be a bit smarter than novelai but it's not smart enough to segment characters purely by prompting like nano banana is, so it will still rely on regional coper if you want to make two bizarre OCs interact which is still worse than novelai's custom coper. dataset is sept2025 which is newer than novelai but still disappointingly old for a model being trained in 2026, they better update it for the full training.
>>8839876
only the huggingface repo and they don't actively post updates or anything really. it's a comfyui-funded project and the trainer isn't active in any of the existing communities, which might be a good thing
>>
>>
>>
>>
>>
>>
>>
File: 2026-03-02071832_stealthmeta.png (2.5 MB)
2.5 MB PNG
>>
>>8839895
nta but probably not, even on bigger models like qwen image it still struggles with that kind of thing which is why novelai, despite the 'just write' meme, had to build regional into their architecture. the model just isn't big enough to learn how to segment every single tag across multiple characters without some kind of guiding tool. the booru datasets don't even specify which tag applies to which character so it would have to be learning all of that from the natural-language portion of the dataset which is lol
>>
>>
>>8839895
Not necessarily that, after all Danbooru's tags don't have namespaces for the separate characters. I would however expect better understanding of NL in general: more leeway in your choice of wording, more consistent anatomy, better prompt adherence, etc. To achieve consistent character separation without regional coping, you'd need to somehow assign the different tags of the image to the characters depicted on it in the training data, which I guess could be possible with VLLMs ("Here's an image and its global tags, look at the image and assign the tags to the relevant characters"), but would probably take a lot of time.
>>
>>