Thread #108645344
File: highlights_g_108639162_1776696951_1.jpg (1.3 MB)
1.3 MB JPG
Discussion and Development of Local Image and Video Models
Previous: >>108639162
https://rentry.org/ldg-lazy-getting-started-guide
>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP
>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows
>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe
>Z
https://huggingface.co/Tongyi-MAI/Z-Image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo
>Anima
https://huggingface.co/circlestone-labs/Anima
https://tagexplorer.github.io/
>Qwen
https://huggingface.co/collections/Qwen/qwen-image
>Klein
https://huggingface.co/collections/black-forest-labs/flux2
>LTX-2
https://huggingface.co/Lightricks/LTX-2
>Wan
https://github.com/Wan-Video/Wan2.2
>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46
>Illustrious
https://rentry.org/comfyui_guide_1girl
>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Collage: https://rentry.org/ldgcollage
>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg
>Local Text
>>>/g/lmg
>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
383 RepliesView Thread
>>
>>
>>
>>
>>
>>
File: emad-mostaque.jpg (145 KB)
145 KB JPG
what happened to him?
>>
File: 1769432142254647.jpg (1.7 MB)
1.7 MB JPG
>>108645453
>he shits in threads.
yeah, Anifart is definitely a nuisance
>>
>>
>>108645344
Thank you for baking this thread, anon
>>108645432
Thank you for blessing this thread, anon
>>
>>
>>
>>
>>
>>
>mfw Resource news
04/20/2026
>Elucidating the SNR-t Bias of Diffusion Probabilistic Models
https://github.com/AMAP-ML/DCW
>(1D) Ordered Tokens Enable Efficient Test-Time Search
https://soto.epfl.ch
>Frequency-Aware Flow Matching for High-Quality Image Generation
https://github.com/OliverRensu/FreqFlow
>From Zero to Detail: A Progressive Spectral Decoupling Paradigm for UHD Image Restoration with New Benchmark
https://github.com/NJU-PCALab/ERR
>China’s Alibaba launches 10,000-card computing cluster
https://www.scmp.com/tech/article/3349335/ai-race-us-intensifies-china s-alibaba-launches-10000-card-compu ting-cluster
>Modly: Local, open source, AI-powered image-to-3D mesh generation
https://github.com/lightningpixel/modly
>DCW: Elucidating the SNR-t Bias of Diffusion Probabilistic Models
https://github.com/AMAP-ML/DCW
04/19/2026
>ZPix: Local AI image generator and editor powered by open image models.
https://github.com/SamuelTallet/ZPix
>Comfy Canvas: Local inline layer based image editor
https://github.com/Zlata-Salyukova/Comfy-Canvas
04/18/2026
>Rose: Range-Of-Slice Equilibration PyTorch optimizer
https://github.com/MatthewK78/Rose
04/17/2026
>ControlFoley: Unified and Controllable Video-to-Audio Generation with Cross-Modal Conflict Handling
https://yjx-research.github.io/ControlFoley
>TokenGS: Decoupling 3D Gaussian Prediction from Pixels with Learnable Tokens
https://research.nvidia.com/labs/toronto-ai/tokengs
>MM-WebAgent: A Hierarchical Multimodal Web Agent for Webpage Generation
https://aka.ms/mm-webagent
>Qwen2D-VAE
https://huggingface.co/Anzhc/Qwen2D-VAE
>ComfyUI HY-World 2.0 — WorldMirror 3D
https://github.com/AHEKOT/ComfyUI_HYWorld2
>Anima Style Explorer: A free web tool for ComfyUI styles
https://anima.mooshieblob.com
>Stanford AI Index Report 2026
https://hai.stanford.edu/assets/files/ai_index_report_2026.pdf
>>
>mfw Research news
04/20/2026
>Towards In-Context Tone Style Transfer with A Large-Scale Triplet Dataset
https://arxiv.org/abs/2604.16114
>Beyond Text Prompts: Precise Concept Erasure through Text-Image Collaboration
https://arxiv.org/abs/2604.15829
>Motion-Adapter: A Diffusion Model Adapter for Text-to-Motion Generation of Compound Actions
https://arxiv.org/abs/2604.16135
>TwoHamsters: Benchmarking Multi-Concept Compositional Unsafety in Text-to-Image Models
https://arxiv.org/abs/2604.15967
>Repurposing 3D Generative Model for Autoregressive Layout Generation
https://fenghora.github.io/LaviGen-Page
>The Amazing Stability of Flow Matching
https://arxiv.org/abs/2604.16079
>DINOv3 Beats Specialized Detectors: A Simple Foundation Model Baseline for Image Forensics
https://arxiv.org/abs/2604.16083
>Sketch and Text Synergy: Fusing Structural Contours and Descriptive Attributes for Fine-Grained Image Retrieval
https://arxiv.org/abs/2604.15735
>AHS: Adaptive Head Synthesis via Synthetic Data Augmentations
https://keh0t0.github.io/AHS
>VEFX-Bench: A Holistic Benchmark for Generic Video Editing and Visual Effects
https://arxiv.org/abs/2604.16272
>Adapting in the Dark: Efficient and Stable Test-Time Adaptation for Black-Box Models
https://arxiv.org/abs/2604.15609
>From Competition to Coopetition: Coopetitive Training-Free Image Editing Based on Text Guidance
https://arxiv.org/abs/2604.15948
>UniEditBench: A Unified and Cost-Effective Benchmark for Image and Video Editing via Distilled MLLMs
https://arxiv.org/abs/2604.15871
>Efficient Video Diffusion Models: Advancements and Challenges
https://arxiv.org/abs/2604.15911
>Making Image Editing Easier via Adaptive Task Reformulation with Agentic Executions
https://arxiv.org/abs/2604.15917
>Noise Aggregation Analysis Driven by Small-Noise Injection: Efficient Membership Inference for Diffusion Models
https://arxiv.org/abs/2510.21783
>>
>>
>>
>>
>>
>>108645600
On the bright side, 99.9% of what's there are shitty fried jeet loras, so not much of worth is being lost.
>>108645947
Yes.
>>
>>108645995
>Yes.
No lmao. Unless you're using some fucked up trainer that has prompt weighting turned on by default (why would you even weight prompts at training time?). Every trainer I know of just processes your captions as is, so you don't escape anything.
>>
>>
File: deLC_zi_00003_.png (2.4 MB)
2.4 MB PNG
>>
>>
>>
>>
>>
File: 1674062767414593.jpg (54.9 KB)
54.9 KB JPG
>install new custom node for great audio
>brick install
>>
>>
>>
File: Screenshot 2026-04-20 190355.png (243.3 KB)
243.3 KB PNG
>>
dont bother trying to ragebait us api chads
we dont post our gens often because we dont want to hurt your feelings
>>
>>
>>
>>
File: Gemini_Generated_Image_543634e5e1e65bb6bcd348.jpg (1.2 MB)
1.2 MB JPG
>>108646398
maybe so, but we can turn it up a notch, no biggie
>>
>>
>>
>>
>>
>>108646491
>I wish there was a tool that automatically does this instead of manually taking screenshots from multiple angles and poses.
"Hey chatgpt can you create a python script to take N renders of this character on *insert 3d software*"
>>
>>
>>
>>
>>
>>
File: qwen_wf.png (538.8 KB)
538.8 KB PNG
is this Qwen-image-edit workflow I'm using still up2date?
including that 8 steps lightnting lora?(what does it even do? is it making the gens faster but resulting in lower quality?)
>>
File: 5125124516146134.jpg (535.4 KB)
535.4 KB JPG
>>
>>
File: 7568653734673473.jpg (327.8 KB)
327.8 KB JPG
>>
>>
>>
>>
>>
>>
>>
File: 1749844493790541.png (111.4 KB)
111.4 KB PNG
>>108647159
there's nothing to talk about
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: animapreview3base_00002_.jpg (667.9 KB)
667.9 KB JPG
>>108647159
I’m recruiting anons from anime general to post here, /edg/, /hgg/, etc.
What are (You) doing for /ldg/, anonnies?
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>108645728
>>108645736
who rattled nigbo's cage again?
>>
>>
>>
>>
>>
File: Z-Image_00005_.png (1.2 MB)
1.2 MB PNG
>>
>>
>>
https://www.reddit.com/r/StableDiffusion/comments/1squ6in/open_source_ crt_animation_lora_for_ltx_23/
this is cool
https://litter.catbox.moe/70a5t3y01avkj291.mp4
>>
File: 1772910684322616.png (1.3 MB)
1.3 MB PNG
>>108648103
>dataset is only 20 clips
yup, thats a hard pass from me.
>>
>>
>>
>>
>>108647159
>holy shit why is ldg suddenly so fucking dead?
Most people left when they realized the petr* schizo is samefagging with bots. It's just non-stop FUD aimed at contributors. Do you want to read FUD all day? Most people don't, so this general died. You should leave too.
>>
>>
>>
>>
>>
File: 1771039218200030.jpg (340.3 KB)
340.3 KB JPG
>>108648228
i gave up fapping a long time ago
if i fap then i stop genning, and i dont want to stop genning
>>
>>
>>
>>
>>108648327
then you're not paying attention
>>108648367
yes, see >>108647897
>>
File: ZiT.png (2.7 MB)
2.7 MB PNG
>>108648367
>can zit ai generate celebrities
it can do some
>>
>>
File: 1752932406353692.png (16.7 KB)
16.7 KB PNG
Has anyone done mask training with OneTrainer? Right now i have masked out faces with a black mask, with everything else i want to train in white. What should these settings be for that case? (i can easily invert the masks if i have it the wrong way around)
>>
>>
>>
>>
>>
>>
>>
>>
File: arigato.jpg (161.3 KB)
161.3 KB JPG
>>108648449
https://civitai.com/models/2466415/cosmos-predict25-2b-base-distilled- extracted-dmd2-lora
This one works in 4 steps, even on Anima Preview 3, it's some sort of back magic. I will give the new one a try however, looks promising.
>>
>>108648449
>>108648500
yeah but you don't have cfg anymore, can't they just turbo'ed their models while keeping the cfg?
>>
>>108648500
That one destroys details to the point of being unusable. It's a miracle it works at all. I think the official turbo lora is much better than any other one I've tested. The latest RDBT 0.25 is pretty good also, but that one is a full checkpoint and has a strong style bias.
>>
>>
>>
>>
>>
>>
>>
>>108648449
>>108648573
Us in the know call him TouchDownRuss.
>>
File: I wonder who's fudding Anima that hard...png (139.6 KB)
139.6 KB PNG
>>108648544
>>108648577
we know who's fuming lmao
>>
>>108648529
The hell do you need high CFG for? If it's negative prompt you're missing you can give parts of your prompt negative weight with this: https://github.com/pamparamm/ComfyUI-ppm Works with Anima.
>>
>>
>>108648573
>>108648582
i call him mommy actually
>>
>>
>>
File: 1758144888888090.png (296.3 KB)
296.3 KB PNG
>>108648604
>THIS GENERAL IS TRASH DUDE WHY YOU KEEP HERE!!!!
SAAR DO NOT REDEEM
>>
>>
>>108648544
>>108648577
>>108648614
now I want russel to post more here, this schizo's melties are so funny lmao
>>
>>108648585
>https://github.com/pamparamm/ComfyUI-ppm Works with Anima.
first time I've heard about it, usually when you want to use negs without a CFG you go for NAG
>>
File: Klein9BDistilled_EmadMuskaque.jpg (2.4 MB)
2.4 MB JPG
>>
>>
>>
>>
File: shinonome.jpg (183.3 KB)
183.3 KB JPG
>>108648640
All the nodes I checked for NAG didn't work with Anima, but this does so I went with that and I haven't bothered to check up if anyone has finally implemented NAG for anima yet. Also, it might have been a skill issue but my experience with NAG was that it completely fucked my images half the time and I haven't had that problem with this.
>>
>>
>>
>>
>>
>>108648700
Unlikely until after full release. The current turbo lora was actually just an experiment with testing the techniques that I'll use for the full Turbo checkpoint. It was good enough that I decided to release it.
>>
>>
>>
>>
>>108648777
No, for optics reasons. But literally just take 1000 photos from any decent pre-captioned photo dataset and train a rank 32 lora, that's the realism lora basically, it's not complicated. Or just wait for someone to inevitably do that.
>>
>>
>>
>>
>>
File: 1776716800944930.png (1.4 MB)
1.4 MB PNG
>>108648791
>>108647620
this is an everyday /adt/ anima gen, but turdrusell prefer to post here, in Lodestone general
>>
>>108648377
>>108648382
interdasting.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: ol yuffie.png (2.7 MB)
2.7 MB PNG
>>108648449
pretty cool
>>
>>
>>
>>
>>
>>
>>
>>
File: ComfyUI_10363_.png (1.1 MB)
1.1 MB PNG
>>108648856
>slimy
>>
File: 1776022168238295.png (79.9 KB)
79.9 KB PNG
>>108648878
>the SOTA dev posting here
a SOTA dev posting on a SOTA general, yeah that fits
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 9967946948484.jpg (632 KB)
632 KB JPG
>>108647068
Ohhh, I think I understand what you meant by that. No, it looks the way it is supposed to. It was trained on Mushoku Tensei anime screenshots. I'm still playing around with the lora settings, but results are promising
>>
>>108648889
i'm not implying anything. i want it to look good and the computer makes it look good. i don't care about things looking "artistic" or any other definition some random person on the internet cares about like "slop"
>>
>>
File: 94784794.jpg (634.8 KB)
634.8 KB JPG
>>
File: DCom3xEYkMbioLO28D62t.png (987.8 KB)
987.8 KB PNG
An anime model being what finally kills this thread was not what I expected.
>>
>>
>>
>>
>>
>>
>>
File: 191562510455854.png (2.2 MB)
2.2 MB PNG
>>
>>108649062
>>108649053
Good gens, you deseve Anima news!
>>
File: 75656885667.jpg (167.5 KB)
167.5 KB JPG
>>108649011
The shitposting and samefagging orbiting around it is dwarfing the zit meltdowns and chroma meltdowns from last year. I don't know if it's just site traffic being down or what, but when 90% of posts are some bored asshole with a script it gets annoying.
>>
>>
File: 674267857621504.jpg (2.1 MB)
2.1 MB JPG
>>108648449
It's very fast, pretty good quality.
>>
>>
>>
>>
File: 705872094488448.png (2 MB)
2 MB PNG
>>108649115
The turbo version captures less of the artist style and more of it's own distilled style, maybe that's why.
>>
>>
File: 813892313650215.png (3.2 MB)
3.2 MB PNG
>>108649115
>>108649131
Curiously though, when using an artist style lora it seems to adhere to it maybe even closer than the base model by itself.
>>
>>
File: 277170810591572.png (3.3 MB)
3.3 MB PNG
>>
>>
File: thumb-1920-842336-1784243730.jpg (147 KB)
147 KB JPG
>>108648449
This news is a nothingburger. It is known that Anima has worse style interpretation than Chenkin/Noob/Mugen, and now you add speed at the cost of more diluted styles.
What use is it to me to use such a diluted model if I can obtain superior results with Chenkin and its superior ControlNet that can capture any scene I propose with Cascadeur?
These little experiments you make are a nothingburger until you release a ControlNet for Anima.
>>
>>
>>
File: 901834508854949.jpg (2.8 MB)
2.8 MB JPG
>>
>>
>>
>>
File: 434430340517180.png (2.4 MB)
2.4 MB PNG
>>108648909
>>108648937
serene
>>
>>
File: 6ef4ccad93fa68ae364288a1601e44cb-3487670913.gif (1.4 MB)
1.4 MB GIF
>>108649239
Google Cascadeur before saying stupid things. It's a free 3D software where you can build 3D scenes and characters with integrated AI. For example, you can move any part of the body and the character moves along with the limb you drag. It's much more intuitive than drawing and much more intuitive than designing a 3D scene in Blender. Then you export that in depth or canny format and you own whatever scene you set out to create. With that, I don't need some stupid new random anime model that has +0.005 prompt adherence in exchange for 70% worse aesthetics.
>>
>>
File: 841977107163987.png (2.2 MB)
2.2 MB PNG
>>
>>
File: 366465205940280.png (1.3 MB)
1.3 MB PNG
>>
ControlNet, regional prompter, and Homing's new mosaic outpainting solved the majority of SDXL problems. I don't need a new anime model that understands "the cat is on the table" when I can mask where I want the cat to be and I can maks where I want the table to be. It seems retarded to not rely on peripheral tools.
>>
>>108649226
>>108649341
Did the journal factory explode
>>
>>
File: 43195659524137.png (1.2 MB)
1.2 MB PNG
>>
>>
>>
File: 7674373473472.jpg (702.2 KB)
702.2 KB JPG
>>
>>
File: z-image_00957_.png (1.3 MB)
1.3 MB PNG
>>
File: ComfyUI_06674_.jpg (1.3 MB)
1.3 MB JPG
Big russ can you train a lora that fixes the fingers?
>>
>>
>>
>>108649517
fpgaminer made this but I have no idea if is decent
https://github.com/fpgaminer/joyquality
>>
>>
>>
File: ComfyUI_03110_.webm (3.9 MB)
3.9 MB WEBM
>>108649531
They said the same thing about Chroma...
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
As someone who followed Chroma until the "official" 1.0 release which was v50 or whatever it was, where can I learn more about all of the actual versions kekstone is training? There has to be a fucking list that explains them, right? It cant STILL all be in trooncord only?
There was some 2k or whatever the fuck that was touted as better than v48 and below, is that still the "best" regular current Chroma model or is there something after?
There is ZIT Chroma thats in training still, Radience pixelspace thats in training still, what else?
>>
>>
>>
File: 1775240292788259.png (1 MB)
1 MB PNG
>crtanim — CRT / Retro Terminal Video LoRA for LTX‑2.3 22B
https://www.reddit.com/r/StableDiffusion/comments/1squ6in/open_source_ crt_animation_lora_for_ltx_23/
https://huggingface.co/lovis93/crt-animation-terminal-ltx-2.3-lora
>>
>>
File: Look at the view count vs upload dates tho.png (477.1 KB)
477.1 KB PNG
This is AI music right?
https://youtu.be/Ikkizhy3nMw
>>
File: ComfyUI_temp_bnyaa_00001_.png (1.2 MB)
1.2 MB PNG
pic related was made on a MacBook Neo with Anima preview 3 using the turbo lora that came out today. 8 steps took 253 seconds in total. This is around the maximum resolution you can do, and I had to use the Q8 GGUF version of Anima as well.
Good to know I can still generate a 1girl every 5 minutes if I'm bored on vacation with my wife's laptop I guess
>>108649773
the wan 2.1/2.2 rollout was one of the top 5 open source model releases of all time. if I didn't crave audio now as a result of trying SaaS models I wouldn't be burnt out on it either. It's also permanently "good enough" for many basic I2V tasks for simple memes or prototype animations so it'll keep being used for years to come.
>>
>>
>>108650144
It's pretty much all in the discord now since most (all?) of the regular posters have left the general and nobody posts on HF. I'll try to cover them all.
Base is really just v48 since v50 introduced some quirks with its outputs. 2k was an attempt to compensate for the 512 training, but I haven't used it so I can't comment on if the anatomy is any better with it or not. Radiance and Kaleidoscope (Vae-less and Klein-based model respectively) have both had training failures and are effectively cancelled. That just leaves Zeta Chroma (Zimage-based) as the only current project. Unrelated to lode, there's a checkpoint called Spark that is currently in progress.
As for which one is better? Some people like the flash models since they clamp down on the model's schizo creativity, but they're more slopped as a result. For my niche use case the base model is still unmatched.
>>
>>
>>
>>
>>108648449
Imagine Anima but more slopped lmao
Looking forward to seeing some epic 'five porcupines on the left anime girl on the right abstract style with a gundam in the background' nonsense gens in /adt/ and /sdg/ that somehow look even uglier than usual
>>
File: 1747029221116112.png (2.6 MB)
2.6 MB PNG
>>
File: 1761168635420497.png (3.1 MB)
3.1 MB PNG
>>
File: 1770360678208696.png (1.2 MB)
1.2 MB PNG
>>
File: 1753678298627795.png (557.2 KB)
557.2 KB PNG
I don't know anything about model training, is that bump good or bad?
>>
>>
File: 1775262013602130.png (1.4 MB)
1.4 MB PNG
>>
File: Flux2-Klein_00078_.png (1.9 MB)
1.9 MB PNG
>>
>>
File: 1757831978637399.png (1.3 MB)
1.3 MB PNG
>>
>>
File: sabergelion alter.jpg (2.6 MB)
2.6 MB JPG
>>108648449
Holy sorcery, first the high-res lora and now this.
>>108648698
Shoukan ni ouji sanjoushita~
1920 height stretches the body. I should test if it can be combined with the high-res lora.
>>
File: 1280x1600_compare.jpg (2.6 MB)
2.6 MB JPG
>>108650793
Works better at 1280x1600.
>>
>>108650793
>>108650809
Looks way worse with turbo lora
>>
File: boxer alter.jpg (2.6 MB)
2.6 MB JPG
>>108650793
>>108650809
Forgot to mention the non-turbo results use 20 steps.
Unintentional text comes out clearer with 12 steps. Coloring is fancier than base's version of Nipi style.
>>
File: hires+turbo.jpg (3 MB)
3 MB JPG
>>108650793
Stacking the high-res and turbo loras works decently. Dat hand tho.
>>
File: pill dond by nttruslan.jpg (2.7 MB)
2.7 MB JPG
>>108650809
Hires + Turbo at 1280x1600. Clearer unintentional text at 12 steps again.
Takeuchi style comes out sharper and more-simplified with turbo overall. Tends to have some AI artifacts, but the speed is fun when 20 steps takes me several minutes.
>>
>>
File: nipi27 (redrop absorbido).jpg (2.5 MB)
2.5 MB JPG
>>108650853
Hires + Turbo with this one almost looks more like Redrop than Nipi.
>>
>>
>>
>>108650982
>>108651029
can you share prompts?
>>
File: WaiAnima1+TurboLora.jpg (2.5 MB)
2.5 MB JPG
>>108650793
Turbo lora works with WaiAnima too. Wai's aesthetic tuning over base preview 3 still shines through.
>>108650853
>Forgot to mention the non-turbo results use 20 steps.
Correction: 30 steps for the gown ones, and 20 for the MMA one.
>>108651160
For the gown ones:
>masterpiece, 1girl, saber alter, fate \(series\), @takeuchi takashi, evening gown, nightclub, bar, looking at viewer, serious
>Seed: 106979513466387
>er_sde simple
>30 steps
For the MMA one:
>1girl, saber alter, fate \(series\), @nipi27, low ponytail, black sports bra, petite, flat chest, black dolphin shorts, black fingerless gloves, bruise, dirty, sweat, heavy breathing, looking at viewer, serious, stretching, boxing ring, crowd
>Neg prompt: letterboxed, border, black border, white border, black background
>Seed: 790426829359823
>er_sde simple
>20 steps
>>
>>
>>108651255
test with prompt
>Detailed photograph RAW of seven smiling friends of different races that are at a nightclub concert with dim lighting that is shining on their faces, behind them is a crowd of people dancing while fighting with large swords, everyone is holding a sword in their left hand and an intricate beer glass with differently colored beer in the right hand. Far behind them above the DJ there is a sign which has "Minimum drinKing age 021!" written on it in stylized cursive letters.
>>
>>
>>
File: WaiAnima1+TurboLora_1280x1600.jpg (2.2 MB)
2.2 MB JPG
>>108650809
>>108651255
Wai+Turbo, 1280x1600. I like the middle one.
>>
File: WaiAnima1+Turbo_00001_.png (2.7 MB)
2.7 MB PNG
>>108651265
Dozo. Wai+Turbo
>>
File: Anima0.3+Turbo_00011_.png (3.1 MB)
3.1 MB PNG
>>108651265
>>108651306
And here's base preview3 + turbo. Hope you like guys, haha.
>>
File: 1773177153069470.png (2.5 MB)
2.5 MB PNG
https://jeoyal.github.io/MegaStyle/
>style transfer lora
nice
>Flux 1
OH COME ON
>>
>>
File: WaiAnima1+TurboLora_boxer.jpg (2.3 MB)
2.3 MB JPG
>>108650853
And finally Wai+Turbo of that. Shinier and grew an extra finger. This is the prompt that does NOT start with "masterpiece", interestingly.
>>
File: Untitled.png (1012 KB)
1012 KB PNG
Is the right one the best one? For nsfw
>>
>>
>>
>>
>>
>>108650853
>>108650905
>>108650982
>>108651029
Saber Slopper, you anime poster traitor. Such a casual slopper, your posts are pure amateur cringe. You only open your UI to pretend you test things when tdrusell posts.
>>
>>
>>
>>
>>
>>
>>
File: WaiAnima1+Turbo_00014_.png (761.2 KB)
761.2 KB PNG
>>108651492
It's a sloppy job, but hey, three testcases over time.
>>
Does anyone found a solution how to use Euler A with anima without the reoccuring weird noisy patterns? This hasn't really improved since preview1, so I don't think it's pixelstretching due to lack of 1024px training. Maybe ancestral sampler not working well with flow-matched model in general?
>>
>>
>>
>>
File: 1757585511364086.jpg (1.6 MB)
1.6 MB JPG
https://www.reddit.com/r/StableDiffusion/comments/1srk0xx/release_comf yui_diffaid_patches_inferencetime/
https://github.com/xmarre/ComfyUI-DiffAid-Patches
https://arxiv.org/abs/2602.13585
>Beyond improving generation quality, Diff-Aid yields interpretable modulation patterns that reveal how different blocks, timesteps, and textual tokens contribute to semantic alignment during denoising.
looks like it can be used on flux 2 klein
>>
File: Gpt-Image 2.png (2.5 MB)
2.5 MB PNG
>Gpt Image 2 is being rolled out to all ChatGPT accounts
local losted'ed
>>
File: 1772307853516045.png (109.8 KB)
109.8 KB PNG
>>108651704
this is impressive desu, it even managed to write this little text here
>>
>>
File: big if true.png (360.9 KB)
360.9 KB PNG
>>108651687
https://www.reddit.com/r/StableDiffusion/comments/1srk0xx/comment/ohfa yx2/?utm_source=share&utm_medium=we b3x&utm_name=web3xcss&utm_term=1&ut m_content=share_button
>I think the more interesting part of this is SDXL, which will now be able to "edit"
GROK IS THAT TRUE??
>>
>>
>>
>>
>>
File: Capture.png (3.5 MB)
3.5 MB PNG
>>108651687
>https://arxiv.org/abs/2602.13585
if this can make Klein less slopped I'm all for it
>>
File: eulera.jpg (934.4 KB)
934.4 KB JPG
>>108651772
>>108651582
for reference. I think you just never looked at your gens then. This is not a debatable issue. Anyone who used euler a has seen this.
>>
>>
>>
>>
>>108651737
dev, responded. the person was retarded.
> No, you are missing the point too. It's not about giving a model edit capabilities. It's about giving a model better prompt adherence (which in turn can boost pre-existing edit capabilities since they are intertwined with the prompt) and overall image quality by modulating text conditioning.
Sure sdxl will probably be better with inpaint tasks with the help of DiffAid, but it doesn't just suddenly give it the capabilities of a proper edit model like flux klein.
>>
File: dem hips.png (653.5 KB)
653.5 KB PNG
>>108651796
That print-looking pattern? I don't seem to get it on mine though.
>>
>>
>>
>>
>>
>>
>>108652028
I'm not expecing much (or anything at all perhaps), but I do like that they've taken the initiative so far; they had a positive response and they built on it, unlike some I could name who had a fantastic reception and followed it with radio silence
>>
these are seedanece 2.0 text2video nsfw, there is still some safety cucking filters and heavy copyright filters. I wish the chinks would stop with the endless wan2.1 and wan 2.2 finetunes and just focus on finetuning ltx 2.3. I really wish lightricks trained the model on accurate detailed nudity and anatomy.
https://litter.catbox.moe/bvwxgx.mp4
https://litter.catbox.moe/hy925n.mp4
https://litter.catbox.moe/0a9y3l.mp4
https://litter.catbox.moe/936897.mp4
https://litter.catbox.moe/9vigar.mp4
https://litter.catbox.moe/n1lgxt.mp4
https://litter.catbox.moe/j6ls3t.mp4
https://litter.catbox.moe/70lrzq.mp4
>>
>>
>>
>>
>>
>>108651737
SDXL bros??? I knew that Anima was the false messiah, the antichrist of anime disguised as a good anime model, I knew it! Nobody in their right mind posts Anime model news here unless they are the antichrist of anime diffusion!
>>
>>
File: 1770573245233488.png (64.7 KB)
64.7 KB PNG
>>108652177
>>108652187
you fell for misinformation
>>
>>
Did russell mention if there is going to be a preview4 etc or is the p3 "base" it now? Since he writes he will release a turbo model for Anima, I assume its also based on preview3 which would imply that that's it, we are done? Havent followed the thread much in recent weeks so idk if he said something to this end.
>>
>>
File: 1762664986876495.png (1.5 MB)
1.5 MB PNG
>can gen kino in 5 seconds now
HOLYYYYYYYYYYYYYYYY
>>
>>
File: 56546872341885848.png (2.1 MB)
2.1 MB PNG
>>108651982
check yourself before you wreck yourself
>>
>>
File: 1772852862285135.jpg (836.2 KB)
836.2 KB JPG
>>108652296
fixed her
youre welcome
>>
>>108652211
>locally with consumer enthusiasts gpu of 32gb vram?
The model would have trained and developed from the begin to designed from consumer level gpus. Currently the SOTA chink and western ai labs don't want to commit to making video models fit within 24-40gbs of vram at fp16. Ltx 2.3 fp8 720p is pushing the limit for a 5090/64gb of ram pc build.
>>
>>
>>
>>
>>
>>
File: 856725727272.jpg (1.1 MB)
1.1 MB JPG
>>
>>108652215
>>108652338
It's still called "preview", so there's clearly more planned.
>>
>>108652338
>>108652441
go for a non meme base model though, why did you go for fucking chronos :(
>>
>>
>>
>>108652361
Can you mix artist tags? Can you weight them? Can you use ControlNet? For such a small and new model it is pretty shit THO!
>>108652441
Can you show me where it says anima 3 preview?
>>
File: 64565.png (169.9 KB)
169.9 KB PNG
>>108652101
why does this happen for all those links?
>>
>>108652456
>For such a small and new model it is pretty shit THO!
what are you waiting for to make something better? you promised to >>108648583
>>
>>108652456
https://huggingface.co/circlestone-labs/Anima/tree/main/split_files/di ffusion_models
>anima-preview3-base.safetensors
>>
>>
>>
arggh why are you guys using anima just use mugen UGHHHH im so mad why cant you use the model we made??? yeah its melted and it works like fucking garbage but who cares? I'd rather shill my garbage model and fud 24/7 the competition otherwise I will realize I have no reason to live and sudoku because im a clueless talentless grifter and failed developer.
>>
>>
>>
File: 1753718602004154.png (959.4 KB)
959.4 KB PNG
>>108652501
kek, it do be like that
>>
>>
>>
>>108652468
looks like it expired too early forgot to change the settings for longer than a hour. here you go
https://litter.catbox.moe/ilq97n.mp4
https://litter.catbox.moe/1pul90.mp4
https://litter.catbox.moe/b2f5jy.mp4
https://litter.catbox.moe/0y5gmp.mp4
https://litter.catbox.moe/m47p16.mp4
https://litter.catbox.moe/xwp2l7.mp4
https://litter.catbox.moe/60u9w5.mp4
https://litter.catbox.moe/jszt0b.mp4
https://litter.catbox.moe/86ypn3.mp4
https://litter.catbox.moe/qdvb2x.mp4
>>
>>
>>
>>
>>
>>
File: 1763912753777134.jpg (51.6 KB)
51.6 KB JPG
>>108652694
we are too smart for china. they know that if they share their new video models, they will never beat us again kek
>>
>>
>>
>>108646491
>>108646541
>>108646515
Not OP, but using 3D models does make consistency way easier.
My workflow involve bringing Daz3D characters into Blender, rendering multiple angles into a 16:9 character sheets, then using a realism Lora. After that, I create all my scenes in Blender and just render first & last frames with the same realism Lora, only doing a face swap using the character sheet when needed.
The initial pre-production is tedious, but you can also just batch run these overnight and wakeup to all the frames you need.
>>
...guys I dont know even know what is going on or what is needed for these python scripts. I am that dumb. I'm not able to just run the scripts as posted in my Python shell, I'm like not even smart enough to ask the right questions. Just this whole git/pip thing eludes me. Trying to run WAN locally, have used StableDiffusion fine