Thread #108659074
HomeIndexCatalogAll ThreadsNew ThreadReply
H
Discussion and Development of Local Image and Video Models

Previous: >>108655751

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe

>Z
https://huggingface.co/Tongyi-MAI/Z-Image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo

>Anima
https://huggingface.co/circlestone-labs/Anima
https://tagexplorer.github.io/

>Qwen
https://huggingface.co/collections/Qwen/qwen-image

>Klein
https://huggingface.co/collections/black-forest-labs/flux2

>LTX-2
https://huggingface.co/Lightricks/LTX-2

>Wan
https://github.com/Wan-Video/Wan2.2

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46

>Illustrious
https://rentry.org/comfyui_guide_1girl

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Collage: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
+Showing all 312 replies.
>>
>>108659074
>my gens arent in the fagollage
trash
>>
>>108659074
requesting this in any scenery and any effects/editing, please
>>
>>108659100
ok
>>
maybe "smooth skinned amputation" will worrk.
>>
>>108659106
thanks but sadly it doesn't look like her
>>
>>108659049
Do you have a patreon too?
>>
local having an absolute meltdown over saas superiority. gpt image 2 is insane
>>
openai actually shills here.
>>
*yawn*
change bait please it's getting stale
>>
>>108659179
the nonstop posting doesn't come across as desperate at all.
btw it has been 7 and a half hours since the last image was posting in the api general, maybe you should post that there before the general falls off the board.
>>
>>108659074
I should by 5090 some day.
>>
File: 130877049.jpg (494.7 KB)
494.7 KB
494.7 KB JPG
>>108659074
>>
There is "people" that actually waste compute on this btw
>>
>>108659234
>there is
holy esl
>people
sirs
>>
>>108658677
>anima + zit
what does your workflow look like
gen with anima then controlnet zit? or img2img?
>>
>>108659100
Is this the fabled "girlfriend" that you wanted a LoRA of?

>>108659133
No. I don't want Jenny coming after me... she's very petty and never forgives!

>>108659179
Go back to bed, faggot.
>>
>>108659289
she's a twisted fy'ah stat'ah
>>
>>108656011
>>108656084
Where the hell are you posting that an OF whore found you?
>>
Is Qwen Edit the best local edit model?
>>
>>108659470
Overall probably yes. Not few people use klein because it's faster and just as good for many subjects though. Qwen is not *clearly better* for quite a lot of edits.
>>
>>108659498
I tried ph00t's AIO but the results are kinda plasticky. Does qwen edit beat klein for realism?
>>
>>108659592
>I tried the shitmix
wow congrats my dude.
>>
>>108659270
nta but anima does most of the heavy lifting out of the box. after that just pick your poison for upscaling/detailing.
>>
>>108659179
This is good but I don't know why seeing other gens like this gives me the impression that they have something similar to SD1.5 slop vibes, the coloring or something like that.
>>
>>108659655
It's that grease layer on the surface
>>
>this is good
It's really not.
>>
>>108659650
I see
>>
>>108659619
well I've used qwen image 2512 and all that spits out is plastic shit. The edit model isn't trained on synthetic data?
>>
>>108659683
try not using turbo loras, they sloppify the output tenfold for qwen
>>
File: file.png (3.5 MB)
3.5 MB
3.5 MB PNG
>>
>>108659650
with the right lora or finetune, anima could be a great model for realism, even if solely as a 1girl machine. It's fast, it's easy to prompt, it's fun - and I say this as a person who hates anime
>>
>>108659696
i was/am surprised by it. i don't use anime models so maybe it's not a big deal, but the ability to prompt "cosplay photo of 1girl dressed as whoever from whatever" and have it replicate that in a realistic and somewhat accurate way is pretty neat. looking forward to the final release.
>>
Blessed thread of frenship
>>
>>108659688
>that ear
Is she half Ferengi or what
>>
>>
>>108659856
>clothed female
>Ferengi


>>108659910
Cute!
>>
https://github.com/AMAP-ML/DCW
Babe wake up, Alibaba improved the inference quality of images, when ComfyUi custom node?
>>
>>108659982
alibaba increasing alibooba quality, love to see it.
>>
>>108659982
snakiolio
>>
>>108659655
>>108659662
>It's that grease layer on the surface
you better get used to it
the chinese are already slurping up gpt image 2 outputs as we speak to train their upcoming model
>>
>>108660027
>the chinese are already slurping up gpt image 2 outputs as we speak to train their upcoming model
all I wanted was Z-image edit...
>>
why are there no diffusion models for frame interpolation?
>>
File: file.png (355.5 KB)
355.5 KB
355.5 KB PNG
>>108659982
https://arxiv.org/pdf/2604.16044
I tend to believe this isn't a nothingburger when the improved numbers are miles ahead the original, and not just some statistical noise, and it seems to be the case here
>>
>>108660027
A little bit gutter oil to fool gweilo
>>
>>108660027
because if the internet was short on anything it was minecraft screenshots and fake infographics.
>>
>>108659982
>>108660045
it looks more detailled indeed, if that removes a bit of the slop I'm all for it
>>
>>108659982
>brutally underbaked top images
Just stop doing benchmarks, niggas. Nobody believes this shit anymore.
>>
>>108660062
no shit, 20 steps for models that need at least 50 lmao
>>
>>
>try the new diff-aid node that inpaints and masks(?) the area you prompted for
>suffers the same issue of all other masked image edit features

Localbros..
>>
>>108660093
get out, this is a local thread >>108653190
>>
>>108660097
inpainting is a meme, dont fall for it
>>
>>108660087
to be fair, if you're able to get a 50 steps quality out of 20 steps, that's a big win in my book
>>
>>108660117
well yeah faster convergence is always a plus, but publishing your benchs when comparing against a 20 step normal target is so fucking dishonest, and even then I still think that 50 step looks better, so I guess it's a middle of the road method they have.
>>
>>
>>108660131
>>
>>108660117
You can do that by not using euler and choosing a more ruminating sampler.
>>
This. Is. Insane. GPT-Image 2 just changed the game for indie devs. High-quality assets delivered in seconds
>>
>>108660295
kek
>>
>>108660295
>>108660322
thanks Julien, very cool
how'd the funding talks go btw?
>>
>>108660295
Evidently not pixel perfect, so nothing has changed. A separate, specialized model that strictly adheres to standards would be more useful, and obviously ClosedAI doesn't care about that.
>>
>>108660295
>>108660099
>>
>>108660338
>Evidently not pixel perfect
you think trani's brain is developed enough to notice that?
just nod and move along, there's no point in trying to reason with subhuman lobotomites
>>
>>108660131
go home sd3 you're drunk

>>108660295
https://x.com/imgborba/status/2046696599389143169/photo/1
gpt-2 is tibia pilled

>>108660338
>and obviously ClosedAI doesn't care about that
neither does local
>>
Lots of seething in response to the latest state-of-the-art model. What's with the crying? We're here to have fun with AI
>>
>>108660338
>Evidently not pixel perfect
So perfect for indie devs.
>>
>>108660376
why do you want to be off topic so bad? this is a thread for local threads, if you want to post api shit go elsewhere
>>108653190
>>108653190
>>108653190
>>
>>108660370
>>108660376
>>108660381
Julienxisters how are we gonna become millionaires now that GPT 2 exists?????
We're CONDEMNED to be the miserable life of a bottom feeder
>>
>>108660131
please don't tell this is Zeta-chroma
>>
>>108660376
>>
>>108660415
Imagine linking your irl identity with this prompt lmao
>>
>>108659074
Bottom right is AI? I don't believe it.
>>
File: 17612451.jpg (882.9 KB)
882.9 KB
882.9 KB JPG
>>
File: file.png (3.5 MB)
3.5 MB
3.5 MB PNG
>>108660415
lame gen
>>
>>108660376
>>108660460
>>108660470
>hates local
>but loves to lurk on a local thread anyway
what kind of mental illness is this?
>>
>>108660503
raped retard syndrome
>>
>>108659074
>Local can't do text
>t. cloudcuks conveniently forgetting that Ernie image exists and already mostly caught up to NBP with text.

The gap really isn't as is big as it's made out to be by cloudcucks.
>>
>>108660470
is this zimage or did openai switch to a quantized model already?
>>
local is left raped and gaped, but at least alibaba remembered to pull out!
>>
>>108660595
it's not just text, it's also realism + being accurate to the original Ui (whether it's twitter, twitch, youtube...) Ernie can't do that, memes are fun when they are accurately transformative to the real world
>>108660496
>>108660555
>>108660477
>>108660103
>>
35 stars?
>>
julien status?
>>
>>108659074
>that peach fuzz
god I love our women so much it's unreal
>>
>>108660657
what a fucking failure of a model. not even 24 hours and we've gone from "wow it can fucking think! and you can read text on a grain of rice" to "well it's pretty good at making fake twitch screenshots."
i never expect much from openai, but man they always find a way to underdeliver.
>>
localcope
>>
>>108660731
what do you mean? people absolutely love that model, it has the highest gap in the history of Arena AI
>>
>>108659688
No
>>
>>108660376
go create your fart api thread already
>>
>>108660792
it's already been created >>108653190
if he's still posting here it means that he's just a troll, and you know how to deal with trolls that do "off topic posts"
>>
>>108660779
well what the heck happened overnight?
from being able to zoom in on a single grain of rice and read text to squiggle slop books and posters is concerning, to say the least.
>>
>>108660814
I think that model has 2 versions, the free version from chatgpt (probably some turbo shit) and the real deal, not sure though
>>
>>108659074
Guys, I’m trying to get into the local diffusion AI generation scene. Is there a guide for this?
>>
>>108660841
the third line of text in the op
>>
new 'toss
>>
>>108660437
you know its ai because her right breast is floating in mid air. at that angle, both breasts shouldnt be clearly visible like that
this happens because local models cannot think like api models can. they see "breasts" in the prompt and insist on showing both, even when it breaks anatomy
>>108660814
>>108660825
if only you had paid attention during the openai livestream, youd actually know how to use the advanced features
>>
thirty five
>>
>>108660854
Where's the buldge and why is he not farting?
>>
>>108660841
Well, what kind of local genner do you want to be?
>The local purist
Sticks to 100% local models with permissive licenses like SDXL. The gens might look 3 years out of date, but its not about the outputs it’s about sending a message. Uses Forge because ComfyUI is API-focused. A gpu with less than 12gb vram is recommended for the true local experience.
>The hybrid enhancer
Takes advantage of power API Nodes through ComfyUI to level-up workflows. Weave in and out of the localspace, harnessing the power of both top API-models and local tooling
>The API Ascension
Fully decked out with a suite of top-class API models ready to roll. Maximum prompt enhancements accelerated by Gemini 3.1, fed into the latest GPT Image-2 and hand-animated by the diffusion deities at Seedance 2. Capable of generating insane production-ready outputs at a fraction of the cost.
>>
does anyone know of a lora or prompt phrasing that makes girls aim and shoot guns instead of striking a sexy pose next to a gun?
im using NovaAnimeXL
>>
>>108660854
where's the mogus
>>
>>108660415
how do I gen a hapa cutie like this on local
>>
>>108660891
>the advanced features
post some gens using the "advanced features"
go prompt "a pile of rice on a table, "GPT image" is printed on one of the grains of rice."
>>
>>108660093
>>108660595
>>108660657
>Cloudcucks still think their models have moat

No you don't. Plus they revealed on stream that GPT image 2 literally looks up images on the internet, once local gets the upcoming Ernie Image edit we will be 90% of the way there. And that is naturally enough to surpass cloudcucks, because local is uncensored. GPT is literally just an autoregressive model, like the one we have, but with added tool usage.
>>
>>108660924
>models with permissive licenses like SDXL
wtf why not use an Apache 2 licensed model like Wan as an example. oh because this is ai generated slop i got baited by slop

you're also forgetting one of the schizo dimensions of a true local purist - if you're running a local model on someone else's PC you're still a cuck
>>
>>108660948
by generating local api calls to image gen 2
>>
>license
>>
Just how astroturfed are the top accounts on civitai?

>1girl with midjourney
>never posts prompts
>no model info
>always makes it to the front page with a ton of likes
There's no way this is organic, indodogs are botting this shit to hell.
>>
>>108660891
ok but how does this relate to local models?
>>
>LOCAL IS UNCENSAAAARED
>>
>>108660814
I tried the model myself, it can definitely replicate small text, like tiny labels on clothes. No idea about images posted here though.
>>
>>108660925
Use anima.
>>
>>108660797
summer is approaching. dylan and jamal have probably already stopped going to school. they know their future lies in becoming taxi drivers kek
>>
>>108660657
It looks worse than Nano Banana Pro at realism.
>>
>>108659234
better expenditure than making celebrities loras
>>
>>
>>108661081
looks shit
>>
>still fudding
>>
>>
>>108660376
>>
>>
>>
>>
>>108661148
>It looks worse than Nano Banana Pro at realism.
debatable. the outputs are less generic than nbp, but at the cost of speed so it takes longer to experiment with prompts
nbp feels like zit (generic but consistent) while gpt image 2 is like z base if was actually good (more variety but slower)
>>
>>
>>108661285
nice style
what do the faces look like
>>
any tips for eye contact? it seems random no matter what i prompt, sometimes the girl will stare at the camera and sometimes not
>>
File: Untitled.png (3.3 KB)
3.3 KB
3.3 KB PNG
What can we do about this?
>>
>>108661354
why should we care? this is /ldg/, if you want to talk about the state of /adt/, go talk about it on /adt/ retard
>>
>SOTA at photorealism

Not so fast cloudcucks. Nothing has surpassed the Flux.2 VAE yet.

>Technically BTFOs Google
>Technically BTFOs ClosedAI

Idk how BFL do it, and the model is open and free! One thing is certain, more compute is simply not enough to be objectively better than local across the board.
>>
>>108661387
Nah, I don’t think so. It started here, so in a way it’s ours, and we’ve got some responsibility for it.
>>
>>108661403
>we’ve got some responsibility
who's "we"? I didn't do anything, I'm here to talk about local models, and so far it's been API trolling and /adt/ concern trolling
>>
>>108661401
is this a joke or something? klein looks like you took the image and cranked up the luminosity and color saturation to the mask, wtf is this piece of garbage??
>>
>>108661414
/ldg/ is a big thread that spun off a bunch of smaller side threads, so it kinda matters to highlight what’s happening in those too.
>>
>>108661438
>it kinda matters
no, I don't really give a fuck about other generals, like I said, I'm here to talk about local models, not about some petty rivalry between generals, dunno why /adt/ rent so free in your head but you need some help
>>
>>108661333
if its already messing up anatomy and generating six fingers from a basic 1girl output, i wouldnt get too excited
>>
>>108661422
Maybe if you are blind. Here's a closer look at the images, which one captures more detail? Even a kindergartener could figure that out
>>
>>108661451
nice
>>
>>108661480
Real image I used for inspiration- https://thumbs.dreamstime.com/b/blonde-woman-washing-plates-sponge-domestic-kitchen-48728607.jpg

Flux 2 Klein knows about real photographs, not the slop you see AI models usually output and it's unmatched in that aspect.
>>
>>108661354
Good point, I think we should conquer them and make them our colony, honestly, it’s probably the only way they’ll survive at this point.

For example:
/adt/ would become: /ldg/’s anime thread
/atg/ (Ace Step General) would become: /ldg/’s local music diffusion thread
/adg/ (API diffusion general) would become: /ldg/’s API diffusion thread

If people see “/ldg/” in the thread name, I think it would attract new anons on its own, since it’s a recognizable and trusted name.
In other words, /ldg/ is guarantee of good quality.
>>
>>108661480
the skin is completly plastic and the light is uniform as if shadows don't exist, are you blind or something? a detailled 3d pixar image is still a pixar image, tell me you're trolling, right?
>>
where is the realism lora, russel
>>
>>108661505
I like the idea!
Could we do the same for the /h/ generals?
Like “/ldg/’s hentai thread”? that could restore trust and quality in our hentai board and put an end to the /hdg/ vs /hgg/ rivalry.
>>
>>108661505
>>108661527
what kind of mental illness is this?
>>
>>108661527
>>108661505
tdrusell here, if /adt/ and especially the /h/ diffusion threads start becoming /ldg/ colonies, I’d officially post my Anima news there, since they’d basically be extensions of this great general.
>>
>>108661507
Maybe genning makes you blind to what a real photo looks like vs. fake photo.
>>
WHY THE FUCK ARENT WIDGETS SHOWING ON ON THE SUBGRAPH?! I'M ADDING A TON BUT NONE SHOW REEEEEEEEEEE
>>
>>108661550
so you're completly delusional? damn, AI psychosis is real
>>
>>108661548
ani here, I'm here to congratulate on tdrusell's model Anima, I admit that I lost and that he's a way more talented lad than me
>>
>>108661498
thx
>>
>>108661550
Also, it's quite sad that a possibly 100B+ parameter cloudshit model performs worse than a tiny open model, and one of the main things it users brag about is already surpassed by it. I want to see this tech advance as much as you do, and that's quite disappointing.
>>
>>108661551
hidden / nested parameters -> Show all?
>>
>>108661582
>>108661551
I asked claude and it knew right away, it's so good at finding answers.

Using basic reroute nodes solves it.
>>
>>108661542
fudophrenia
>>
>muh magic flux 2 vae
lmao
>>
Any anon using a local LLM agent to give proper tags/writing to a gen in a process of chat/refining ideas?
I'm still using the old and crude method of randomizing tags but I'd like something more focused.
>>
>>108661452
>petty rivalry between generals, why /adt/ rent so free in your head but you need some help
This is a drama invented by (You), /adt/ whiteknight
>>
>>108661452
why do you waste your time answering a sharty retard? you are basically talking to a wall
>>
>>108661672
ok anifart whiteknight
>>
>previous thread
>"ldg is worthless, 4chan is worthless, no one posts here and everyone left"
>ITT
>"ldg is a big thread we should help the smaller threads"
???
>>
Is there any Flux Klein LORA for turning images into video stills? The I2V services and local models tend to do better with images that look like a video still (for obvious reasons) but I've had pretty mixed results with asking Klein to change the style of photos to look like home video stills/etc. Seems like something that could be refined a lot with a well-made LORA
>>
>>108661779
What video style? Either just prompt it or if it's not defined enough, get a lora.
>>
I will not be impressed by an image model until it has spatial intelligence, it's the god given mission of image models to server as world models with spatial intelligence.
This is GPT-Image. Better than every other model but not that big of a jump.
>>
>>108661787
I just want it to
- change focal length to a more realistic one for a video camera
- remove photo postprocessing effects that would not be present in a video
- remove any visual signs that this is a physical photograph (dust, scratches, glossy surface, etc)
- add mp4 compression artefacts
- add blur if appropriate
- leave the subject matter unaffected as much as possible (it won't be perfect obviously)
etc.
>>
>>108661804
If it was captioned then you can prompt for it. For blur and mp4 I'd unironically rather use a normal postprocessing (non ai)
>>
ani here
i will soon fartmaxx on a flight to japan for talking with investors <3
>>
>>108661815
I disagree. https://chatgpt.com/share/69e8e833-83f8-83ea-a453-75a7c3e53af5
>>
>>108661835
ai;dr
>>
>>108661772
In this dying website where everyone left, /ldg/ is still a big thread, we shold help the smaller threads (by conquering them)
>>
>>108661847
It's just a very exhaustive list of the small perceptible differences between a postprocessed image and a true video still, which would be tedious to write out manually. Just skim it.
>>
>>
>>108661854
>we shold
saar?
>>
>>108661639
Relax API cuck. This time I included a real image for comparison because I'm not kidding when I say you don't know what real images look like.
>>
>>108661895
Not even trying to troll (even though anon trolled so hard with GPT2 the other day...) but unironically how is GPT2 so bad at 1girl? That's nuts.
>>
>>108661895
why her hair is aliased on flux 2 klein? looks like it's from a video game lmao
>>
>>108661908
Look at their livestream. Then look at every Chinese model that's ever been released. That explains why all their models are slopped.
>>
>>108661895
use zit instead of flugggs :D
>>
chinese researcher GF status?
>>
>>108662000
cao ni ma'd
>>
>>108661895
we can do this all day buddy, your little toys dont stand a chance
>>
>>108662046
@grok jizz on this girls face NOW
>>
>>108659074
Hows LTX 2.3 1.1 ??
>>
i feel bad for ani
>>
>>108662121
The output looks good and it's fast but I can't get it to perform the specific kind of degeneracy I want.
>>
>>108662135
nsfw lora works, right ?
>>
>>108661639
I can't stand the patterned noise that it has. NBP images look clear.
>>
>>108662135
my issue with it was the awful audio quality, is it better with 2.3?
>>
>>108661401
Agreed. Most people here are a bit blind. I've learned a lot from here about how much 'visual sensitivity' or 'visual IQ' (or whatever we might call it) varies among the general population. I'm big on the blurry older style photos, but that's not the only kind of photo that really exists. The fact that the Flux example is a hair more "plastic" (which sort of disappears if you zoom out a little) is not sufficient to declare it the loser when it is obviously giving a more real overall impression (and it's not close)
>>
>>108662135
Will it work why my Workflows ? Like MrXin Workflow on Civit
>>
>>108662209
>The fact that the Flux example is a hair more "plastic" (which sort of disappears if you zoom out a little)
>which sort of disappears if you zoom out a little
the absolute state of localkeks
>>
>>108661616
get and set nodes are working inside blueprints?
this is a game changer
>>
>>108662046
>Gets rekt at realism by local
>Still thinks his shitty slopped model stands a chance
>>
>>108662306
why are you using flux 2 klein to make your point? Local has a better representative of realism in the name of Z-image turbo
>>
>>108662306
gpt image 2 is more lewd in this pic lmao.
anyone with eyes can see that flux klein is getting completely mogged here. no work needed from me
>>
>>108661401
for all its faults I still think Chroma is best
>>
how would i prompt for a blank stare like mind broken
>>
>>108662439
just prompt for anime
>>
>>108662384
>why are you using flux 2 klein to make your point?

I am interested in small details. E.G. skin, hair textures, every grain in the image. I know it's hard to notice, but the Flux.2 VAE captures those details very well which makes it stand out in realism, while every other model that doesn't use this VAE (including NBP/GPT Image 2) is less than perfect at capturing such details.
>>
>>108662306
the proportions of the klein image are all wrong
the store doors are too big and the street is too small
>>
status on new chroma?
>>
>>108661246
What artist style is that?
>>
>>
>>108662462
she's just tiny
>>
>>108662439
empty eyes, expressionless,
>>
>>108660376
What kind of prompt you used to get that old analog/film look? or you added that effect manually?
>>
>>108662437
It is best at amateur iPhone style realism and is unmatched at being utterly unhinged, though there likely aren't many funds left for a proper Klein Chroma tune.
>>
>>108662602
imagine if lode didnt waste time on schizo experiments and just focused on one single good finetune
>>
reminder
>>
>tfw you're completely lost, have lost track and have no idea of what the present state of the art meta is, and hopping between tag based and natural language prompting has caused your prompt technique to atrophy into pidgin prompt nonsense that is the worst of both worlds.
>...and then you updated comfy and everything just stopped working
I'm tired boss
>>
>>108662629
>reminder
we must be reminded that you needed the help of an API model to make this image?
>>
>>108662629
>>108662657
samefag
>>
>>108662657
>>
>>108662753
It's rare that I agree with a "samefag" comment but in this case it seems really obvious. But maybe it's just a very obvious troll and someone who took the bait
>>
>>108662753
>>108662793
samefag
>>
>>108659289
>Is this the fabled "girlfriend" that you wanted a LoRA of?
i dont' know what you're talking about. i have no knowledge of ai stuff. i just want her with any ai edit. maybe cyberpunk or prehistoric hunter-gathering style, please. >>108659100
>>
File: file.png (3.4 MB)
3.4 MB
3.4 MB PNG
#comfy
>>
>>108662850
I fucking hate people who upload their workflows and hide all their messy shit behind a node
>>
>>108662793
sharty bullshit is kind of obvious
>>
>>108662850
>penis
are you from /hgg/?
>>
>>108663039
wait til you check out the messes they hide inside "subgraphs"
>>
>>108663046
source?
>>
File: 1_00031_.jpg (3.3 MB)
3.3 MB
3.3 MB JPG
so its true that cloudcucks cant generate any goon material?
>>
What are the websites for this, civitai and tensor art? Which is better?
>>
>>108663187
http://127.0.0.1:8188 is the default for comfyui, you can specify a port in the launch options
>>
>>108663205
how did you get access to my comfy?
>>
>>
File: toss.png (3.7 MB)
3.7 MB
3.7 MB PNG
>>
>>108663223
lol
wheres the lora
>>
>>108663223
awoooooga
>>
>>108663217
fine curves
>>
>>108663223
where amogus?
>>
>>
>>108663270
>tattoos
*vomits*
>>
>>108663270
cute face, minus the tattoos
>>
Is there a way to improve the accuracy of klein when upscaling or editing? It keeps warping parts of the image.
>>
>>108663394
https://huggingface.co/Yyyueyu/Flux2-Klein-9B-Consistency
>>
>>108663409
awesome, thanks.
>>
>>108662823
>i just want her with any ai edit
Can't help you bully girls at school, sorry!

>>108663223
I don't see it.
>>
>>108663409
>>108663431
dunno why that anon is linking that, is not from the original author, might be just a mirror, but beware

https://huggingface.co/dx8152/Flux2-Klein-9B-Consistency
>>
>>108663448
good call, it was the first thing I got with search
>>
>>
>>
>>108663484
asshole mouth lol
>>
>>
> reading on phone
> screen has crack
> made it look like Qwen in the OP said "Owen" for a second
kek
>>
>>108663498
*blowjob mouth
>>
love me some 'eedance
>>
>>108663524
nah as an anus enjoyer I recognize an asshole when I see one
>>
>>108663484
>>108663511
Nice anon, I was checking out older gens from Chroma and DC-2K is a pretty good model, I dunno why I switched to the hd-fp8 one
>>
>>
>>108663629
>tattoos
>>
>>108663561
love that pose
>>
>>108663561
>dunno why I switched to the hd-fp8 one
speed!

>>108663629
very cool, is it just anima3 or with lora?
>>
>>108663648
anima3>zit refiner, workflow from here >>108642763 with some tweaks
>>
not seeing any rentry links about inpainting, any decent models come out in the past few months? been trying a few based on sdxl from a while back and not having great luck
>>
[general]
shuffle_caption = true
keep_tokens = 2
caption_extension = ".txt"

[[datasets]]
resolution = 1024
batch_size = 2
enable_bucket = true
min_bucket_reso = 256
max_bucket_reso = 2048
bucket_reso_steps = 32
[[datasets.subsets]]
image_dir = "E:\\sd-scripts\\mylora_images\\myfirstlora"
num_repeats = 96
is_reg = false

@echo off
cd /d E:\sd-scripts

accelerate launch --num_cpu_threads_per_process 1 sdxl_train_network.py ^
--pretrained_model_name_or_path="E:\ComfyUI_windows_portable\ComfyUI\models\checkpoints\illustriousXL_v01.safetensors" ^
--dataset_config="E:\sd-scripts\config.toml" ^
--output_dir="E:\sd-scripts\mylora_output" ^
--output_name="myfirstlora" ^
--save_model_as=safetensors ^
--network_module=networks.lora ^
--network_dim=16 ^
--network_alpha=32 ^
--learning_rate=0.0003 ^
--unet_lr=0.0003 ^
--text_encoder_lr=0.00003 ^
--optimizer_type="Adafactor" ^
--optimizer_args "scale_parameter=False" "relative_step=False" "warmup_init=False" ^
--lr_scheduler="linear" ^
--max_train_epochs=1 ^
--save_every_n_epochs=1 ^
--mixed_precision="bf16" ^
--gradient_checkpointing ^
--cache_latents ^
--clip_skip=2 ^
--min_snr_gamma=5 ^
--multires_noise_iterations=6 ^
--multires_noise_discount=0.3 ^
--max_token_length=225 ^
--seed=42
pause]

This is my first animu character lora. I’m making it on illustriousXL_v01 to start with the basics, kind of like reading the Greeks when you begin studying philosophy.
My dataset consists of 22 images ,I know it’s smal, but I want to see what I can get out of it.
For the next version, I’m planning to increase the dataset to around 50 images and include tags that I didn’t add before.
Right now, all the images have white backgrounds, so it might be overfitting, but I still feel the need to test how the model behaves, at least for this first attempt.
>>
i begin to think that /ldg/ is its own cute sister general
blessed
>>
>>108663704
>illustriousXL
added to my filter, thanks
>>
>>108663536
>no sound
that's why this thread is made for this >>>/wsg/6132196
>>
>>108663704
>Right now, all the images have white backgrounds,
can't that be ignored by tagging it? haven't touched lora training since Kohya_ss in 2022
>>
>>108663688
cool, have to check it out

>>108663704
Don't train text encoder, it's already fried to shit
>>
>>108663704
As of 2026, /ldg/ has officially ended support for SDXL, this model is no longer receiving assist. Please upgrade your model at your earliest availability.
>>
>>108663704
Increase num epochs and decrease repeats, more useful output. I usual do 10 epochs or less. If you end up frying you can use earlier epoch output this way.
Invert alpha and dim values
That LR might fry it by being too high.
Don't attempt to specify different lr for unet and te on your first try.
Either replace optimizer with AdamW or Prodigy (and set LR and alpha to 1 if you choose the latter)
lr schheduling should be cosine
Don't smash random shit you don't understand into optimizer args
Multires noise crap is supposed to work with older epsilon models but I don't recall having success with it, I would recommend skipping that shit.
>images have white backgrounds
I would just bucket paint different colors to a few of them.
I would bump weight decay a bit in the optimizer args and add <0.1 dropout when training with a small dataset, but that depends on how noisy the data is.
>>
>>108663768
>it's already fried to shit
Yes but some loras are extremely difficult to train without also training text encoder on SDXL. I would keep it in general.
>>
>>108663688
>>108663629
tdrussel should create an Anima realism and furry fine tune and put it behind a paywall to finance the anime one.
>>
>>
>>108662209
>Some people have low visual sensitivity or, heh, i call it visual IQ *does epic fedora tip to establish dominance*
>Anyway, if the image looks bad you can simply impair your vision. Try taking off your glasses or walking to the other side of the room
I do not think it is a 'visual' IQ issue we are dealing with here
I think Klein gets the edge in that example too but c'mon, don't be silly
>>
>>108663704
gpt image 2 just dropped and this goofy ahh talmbout illustriousXL
wtf yall doing
>>
anima is lit bruh, fr fr
>>
>>108663561
Is dc-2k from silver oxides? I can't remember where it is on HF.
>>
>>108663536
Shame that this model can do full blown porn yet its potential is not being fully realized. Such a waste.
>>
>>
>>
>>108663753
I tagged everything I saw using the appropriate Danbooru tags. Maybe it won’t be so overcooked.
>>108663768
Thanks. Which option is the text encoder? Is it `--text_encoder_lr=0.00003`? Should I set it to 0?
>>108663800
I’ll take that into account. Do you have a good all purpose SDXL lora or a user who makes good lora that I can check to look at their metadata?
>108663867 >108663726 >108663774
I need to start somewhere. My next step will be Anima, but before that I have to pay tribute to Noob after this lora.
>>
>>108663960
set to 0 and network_train_unet_only = true
>>
downloaded a1111 and got hello asuka to work, whats the next step?
>>
>>108664077
uninstall it and install comfyui
>>
>>108662437
jesus christ thank you so much for this laugh I really needed it KEK
>>
Is there a model that's actually good at breaking down images to guidelines/sketch? I wanna study a bug girl, and NAI is failing me comically bad.
>>
>>108664108
>https://github.com/lllyasviel/Paints-UNDO
>https://lllyasviel.github.io/paints_alter_web/
closest implementation I know of
>>
>>108659074
POST the fucking catbox for the knight girl holy FUCK
>>
>>108664127
Thanks bro, much appreciated.
>>
>>108662850
I just started genning locally and this shit is still confusing to me
>>
>>108663688
>anima3>zit refiner, workflow from here
does the workflow need https://catalog.ngc.nvidia.com/orgs/nvidia/teams/maxine/collections/maxine_vfx_sdk this shit
>>
>>108664127
wtf this is neat! thanks!
>>
>>
still no great realistic anima but the 2d part is awesome
>>
>>108664190
when i imported the workflow i got missing node errors about that so I just removed it from the workflow. maybe it makes it better?
>>
>>108664349
I ran this in comfy folder and it worked
 .\python_embeded\python.exe -m pip install nvidia-vfx --extra-index-url https://pypi.nvidia.com 
>>
>>108664349
holy fuck i'm in love
>>
>>108664108
>Is there a model that's actually good at breaking down images to guidelines/sketch?
All the edit models? Flux, Nano Banana, GPT Image, etc.
>>
>>
>>108663960
>Do you have a good all purpose SDXL lora or a user who makes good lora that I can check to look at their metadata?
I can't endorse anyone. Just checks loras you have been using and know that aren't underbaked or fried.
This is the command I used to run for SDXL loras btw (for Noob):
>python sdxl_train_network.py --v_parameterization --pretrained_model_name_or_path ~/models/NoobAI-XL-Vpred-v1.0.safetensors --tokenizer_cache_dir ~/lora/tokenizercache/ --train_data_dir ~/lora/images/ --shuffle_caption --caption_separator , --caption_extension .txt --keep_tokens 1 --resolution 1024 --cache_latents --cache_latents_to_disk --enable_bucket --min_bucket_reso 256 --max_bucket_reso 2048 --bucket_reso_steps 64 --dataset_repeats 8 --output_dir ~/lora/output/ --save_precision fp16 --train_batch_size 2 --max_token_length 225 --xformers --max_train_epochs 10 --persistent_data_loader_workers --max_data_loader_n_workers 1 --seed 44453 --gradient_checkpointing --mixed_precision bf16 --logging_dir ~/lora/logs --log_with tensorboard --zero_terminal_snr --loss_type l2 --training_comment "Trigger word is blabla" --save_model_as safetensors --optimizer_type Prodigy --learning_rate 1.0 --max_grad_norm 1.0 --optimizer_args weight_decay=0.01 decouple=True d_coef=1 use_bias_correction=True safeguard_warmup=True betas=0.9,0.999 --lr_scheduler cosine --lr_warmup_steps 0 --min_snr_gamma 5 --prior_loss_weight 1.0 --network_dim 16 --network_alpha 1 --network_dropout 0.08 --network_module networks.lora --save_every_n_epochs 1
Use 0.10 warmup if using AdamW
>Noob after this lora.
Noob uses v-pred so ditch min_p too on top of noise offset if you haven't already when you move on to that.
I will also for the last time recommend training the text encoder, but I guess you should see it yourself how it works out.
>>
>>108662629
I'd do local but Gemini is just better in everything except it can't do porn and deepfakes, why can't local models be as good?
Why can't I throw an image and tell it "Hey, put her in a bikini" and just understand and do that instead of generating some other bitch in a bikini with 3 legs?
>>
>>108664522
>why can't local models be as good?
because API models are 100b at least, we can't run that
>>
>>108664536
By 100B what are you referring to? I can believe the TE/(V|L)LMs are a few hundred something billion MOEs with say 20-50 billion active params, but I doubt they are running massive UNETs. It would be too economically impractical.
>>
>>108664572
>It would be too economically impractical.
it is, there's a reason why OpenAI stopped Sora 2, they were losing 15 millions dollars per month
>>
>>
>>108664579
I mean yes sure but that is not as directly related as you think. After all these models are also on the API and being offered presumably not at loss.
Z-Image base is 0.01$ per megapixel of FAL and banana 2 is 0.08. They are bigger than the average local model but I don't they are 100B big. Comparing API inference costs we can assume they would bleed a lot of money with current prices if they were that big.
>>
>>108664572
we have local models that are competitive, qwen and deepseek series, they just require 30k in equipment to run properly
>>
>>108664522
because the local community is stuck in a perpetual emperors new clothes delusion, convincing themselves that cucked models like flux klein 2 are actually good.
why would china or whoever bother making a genuinely good model when they can just capitalize on the shit eaters who will hype any new slop release (like ernie) as long as its "local"?
all the serious genners have been using apis exclusively for a while now because local models have hit a brick wall in realism and surrendered to churning out anime models instead
>>
>>108664503
Lovely
>>
>>108664499
I got a pretty good breakdown from Nano Banana Pro, I'm hoping it's actually decent. I'm noticing the furthest left claw has the wrong number of joints and the numbering on the body segments is wrong, so we'll see...
>>
>>
>>108664652
For /lmg/ yes.
For here that's a cope.
>>
>>108664716
>For here that's a cope.
it's not, do you really think a 6b model can rival Seedance 2.0 or Gpt Image 2? it's not happening
>>
>he doesn't know what tool calling is
>>
cloudkeks are still at it? youd think theyd be having fun genning memes or whatever with their new model but it seems not
>>
>>108664736
is that like a booty call? because I know what that is.
>>
>>108664728
I meant cope as in "Even if you sink many thousands of dollars getting your hands on an H200 or whatever there are no local models released that can rival Banana pro or GPT Image 2 for you to run."
>>
>>108664742
genning while posting on 4chan seems to be completely out of the realm of possibility for localcucks
>>
>2x buzz rate
>ongoing site-wide issues
what did they mean by this?
>>
Fresh

>>108664784
>>108664784
>>108664784
>>108664784

Reply to Thread #108659074


Supported: JPG, PNG, GIF, WebP, WebM, MP4, MP3 (max 4MB)