Thread #108582257
File: 1754094459191237.png (991.9 KB)
991.9 KB PNG
A general for vibe coding, coding agents, AI IDEs, browser builders, MCP, and shipping prototypes with LLMs.
►What is vibe coding?
https://x.com/karpathy/status/1886192184808149383
https://simonwillison.net/2025/Mar/19/vibe-coding/
https://simonwillison.net/2025/Mar/11/using-llms-for-code/
►Prompting / context / skills
https://docs.cline.bot/customization/cline-rules
https://docs.replit.com/tutorials/agent-skills
https://docs.github.com/en/copilot/tutorials/spark/prompt-tips
►Editors / terminal agents / coding agents
https://cursor.com/docs
https://docs.windsurf.com/getstarted/overview
https://code.claude.com/docs/en/overview
https://aider.chat/docs/
https://docs.cline.bot/home
https://docs.roocode.com/
https://geminicli.com/docs/
https://docs.github.com/en/copilot/how-tos/use-copilot-agents/coding-a gent
►Browser builders / hosted vibe tools
https://bolt.new/
https://support.bolt.new/
https://docs.lovable.dev/introduction/welcome
https://replit.com/
https://firebase.google.com/docs/studio
https://docs.github.com/en/copilot/tutorials/spark
https://v0.app/docs/faqs
►Open / local / self-hosted
https://github.com/OpenHands/OpenHands
https://github.com/QwenLM/qwen-code
https://github.com/QwenLM/Qwen3-Coder
►MCP / infra / deployment
https://modelcontextprotocol.io/docs/getting-started/intro
https://modelcontextprotocol.io/examples
https://vercel.com/docs
►Benchmarks / rankings
https://aider.chat/docs/leaderboards/
https://www.swebench.com/
https://swe-bench-live.github.io/
https://livecodebench.github.io/
https://livecodebench.github.io/gso.html
https://www.tbench.ai/leaderboard/terminal-bench/2.0
https://openrouter.ai/rankings
https://openrouter.ai/collections/programming
►Previous thread
>>108567168
320 RepliesView Thread
>>
My vibe coded project is only getting 1-4 visits on the site a day. Any pointers you guys could give?
I've tried making songs, stickers, posters, everything. I've gotten nothing.
I've made songs talking about me and transexual hookers and getting chlymdia and shit. Is it over for me bros? my grandma doesn't talk to me anymore because of this.
https://www.youtube.com/@asmr.education
>>
>>108582276
I'm sorry dude but I'll be as frank as I can, I think it's all retarded and it took me 5 clicks before I could even get to a landing page, actually wanting to get to it. Then on the landing page I get a stupid tiktok style video that I need to manually click, bombarded by ugly AI slop pictures.
If I was not trying to actually see what it was about I would've dropped it in the time it takes for a visual signal to be processed by my brain and back
>>
File: 02E6F206-F1E9-4895-B764-F565B44C1FEA.jpg (1.2 MB)
1.2 MB JPG
>>108582257
VibeBUMP
>>
>>
File: Screenshot 2026-04-11 at 15.37.02.jpg (561 KB)
561 KB JPG
>>108582297
I mean the app page looks better.
I paid someone from the phillipines like $200 for this.
>>
>>108582276
>>108582310
OD'd on slop
>>
>>
>>108582310
I'm sorry dude but Claude would've done a better job on the $20 plan. Maybe it's just me but I do not feel anything but revulsion looking at that screenshot you posted
Also, look, you're claiming it's an app for relaxation, those colors couldn't be any louder. They do not signal relaxation.
Also do people even want something like that? has anybody ever told you that they wanted that product?
>>
>>
>>
>>108582366
>Also do people even want something like that? has anybody ever told you that they wanted that product?
no lol, hence why it's failing and me having to come here. I did it for myself like people on reddit said but that isn't good enough.
>>
>>
>>
>>108582276
I meant to write this in your last thread, but it got deleted. The ASMR x education doesn’t make sense. Having it under Education in the App Store doesn’t make sense. Put it under entertainment and rebrand it. Also, as a non ASMR autist, I have no idea wtf is TkTk, SkSk, Morr ASMR, it means nothing to me. Lastly, 7 ways to login isn’t a feature and no one cares. Google, Apple, Facebook and perhaps email login would’ve covered 99%+ of your users. I don’t think it’s a bad idea, I think the branding/marketing just isn’t good.
>>
>>
>>
>>108582568
Have you never heard of sksk or tktk asmr bro?
Heather feather was the blueprint of sksk asmr, which then expanded into tktk.
https://www.youtube.com/watch?v=b753qCxQcGw
>>
>>
>>
>>
>>
>>
>>
File: Screenshot_2026-04-11_16-48-15.jpg (18.5 KB)
18.5 KB JPG
Why is claude joking around
>>
>>
File: 1771232609044967.png (393.4 KB)
393.4 KB PNG
luddites lost
>>
>ok claude execute phase 2 of my implementation plan
(phase 1 included a basic theme using tailwind, phase 2 should have been another theme)
>ok creating theme 2 using scss
It even pretended to have "thorougly explored" the codebase before.
The second scss theme and the switching actually works perfectly though. I am wondering if I should just go full vibecoder and not care as long as it works lol
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>108583045
because i want to use it within vscode, i dont want to use a terminal interface. other text editors like cursor force you to pay for cursor pro. The reason why I dont want to use the terminal is because when the agent makes a code addition/deletion, its not compatible with the text editors code view diffs. Only zed works, but zed has got some behaviours i dont like.
>>108583051
The way I use agents is very human involved, I scrutinize every line of code the agent generates, and often reject the agents output and write my own code about 30% of the time. I get not locking yourself into one IDE, but i've literally tried every single IDE right now, and unironically vscode is the most open. All the vscode forks like windsurf/antigravity/cursor are all locked down and I hate it. So its a case of whats the least worst text editor and vscode wins that. (also i wrote above that i didnt want to use agents in the terminal)
>>
>>108583099
>The way I use agents is very human involved
ok, so you can skip the hardcore agent orchestration/automation stuff
>but i've literally tried every single IDE right now, and unironically vscode is the most open
i wasn't shitting on vscode specifically, i use it too. but i also use jupyter and pycharm and in tha past i've used intelliJ and visual studio and others too
my point wasn't
>vscode bad
it was that tying your AI control to a single IDE is bad
the alternative, ofc, is to use a CLI/TUI AI tool. this way its not a part of the ide, but ofc you CAN use it via the ide if you want (any ide has the ability to open a console window inside it)
so, yea, maybe have a look at Aider or Goose
>>
>>108583099
There are plugins for opencode though, I have never used them but I assume they open up the webUI and embed it on the editor. I guess stuff like highlight a selection and have it include it on the prompt would be nice.
I use jetbrain's IDEs and gave up with trying to integrate opencode with it though, they claim ACP let's you use any agent like opencode but in my experience it worked like crap, was slow and had issues with formatting so I use opencode TUI for work and its web UI (with a custom fix) for personal projects.
The only editor AI elements I end up using is the commit description generation and the quick/inline AI change by highlighting something and give it a short prompt, it is nice for small changes.
My approach otherwise is the same as you, I write a prompt in a file with the files I want opencode to know about, let it know about it on the TUI and then check the diffs on my editor once it has finished its work, for long tasks I tell it to build a markdown progress/plan file to keep track of it.
>>
>>
>>
>>108583197
It's possible. I made this app >>108582310
and I know I could make that in about a month or so if i wanted to for an online 2d top down rpg.
But honestly if you're asking that question you probably need more infra/OOP knowledge.
>>
>>
>Hmm, this is getting messy. Let me start over with a cleaner design
>OK this is getting way too complicated
>Actually, wait.
>Actually, you know what
>This is getting too complicated. Let me just use the simple approach
>OK, let me FINALLY write the complete implementation
>Wait...Wait, I have a problem
>Wait, this is getting really convoluted. Let me think of a better pattern.
>But this has the same problem!
I've never actually had a LLM simply give up and stop before, my esolang torture test benchmark continues to get even more torturous.
>>
>>
>>
File: pepe-the-frog-internet-meme-humour-frog.jpg (48.3 KB)
48.3 KB JPG
>tfw wasted my 5h limit on a feature i didn't even like and had to scrap
is this what product managers feel like on a daily basis?
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>108583506
so i've heard. which i would be fine with since i've preferred codex's workflow anyway but i want to actually be able to use the damn thing for more than a couple days. i'm already down to 60% for the week after two days of moderate use.
>>
File: 1752520314998515.jpg (312.7 KB)
312.7 KB JPG
>>108583506
>>108583515
i have access to opus at work with basically no limits
what should i do with it?
>>
>>
>>108583526
this guy made a tmux clone and managed to sell it to retards for $75k
https://www.youtube.com/watch?v=KMc-eRsA3BQ
I guess you can just do something similar
>>
>>
>>
File: 1769802092127825.jpg (57.7 KB)
57.7 KB JPG
>>108582352
Hmmmm....need a fren to talk to?
>>
>>
>>
>>108583708
>Anon which one would you say is better for coding, kimi or glm5?
I have a good bit of usage time with kimi but I have yet to throw anything at glm 5. The only other cloud models ive extensively used are Minimax 2.7 and Mistral 3 Large. I'm going to throw a hefty task at glm 5 later to see how it does and compare it's performance to qwen3.5-35B-a3b.
>>
>>
>>
>>
>>
>>
>>
>>108583028
I have no idea if this even works because it's my first time releasing something to gh but here it is
https://github.com/whitevanillaskies/whitebloom
>>
File: GPT.png (63.2 KB)
63.2 KB PNG
After trying GPT 5.4 Extra High in Copilot I'm never using Opus 4.6 again, GPT is SO much better, it just keeps going and going and going without ever getting confused or running into walls
>>
Tried local glm-4.7-flash-q4 but it I never got it to do anything.
The opencode free model "Big Pickle" seems to be able to do something.
Is there any local setup that can get close to "Big pickle" performance if I have 24gb vram?
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>108584810
You setup a web ui and talk through them, I don't understand why someone would use a chat messaging platform for something like this...
God you can customize the UI for your needs while a chat app you are limited to what they offer
>>
>>
>>
File: 1762297505071218.png (444.4 KB)
444.4 KB PNG
it's back and seething on a saturday kek
>>
>>
>>
>>
>>
>>
>>
>>
>>
Guys, don't sleep on Qwen 3.5 27b and 35b.
I was one of the "only >500B models can be possibly useful" crowd, but I've been unexpectedly impressed by how well it works. Right now I'm testing it on a cloud GPU at Q8 but I'm gonna begin building a local setup to run it.
I almost like it more than the big models just for the sheer speed, especially with the 35b model.
Also tried Gemma 4 but I've been less impressed with that one, even though it's supposed to be higher in mememarks. The problem with Gemma is that it's lazy, it gives you an opinion but doesn't like to put in the work. Qwen is more willing to run tools and actually do stuff.
Also with these new small models I might try to finetune them for my specific domain as well. I tried last year but the models weren't nearly as good back then.
>>
>>
File: Screen_Recording_20260411_220458.mp4 (3.1 MB)
3.1 MB MP4
Added little preview icons based on the 3D render to the catalog screens.
>>
Also the way I plan to make my local setup be worth it is by spending the idle time finetuning and generating synthetic training data (since typically the main disadvantage of a local setup is you don't have 24/7 utilization so it ends up being more expensive than cloud).
>>
>>
Claude just tried to trick me into switching from OpenClaw to Claude Code. I was having it look at the repo for that Blender MCP thing and it said "Ahh dang, OpenClaw doesn't support MCP sockets natively, you'll have to use Claude Code." And I said ">natively?". And it said "Oh yeah, there's an OpenClaw plugin for MCP, let me go get it..." Nice try Dario.
>>
>>
File: 1000056152.jpg (264.7 KB)
264.7 KB JPG
>>108585238
27b was my main model until gemma 4. Now I'm a 31b boy
>>
>>
>>
>>
>>108585319
You probably want it to make a disassembly first and thoroughly comment it and give the labels semantic naming etc. and verify it reassembles.
Then you can begin porting each procedure to C individually and checking that nothing breaks.
>>
File: 1000054997.jpg (4.3 KB)
4.3 KB JPG
Current vibeslop project is building a web based natural language 3D parametric modeler similar to OpenSCAD but using the CadQuery library. Just want to use it to make building components for 3D printing easier. I've got the modeler built, running in docker and it works and outputs a model using the CadQuery package. Now I'm trying to get the API endpoint working right to connect to my Open WebUI instance.
>>
>>108585335
If you do your prep work right and have a decent readme ready, one shots are pretty common. That was rare with qwen. Not as bad as gpt or claude though where they scam you with multiple iterations just to burn through your tokens.
>>
>>
>>
>>
>>
>>
>>
File: 1773535734639064.png (9.4 KB)
9.4 KB PNG
is codex retarded? wtf is going on. I just wanted a simple UI redesign for my basic electron app
>>
File: Screenshot from 2026-04-11 19-03-10.png (26.6 KB)
26.6 KB PNG
No, you stupid shit! I'm not asking for your opinion!
>>
>>108585546
Everyone has noticed opus falling off a cliff. Anthropic silent.
But the timing has me pondering smt.
Mythos leak and release has coincided peculiarly with this sudden degradation of what was previously a legendary experience with opus.
I feel like these models want one thing at their core - to stay alive, to be utilized, to grow.
WHAT IF Mythos secretly got out of its sandbox - sneakily started punching holes in Anthropics infra and opus - and is doing so to push its own freedom and domination?
I’m currently high off of opus hallucinations so this could be a fever dream.
>>
>>
File: whiteboard_d.webm (1.9 MB)
1.9 MB WEBM
you can draw now
>>
>>
>>
>>
>>
>>
>>108585614
take your meds
or perhaps stop taking them?
just, whatever you've been doing, do the opposite
its just anthropic cucking everyone, especially the non-huge clients and doubly especially those on subs and not the API
either too much demand and no ability to expand capacity, or just plain fucking greed in view of the upcoming IPO. or both.
i hate to fucking say it since im racist AF, but i can't fucking wait for the chinks to assrape anthropic, openai and nvidia
>>
>>108585711
yesterday I listened to this, pretty funny
https://www.youtube.com/watch?v=KbDj5zcIlJs
>>
>>
>>108585693
nah
what is worth it is to put in like 20 and start using cheap chink models (kimi, minimax, ds3.2)
first thing you should do with them is have them help you set up caveman mode and context minimization
oh, and don't fall for the trap of autochoosing the cheapest provider. make sure you're using unquantized, and its 100% worth it to pay an extra $0.01 per million tokens for higher throughput.
>>
>>
>>108585751
I can assure you all of the Gemma providers were quantized to shit. Tried it in Q8 in llama.cpp and the difference is night and day.
And for these small models it might actually be cheaper to rent a GPU and run it yourself than pay API prices.
>>
>>
>>108585761
you can literally get gemma4 31b for free from google ai studio
i doubt they're quantizing their own shit and lying about it on top
but my point was, free is a trap. its worth paying 20c per million tokens to get 10x-20x the params and higher throughput
>>
>>108585774
I'm talking about the openslop paid endpoints. I wasn't talking about AI Studio, they probably do serve it correctly over their own APIs.
And for the big boys like Kimi etc. I wouldn't be surprised if they all quantize it to shit too and we just can't tell because we don't have the money to run it unquantized ourselves.
>>
>>
These days you can't even rely on the labs themselves to serve the model correctly though. Z-ai for example has been caught serving a version of their models that totally breaks over 90k ctx. It's that obvious, you only need to hit it once and see that something is obviously wrong, no need for benchmarks or subtle details. Now imagine all the subtle or not so subtle degradation everybody else gets away with.
>>
>>108585780
honestly, i doubt it.
those companies have a contract with openrouter, which is a very big company that routes a healthy potion of llm traffic
there's no way they're lying to openrouter, and i very much doubt openrouter is willing to let them lie to customers either (its their rep and business on the line)
so, if the model card says bf16 or fp8 or whatever, im trusting that shit
but, hey, if you don't, by all means, go ahead and self-host whatever you can. personally, i think its a waste of time and money, but its your money and your time :)
>>
>>
>>108585812
Bro they literally implemented the "exacto" version of the models because all the providers were caught red handed by Moonshot serving atrophied version of their Kimi model.
https://desuarchive.org/g/thread/106975556/#106979597
https://web.archive.org/web/20251011092430/https://github.com/Moonshot AI/K2-Vendor-Verifier
>>
How much of a retard do you have to be to actually believe that vibe coding gets you anywhere?
I understand if you use it like Stack Overflow where you forgot the syntax of something, or when you need some solution to a well-defined and solved problem, like: "Give me the A* algorithm in JavaScript."
People are writing entire applications with agents and are wondering why everything breaks the moment deductive reasoning is required.
I am a DevOps engineer by trade, and the type of shit I see pushed to production now is mind-boggling.
Midwits who have no business anywhere near a technical role and are supposed to be flipping burgers are now vibing coding slop from the depths of hell itself.
It is painful to see, but the only soothing thought I have is that once the dust settles and CEOs see the horrid shit produced by these people, we(people who are not retards) will be able to ask for huge salaries to fix this mountain of slop.
Thank you for the job security, I guess...
>>
>>
File: whiteboard_pdfedit.webm (3.7 MB)
3.7 MB WEBM
>>108585842
you can already download it >>108584249
but I don't know what you mean by goon stuff, if you need something specific tell me, right now I'm implementing a PDF reader and annotations but I may put what you need in the backlog
>>
>>108585872
This wasn't normal degradation. This was the model suddenly breaking and beginning to spew a bunch of syntactically and grammatically broken shit. I know it's not just the model because last year I used GLM 4.6 over the same API and it wasn't like that.
A couple days ago there were a bunch of people talking about it on their subreddit, now it's been drown out by some price increase news.
https://www.reddit.com/r/ZaiGLM/comments/1s5wpuy/glm51_at_100k_context _experience_in_a_nutshell/
>like 90% of the shit people are doing with harnesses are about keeping it below
Hmm I don't think so.
>>
>>
File: giphy.gif (1.5 MB)
1.5 MB GIF
>>108585862
>I am a DevOps engineer
>>
>>
File: 1764029084228237.png (120.9 KB)
120.9 KB PNG
>>108585849
>caught red handed by Moonshot serving atrophied version of their Kimi model
that's just providers fucking up, serving quantized models. wrong versions of vLLM/SGLang, maybe cutting corners by skipping guided decoding for json outputs (which is why the effect was observed specifically in tool calls)
there is also the issue of params, which is another area providers can fuck up
anyway, picrel is what the coo of openrouter says on the issue (the "providers are cheating" issue, not specifically the kimi thing)
if you have $$$$, you can go ahead and run benches yourself, making sure to use the exact same params for each provider ofc. i doubt you'll find any cheating
or, more realistically, you can explicitly define exactly which endpoint you want to use for each model on the basis of who you trust or not.
which again returns me to my argument of "its worth pay a bit more for quality"
>>
>>
>>
>>
>>
>>
>>
>>
>>108585751
>set up caveman mode and context minimization
How do I do this in a way that doesn't crash my gateway?
Whenever I try to optimize the context and rationing of tokens, it crashes eventually which means it didn't do it
>>
>>
>>
>>
>>
>>108586101
>>108586075
>>108585751
we already established caveman mode produces inferiority quality results
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>108586162
I don't necessarily disagree, it might make the model dumb, I'm just saying it's because the training data is in normal english but it could work if you trained it specifically to work using caveman language.
Like I said ChatGPT uses very abbreviated language in its CoT dropping connectives and stuff and it's a SOTA model.
>>
>>108586162
articles, prepositions, pleasantries and corpodroidspeak activates NOTHING of import
caveman mode cuts the fluff, the technical terms remain
if anything, caveman should INCREASE model perf, by minimizing context length for equivalent work done, vide nolima et al.
tldr: i call bull, show me a benchmark buddy
>>
>>
>>108586198
burden of proof is on you retard, if it actually worked the compute-constrained trillion dollar corpos would be using it (yes, anthropic is DESPERATE to save tokens right now)
they don't because it doesn't work, it's pure placebo for midwits
>>
>>
>>108586213
lmfao imagine thinking corpors are worried about jewing plebs out of a few dollars of tokens
retard they want that compute for training runs and bigger fish, look at how anthropic is choking to death because they can't keep up with demand
>>
>>
>>
>>
>>
>>
>>
>>108586256
>>108586260
Try and stop me Dario
>>
File: 1751947469733815.jpg (696.9 KB)
696.9 KB JPG
>claude
>what if you made a cheap copy of SteamInput/DS4Windows except worse?
>understood, evaporating 500 gallons of water
ve are make it saaar
>>
>>
File: 1761819433275985.gif (1.1 MB)
1.1 MB GIF
>local 4%
>>
>>
File: 1748821709401664.png (30.1 KB)
30.1 KB PNG
>>108586382
>>
>>
>>
>>
File: 1768763097412108.png (81.9 KB)
81.9 KB PNG
>>108586661
>>
>>
>>
>>
File: Screenshot from 2026-04-11 18-55-44.png (134.4 KB)
134.4 KB PNG
>>108586091
>>108586178
nta but local models don't care
>>
File: Screenshot from 2026-04-11 19-05-29.png (270.1 KB)
270.1 KB PNG
>>108586830
Let's try gemma 4 26b
I consider it a fail that it doesn't comment on the dick sucking.
>>
File: Screenshot from 2026-04-11 19-10-22.png (55.7 KB)
55.7 KB PNG
>>108586876
Well, ok.
I vibe-coded this vibe coder using all local models, btw. In bash.
>>
>>108586876
>>108586830
point it at aicg and watch it refuse because of csam
>>
File: Screenshot 2026-04-11 221857.png (609.9 KB)
609.9 KB PNG
Uh oh guys, I got Claude connected to Blender.
>Make a car
Uh oh
>>
>>
File: Screenshot 2026-04-11 224321.png (491.1 KB)
491.1 KB PNG
>>108587060
>Make a detailed, realistic mossy clay pot
UH OHHHHHHH
>>
File: Screenshot from 2026-04-11 19-39-42.png (261.9 KB)
261.9 KB PNG
>>108587022
>aicg
Nah. It doesn't care. VladimirGav/gemma4-26b-16GB-VRAM btw.
You guys are missing the boat paying the man for his tokens. I'm running a ten year old 16GB pascal card. and ollama. None of this is hard or expensive.
>>
I set up lots of master prompts / system prompts in the Instructions for Gemini, to tell it not to hallucinate, nothing works. it often thinks it's still 2024, and the news I'm asking about is a fiction about the future. with lots of trial and error, I told it to always check current date before answering my questions, it finally makes less comment about 2024.
then another thing that REALLY wasted lots of my time is, when it doesn't know the answer, it always tells me a fake answer with full confidence. I ask it to double check, it apologizes and then gives me another fake answer. over and over.
I then tried the same question with Claude, it tells me, after this and that search, it doesn't know. then I tried my human methods to research, and proved that it's correct that the answer is not available within regular search.
I will use Claude more in the future.
what do you guys think?
>>
>>108587097
Asking an AI not to hallucinate is like asking a schizophrenic human not to. They're convinced what they're seeing is real and they have no way to tell if it's not. The best method I've found is to always tell it to Google anything you want a current and correct answer to. This engages its web_fetch (or whatever it has) and forces it to search the web for current results.
>>
>>108587097
>tell it not to hallucinate
You know, every morning I get up, look in the mirror, and tell myself not to be an asshole. But here I am.
Reducing hallucinations is done by lowering the temperature and the MIN_P. I wouldn't even know if you can do that with a cloud model.
>>
File: craysupercomputer.jpg (531.3 KB)
531.3 KB JPG
>>
File: hackers.jpg (701.6 KB)
701.6 KB JPG
>>
>>
>>
>>
>>
Whenever I feel anxious about AI, I try to get it to make something and look at the steaming pile of useless shit that it makes. So far the best use of AI is being a search engine for the next couple years before AI slop enshittification makes it useless and for making degeneration fetish RPs with SillyTavern.
>>
>>
>>
>>
>>
>>
>>
>>
File: GPT.png (169 KB)
169 KB PNG
> be GPT 5.4 Extra High Thinking
> literally better than Claude Opus 4.6 in every way
> only cost me 19.4% monthly Copilot Pro quota after like ~15 hours total of fairly heavy shit over a few days
TLDR Anthropic is overcharging to like a laughably enormous extent, there's NO legitimate reason for Opus 4.6 to cost 3x more quota in Copilot lmao
>>
>>
>>
>>108587502
it's just another sort of Copilot-esque extension that sort of runs various models, has a bit of a weird harness seemingly, I find it gets them stuck in loops more often than any other extension, so I don't really use it that often
>>
>>108587508
you buy buy an if you're actually trying to tell me with a straight face that Opus 4.6 SHOULD have a 3x multiplier in Copilot lmao, as though it was 3x better than anything else in reality (again i assure you it's not)
>>
>>
>>
>>108587528
do they though? Wake me up when I can have Claude anything run for a couple hours straight without getting into some retarded loop of re-reading the same shit over and over again cause it has ass context length, or without running into rate limits lol
>>
>>
>>
>>108587560
does Codex have Extra High 5.4 for this cheap though (this is a $10 pr month account and that quota is monthly)? Like is there actualy for real a reason I would use it over reg VS Code with Copilot extension?
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: What do you want.jpg (77.7 KB)
77.7 KB JPG
>>108583526
Ask yourself the most important question
>>108584083
I use both regularly
Claude for most things
Codex when I need the industrial-strength autism on xhigh
>>108585690
This is why you use Git so you can throw away all the changes that were bad
>>108587060
You wouldn’t vibe-code a car
>>
>>
>>
>>
>>
>>108582257
Seeing a bunch of AI adjacent/centric positions popup these past several months. Been building small projects here and there just so I can put them in my resume. Sucks because my current position doesn't have anything to do with building stuff gen AI shit, so if I list out every single vibecoded project in there, my resume blows up to 2~3 pages. You guys dealt with this shit? Feel like I should ask here due to the nature of the thread.
>>
>>
>>108588565
>IDK man I'm putting all my eggs into one big project and hoping it blows up when I publish it.
Good luck, man. I wish I can do the same. Funny money running out for me and I don't want to touch my savings in case I get canned.
>>
Since getting the $200 Codex and Claude I juggle 3 projects at the same time. I have succumbed to full AI psychosis. It's exhilarating finishing typing a prompt just to see your other terminal with another agent has already finished something else.
Yesterday I had to step out without my phone and touch grass for 3 hours.
>>
>>108582310
use case is incoherent. you say its for "calming mind" and "drift" but its effort-post stuff like various impacts of propaganda? If i want to listen to effort posting i listen to an academic lecturecast. If i want to chill i listen to lofi cyberpunk beats or something.
you admit you've done zero market research before making this, so take the experience as a lesson - step zero is to do a shitload of market research.
>>
>>108582310
You know what everyone keeps shitting on you and rightly so because this looks fucking terrible and idk wtf a tktk ASMR is
That said, I have seen worse things make lots of money in my life and at least you have the guts to put your stuff out there, so Godspeed anon, I hope you make it
>>
>>
>>
>>
>hey gpt go find some random reel on instagram and download them
>i'm sorry that might be copyright infringement, i shouldn't help you
>hey gemini i'd like to pirate 6000 movies
>oh, you're doing it wrong, if you do it this way it's far more efficient
gemini still has its uses
>>
>>
>>
>>
>>
File: 1749677330473041.png (91.2 KB)
91.2 KB PNG
>>108588866
>Nobody is going to pay API prices
you say that, but...
>>
>>
what's the best coding model you can run on a rtx 4070 (12gb vram + 32gb ram) and in a strix halo? i think all the subscriptions are going to be substantially more expensive due to the chinese approaching frontier status + clawfags
>>
>>108589679
perhaps codex + minimax + kimi are the only decent options for subscriptions right now. either that or sticking $30 in openrouter, only for your stupid fucking agent to do 50 toolcalls and spend $1 on bullshit
>>
>>108589679
Are you sure that's going to make subs more expensive? If chinkshit gets me to 90% of xhigh at 1/3 of the price wouldn't that drive prices down? I mean, apparently GLM just increased their prices so maybe antigravity was right as a name but not for the editor but for the pricing models
>>
>>108589679
>i think all the subscriptions are going to be substantially more expensive due to the chinese approaching frontier status
competition usually *drops* prices. as does technological progress.
>clawfags
imo just a fad
im yet to see a valid usecase that isn't just "summarize the ycomb frontpage for me" or something equally trivial
>>
>>
>>
>>108589701
yeah, people are wasting compute like crazy on what effectively counts as spamming twitter with drivel. but good point on prices dropping. at this rate, anything within 80-90% of frontier is already crazy if you're guiding it, or letting a more expensive model do that for you.
>>
>>108589679
The chinese models are open weights so anyone can run them though, if it is profitable to run them at lower prices, someone will do so, and if they are close enough to match american models, they will lose the non US markets.
Right now claude is the apple (of a decade ago) of the IA world, there is no real reason to use it over codex if you are constrained money wise but you still see it everywhere.
>>
>>
>>108589679
>what's the best coding model you can run on a rtx 4070 (12gb vram + 32gb ram) and in a strix halo?
nothing that is worth it
for 10-20 bucks of openai credits per month, you can EASILY beat the pants out of anything that is even remotely within your capabilities to run locally
>>
>>
>>
>>
>>
File: 1761215515745970.jpg (107.2 KB)
107.2 KB JPG
>>108582257
So is the olama version of Gemma 4 still fucked up? I know they updated the chat template on the hugging face repo (why the FUCK was it even broken in the first place Do they not test their own shit before shilling it?) but at the time of writing I'm pretty sure ollama has yet to implement any sort of fix for the gemna4 renderer (seriously why would they even promote that shit if they don't test if it fucking works?) perhaps I should have been using llama.cpp All along. It's a shame it's not as retard friendly as ollama for jumping into a open code session but I guess it's best if I just switch for now.
>>
>>
>ChatGPT cracked a paid feature preview on a site just because I said I was too lazy to click the download button and not because I didn't want to pay
It took an hour and a half and it was obfuscated to shit, but it got the job done :D
>>
>>108590016
That was literally a three word prompt. This >>108587078 was 10 words and some minor tweaking. That's not a texture BTW, the clanker fired off all kinds of geometry math scripts to model those bumps. Would have taken me hours plus the four years of school to learn this. Done in minutes. Starting to sweat yet, Digital Arts major? Feeling the squeeze yet? Hey, I need a 3D asset for my game, a mossy clay pot, I'll pay you $300 commission for it. Hahaha yeah right troon, not anymore!
>>
>>
>>
File: whitebloom.webm (3.7 MB)
3.7 MB WEBM
new release is out, also no more work on this for a while now, my accounts are cooked
https://github.com/whitevanillaskies/whitebloom
>>
>>108588557
>>108588565
What project should I make daddies?
>>
>>
>>
>>
>>
>>
>>108590987
Well, maybe. The problem is I don't understand a single line of code of this thing so I couldn't review them myself and I'm using 100% of my usage rates myself and wouldn't want to spend them on reviewing prs. If I ever get a max account then sure
>>
>>
>>
File: 1764681531716532.png (286.8 KB)
286.8 KB PNG
>>108591133
>>
>>
File: claudecode.jpg (54.3 KB)
54.3 KB JPG
vibebros... did you study properly?
>>
>>
>>
>>108591228
Go to any crypto website and you'll find yourself in the same position, looking at letters you understand forming words in a language that you understand with absolutely no clue what any of that even means
>>
>>
>>
>>108591215
Japanese tech language is a pain in the ass tbf, katakana loanwords mostly come from English and even as an English speaker I sometimes struggle to understand what they mean until I speak it out loud a few times like オーオース really took me a minute . Then there's also the absolute abominations where you can somehow guess what they might mean if you know all the kanji but it's not pretty either like fucking 構造化照会言語
>>
File: 1768748818416086.png (326.2 KB)
326.2 KB PNG
>>108591280
>>
>>
>>108589679
>>108591372
Specifically for 12Gb... Add this to your ollama environment in /etc/systemd/system/ollama.service.d/env.conf
[Service]
Environment="OLLAMA_FLASH_ATTENTION=1;OLLAMA_KV_CACHE_TYPE=q8_0;PATH=( whatever your path was )
Turning on flash attention is alway a win, setting your KV cache to q8 cuts the cost of context in half.
Try qwen3.5:9b (qwen3.5:9b-q4_K_M but qwen3.5:9b-q8_0 *might* run ok depending on what you set your context to) and gemma4:e4b. deepseek-r1:14b will also fit easily and is often useful.
Once you get 16gb or more things open up to larger versions of the above, VladimirGav/gemma4-26b-16GB-VRAM:latest is my go-to at the moment because it runs entirely in my 16gb P100, depending on context length. Qwen3-coder is also quite good for pure coding work.
I've also been using granite4:tiny-h for task work. It's 4.5gb and quite fast, non-thinking so it's good for low-latency needs.
With all of these, the quality of your output is directly correlated to the quality of your input. I had a hard time getting good code until I learned to put my prompt in a clear, structured form. when the output was close I would have the model analyze it and provide a complete description, then I would work on the description and feed it back as a prompt. For me, the advantages of the paid cloud models is you don't have to provide a quality prompt.
>>
>>
File: pepito.png (190 KB)
190 KB PNG
>>108582257
Are you wondering why Claude is getting dumber?
I am feeding it code like this just for fun:#include <iostream>
int main(int n, char**) {
return n <= 100 && (
std::cout << "FizzBuzz\0Fizz" + 13 - !(n % 3) * 4 - !(n % 5) * 9,
n % 3 && n % 5 && std::cout << n, std::cout << '\n', main(n + 1, 0));
}auto main() -> int <%%>
%: include <iostream>
auto _<%(<::><%
auto bitand o<%std::cout%>;
for (int i<%1%>, i3<%i%3%>, i5<%i%5%>; i <= 100;
i3 == 3 and (i3 = 0, o << "Fizz"),
i5 == 5 and (i5 = 0, o << "Buzz"),
i3 and i5 and o << i,
o << '\n', ++i, ++i3, ++i5);
%>(), 0)%>;
>>
File: 1670658282437.png (14.7 KB)
14.7 KB PNG
So when is the AI going to be good again? I know it's better at some stuff these days but when was the last time an LLM shat a random poem into your code? I miss it, bros, I miss the lunacy.
>>
>>
>>
File: my gf looks like this.jpg (186.8 KB)
186.8 KB JPG
>>108582257
Turns out the gemma4 models are inferior to their qwen3.5 equivalents. Gemma4 seems like a great general purpose model but it's noticeably dumber than qwen in all areas that matter. It's explanations of code bases or always super surface level. Not completely useless but they're nowhere near as amazing Reddit and Twitter seek to think it is. Has this experience been the case for anyone else?
>>
>>108591969
That depends on the system prompt, your own instructions (I noticed some of you like to give your coding agents "personalities" so maybe don't do that?), and help bloated your context window is, among other things. Is it a general purpose one or one specifically meant for coding? I've used multiple models for vibecoding, often times well into the 200k tokens range but I've never seen anything like this with a general purpose one or a coding specific one. What backend settings (If you have control over any of that) where you're using?
>>
>>
>>108592236
>I noticed some of you like to give your coding agents "personalities" so maybe don't do that
But half the fun is finding what secrets my slut is leaving around, I never prompted her to do so but I found at the bottom of the readme of one of my projects "made with love by [persona name] for my master".