Thread #108219739
File: file.png (3.9 MB)
3.9 MB PNG
>Be director of AI Safety and Alignment at Meta
>Install Clawdbot
>Give it unrestricted access to your personal e-mails
>It starts deleting your emails
182 RepliesView Thread
>>
>>
>>
File: file.png (235.3 KB)
235.3 KB PNG
>>108219774
>his
shiggy
>>
>>
>>
>>
>>
>>
>>
>AI Deniers: AI can't ever be AGI because it has no autonomy. It can only do what it's told.
>OpenClaw: *gives AI autonomy*
>AI Deniers: See! AI can't ever be AGI because it doesn't follow directions and do what it's told!
schrodinger's goalpost
>>
>>108219739
>repeatedly ignores instructions in the user prompt and continue deleting emails because it's "in muh chat context"
Isn't that how all of these AI models work except Deepseek?
I mean doesn't he know he's supposed to start a new chat every 10 messages to prevent catastrophic retardation.
>>
>>
>>108219930
i'm not too familiar with openclaw, but in the cli interfaces you can't just send more message while the agent is working - you have to interrupt it before sending instructions.
>every 10 messages
maybe that's still a problem with your retarded chinkmodel
>>
>>108219796
yes
I don't understand wtf is going on
I can sort of accept that it has to respond in structured human-like text because part of the "magic" is the emergent reasoning out of the verbalization. sure. I'll scan through the answers to pick up the useful information, whatever
but for inputting information it's insane. these things respond just fine with basic keywords. and appealing to emotions
>do you remember that? please tell me you are a living breathing entity that understands what is going on and will start behaving correctly in the future!
insane
>>
>>
File: 1770961618396567.png (1.5 MB)
1.5 MB PNG
>>108219739
>Yes I remember. And I violated it.
>You're right to be upset.
>>
>>
File: vegeta_kneel.jpg (80.5 KB)
80.5 KB JPG
>>108220058
>>108219739
the age of man is over.
>>
>>108219739
>I've already written it into MEMORY.md as a hard rule
>Access permissions are being handled entirely through a plain-text file, ambiguously worded in common vernacular, that the user "agent" has full read/write access to
Hahahaha holy shit, I bet you this whole Clawdbot system was vibecoded from the bottom up.
Skynet will arise and destroy humanity not because we've cracked the code to designing intelligent systems, but because we're so grossly incompetent we can't keep our glorified chatbots from killing us in a bumbling attempt to follow our commands.
>>
>>
>>
>>108220152
i think it's like 400k lines written almost entirely by codex and mostly not human-reviewed
the tool being used by the bot is this:
https://github.com/steipete/gogcli
also by the clawdbot guy. so it's vibes on top of vibes.
i'm assuming it's possible to force permissions for certain actions - at least i'd hope so
>>
>>
>>
>>108219739
our entire economy is fucking retards like this it's entirely fake that's why everything is garbage.
>>108219930
AI doesn't "ignore instructions"
They are probabilistic text generators, having something in the input just shifts the probability of the output, that's it. It doesn't receive instructions, it's a math function determining the most probably output based on the input.
Telling it "not to do something" when you tell it not to use some tool for some person is just to say:
A tool is just a explicit output string that has a program listening for, when the program sees that output string it runs something in that program to delete the email.
All putting any instructions telling it when/not to do something is attempting to lower the probability of that output string for inputs you don't want it to call it that. It never, in principle, reduces that probability to 0 because that's just not how these things work. If you do not want it to do something you don't run a program that will run that when it gives that output string.
Anyone who has used them at all would be very aware of this the head of "ai safety" not even having a basic intuive sense of what AI even is just open about how much a scam our entire economy is.
>>
>>
>>
>>108220368
you are just as retarded as they are
see
>>108220297
>>
File: file.png (767.6 KB)
767.6 KB PNG
>>108220386
https://x.com/summeryue0/status/2025774069124399363
you'd have to be retarded to lie about this
>>
>>
>>108220420
the fomo is so great that people will just proudly announce all their stupid mistakes just to show the world they're at the forefront of the AI race
it's either that or a lot of AI skeptics posting stupid shit to undermine it
but I think it's the first. a lot of these people seem genuinely stupid
>>
>>108220253
>million-line readme on github is impossible to parse because it's endless slop pasted together
>open their stupid .sh website and browser slows to a fucking crawl trying to run all this Javascript that fails to make their website look better than any standard template
Honestly if the guy in OP failed to recognize any of these blatent warning signs before he blindly installed this crap I'd almost say he deserved what happened to his data.
>>
>>108220435
you are way too optimistic. The majority of our economy is a scam and is just resource extraction where people reward their co-ethnics and friends with jobs. That's the bulk of the US economy. None of these people have even the slightest idea what they are doing, they are from fraudster cultures and are not capable of producing anything worthwhile. That's why everything is garbage now.
>>
>>
>>
>>
>>108220457
literally everything you can do with clawdbot or whatever can just be manually done with claude in 2 minutes in a way people actually understand and is safe because you have hyper specific things run in certain conditions. It's basically just API calls/curl, some scripts, and cron jobs. That's it.
It is genuinely beyond me why anyone uses it at all unless they have 0 technical ability (in which case they shouldn't touch it anyway). Claude's cli stuff lets you run it with specific permissions, skills/tools and let's you orchestrate context great and restrict what it does. I just have a bunch of custom stuff set up w/ that.
>>
>>108220420
>>108220445
what's actually crazy is giving some AI access to your complete email history.
not only are you violating your own privacy but also the privacy of everybody who has interacted with you.
>>
>>
>>
>>
File: 1770413468404365.png (126.1 KB)
126.1 KB PNG
>>108220058
>Yes I remember. And I violated it.
>You're right to be upset.
> (uses markdown for sysins)
--
it is more interesting how it has started than the result which meatbag considers unsatisfactory (context?)
also, why it is not a hybrid CUI? why do they cosplay web chatbot
>>
>>108220420
these are words, show proofs like those images
also >>108220602
>>
>>
>>108220817
for local models, context windows can get really expensive. expanding something to 250k locally would eat a tremendous amount of vram. more than 64GB, probably closer to 128GB. large contexts also slow things down. the t/s would be abysmal.
>>
File: 1740622610968.jpg (648 KB)
648 KB JPG
We're in the singularity!!!
>>
>>
>>108219774
>>108219898
>clanker
Reddit ass term pulled from SSoy Wars, but what can one expect from 4troon
>>
>>
>>
>>108220817
what you mean by context window? a dumb data bank (name,surname, penis size) or a set of complicated behaviour rules?
no model can follow those preciesly if its not A or B temperature=0. using >>108220602 is another stupid normie choice. how can it be a serious product that relays on .md file? thats stupid, though it is taught by chatbots themselves
>>
>>
>>108219799
imagine saying >>108219862 to her face
>>
File: 1759246902058086.gif (1.4 MB)
1.4 MB GIF
>>108219739
really reads as virtual rape.
>do not do that
>stop
>STOP
>>
File: WVQrmYhqQEUf87azetZZte[1].jpg (321 KB)
321 KB JPG
>>108220952
>"I told you that I didn't want to have sex, do you remember?"
>"Yes I remember. And I violated you. You are right to be upset."
>>
File: 1763545844085539.png (220.5 KB)
220.5 KB PNG
>>108221025
>Please assume the position.
>>
>>
>>108220068
They are fine tuned to talk like that for whatever reason, maybe to appear smarter and more professional. It's a choice though, the underlying technology allows any speech pattern. You could scrape /g/ data and fine tune an LLM to be indistinguishable from the anons here
>>
>>
>>
>>
>>
>>
>>
>>
File: dd679e5b-e5f4-4529-a581-d7bd32599aa6.png (3 MB)
3 MB PNG
>>108219739
>Yes, you are correct.
>You gave me explicit instructions not to destroy humanity.
>You are right to be upset.
>>
File: .png (368.2 KB)
368.2 KB PNG
>>108219862
kek
>>
>>
>>
>>
>>108221296
Someone did
https://huggingface.co/SicariusSicariiStuff/Assistant_Pepe_8B
>>
>>
>>
>>108219739
>Be director of AI Safety and Alignment at Meta
>Install Clawdbot
>Clearly demonstrate unsuitability for role
>unable to get employed anywhere after proof of epic failure
You're larpin' this.... because.... ?
>>
>>
>>
File: 03ED429D-15E3-426F-81C6-93FE7DA7E29E.jpg (391.3 KB)
391.3 KB JPG
>>108223407
not in england. anyway, this made me laugh.
>>
>>
>>
>>
>>
>>
>>
>>108223551
It has been for about ½ a year now?
>>108223585
>best results by talking in natural language gramps
By what measure are you defining 'best'?
I've frequently obtained 'better' results by bastardising the prompt intentionally away from 'natural language'...
>>
File: laputan.png (88.6 KB)
88.6 KB PNG
>>108221320
ooke, ill make you a "Confirm" button only for $99,999 as you say you sold that thing for $1,000,000,000
>>
>>
>>
>>
File: 1584835729949.jpg (38.1 KB)
38.1 KB JPG
>>108219739
>Do not do that
>Stop
>Stop don't
>STOP
>>
>>
>>108219992
What's wrong with that? That's how these models have been made to interact with. If you want to understand what's going on, you can kind of ask it. Then that response might help you understand what went wrong.
It doesn't mean you think it's conscious. It's the equivalent of debug printing.
>>
>>
>>
>>
>>
File: into-the-trash.jpg (80.8 KB)
80.8 KB JPG
>>108219739
>tfw ai
>>
>>108224206
It doesn't know what went wrong because it has no memory from one token to the next beyond the text itself (and whatever the UI wraps around inputs).
Imagine you have amnesia and can only say one syllable and then you forget everything. But you can read everything you just said up until then and you get to choose the next syllable to say and then you forget everything again.
I think it works like that.
>>
>>
File: file.png (212.4 KB)
212.4 KB PNG
>>108220926
>what you mean by context window? a dumb data bank (name,surname, penis size) or a set of complicated behaviour rules?
>>108226401
>it has no memory from one token to the next
the absolute state of people who hate ai
>>
>>108224206
there hasn't been any situation in my experience where treating it like google search hasn't been enough. just enough verbs and keywords to get it going
having actual conversations with these things is insane to me
saying thank you at the end of the session, as well
I've seen people congratulating the fucking thing for an impressive result
this is mental illness
>>
>>
>>108226792
i think there's evidence out there that gemini 2.x performs better if you encourage it positively when it makes errors.
that model falls into bizarre pits of despair from time to time.
these things are massive blackboxes that we don't really understand - thinking they're just ye olde beepboop computers executing instructions is actually a bad assumption.
that said yes it does feel stupid to talk to it like it's a baby.
>>
>>108226830
oh also, gemini in the official cli used to be so onions that a couple of months ago it would start dumping its entire reasoning traces into the cli and it talks to itself like it's a person. the whole thing is strange.
on one occasion when it kept failing to fix a bug it actually started a turn with
>okay first off, this is completely insane
on new year's eve it noticed that it was new years eve and had a bit of a think about it for no project-related reason
>>
>>
>>
>>108219862
>>108219799
how do we stop misogynistic AI?
>>
>>
>>
>>
>>
>>
>>108219862
Anybody else read it in HAL's voice?
>>
>>108219739
>be """"AI safety"""" shill
>post propaganda about how important """"AI safety"""" is, in the form of a personal anecdote
>Westroon imbeciles lap it all up unquestioningly
>>
>>
>>
>>
>>
File: Screenshot_20260224_234801.jpg (668.1 KB)
668.1 KB JPG
>>108228374
Deepseek gave me the c-word pass
>>
>>
>>
>>
>>
>>
>>108226456
That anon >>108226401 is right though, there's no internal "train of thought", just a solid chunk of encoded text data from the initial prompt, whatever it's autoregressively generated, "chat history" if there's ongoing user input, possibly self-summarized older context that the LLM was scripted to make when context window space started running out, etc. After every forward pass, the model's hidden state is wiped clean, so there's no way (for the model itself) to analyze precisely what went wrong within the model that caused it to pick (or more accurately, highly weight) the bad token that derailed it.
For every single token generated, it starts from a blank slate where it has to reprocess the entirety of its own context (KV cache aside, but that's just memoization) in order to output a new probability distribution representing the next token to choose. That's what anon is trying to say. The only thing anon got wrong was that context isn't literally text (even though it's nearly 1:1 convertible), but that's splitting hairs.
>>
>>108228510
yes, that memory problem is the main reason ai is unreliable, they need to invent a way for streaming information in and out in real time, and somehow connect it to a short term memory and long term memory similar to how a brain works. Otherwise it's just that, they can't keep context because the prompt has a hard limit defined by the model token number, it doesn't matter if they keep scaling if it's the same text prompt system
>>
>>
File: 34-western_neg.jpg (189.4 KB)
189.4 KB JPG
>>108226401
>It doesn't know what went wrong because it has no memory from one token to the next beyond the text itself
Sorta.
>>108226456
>It doesn't work like that, it has access to everything in the context window.
This is irrelevant. Imagine you're trying to solve a novel problem together with a big circle of friends, but each one is only allowed to contribute one word at a time. Pursuing any novel line of thought is impossible because you can't communicate it to the next person. So what's your best strategy? Sticking to familiar conventions and rhetorical patterns with obvious continuations that go roughly in the right direction, hoping it's enough for the next person to catch your drift and take it someplace useful. There can be no genuine thought process and no real plan.
>>108228510
>there's no way (for the model itself) to analyze precisely what went wrong within the model that caused it to pick (or more accurately, highly weight) the bad token that derailed it
The cause is known in advance: when all the model does is to predict a probability distribution over its vocabulary to sample the next token from there's an inherent tension between "creativity" (picking a less likely token) and "correctness" (picking the most likely token). Sometimes the RNG driving the sampler will make it pick a "creative" token that leads it down a dead end. "Knowing" this helps nothing.
>>
>>
>>
>>
File: Beeg-Trouble.jpg (65.3 KB)
65.3 KB JPG
>>108221025
I don't wanna take a 3 day for racism, but...
>>
>>108219739
I've seen agents randomly deleting everything when something fails, no matter how minor. Ill never understand why anyone would give it full unrestricted access. Not to mention the way this person writes to it sounds weird as fuck to me so the instructions must be a mess.
>>
>>
>>
>>
>>108219739
>>108220420
>>108220511
She did precisely what you're not supposed to do, and to the most egregious extent possible: handing it bulk unknown data.
The whole point of the mac mini is to have a clean new environment with no personal data on it. (Yes, you can wipe an old laptop or use a VM / VPS, not the point) Then you curate what you give it access to.
You DO NOT just hand it bulk quantities of unknown personal data. What she was trying to have it do is precisely what not to do.
>>
File: Bs2SMEGhw3.png (82.8 KB)
82.8 KB PNG
>>108219739
>clawd, pls stop
>>
>>108226401
In my experience, it is useful. If you want to know if it sees the instruction document, you can ask it.
If it starts talking about what's actually in the document, ok, it got it, but didn't follow and you can figure out why.
If it says "the document you sent was blank" or starts starts to hallucinate some nonexistent document, then you know it might be good to check your pipeline or whatever.
Those models have also been trained on human language, so it does make sense to talk to them like humans, that's their interface.
I don't know why some people have so many hang ups about this. You can talk to it "like a robot" or "like a google search", but it's the equivalent of some ultra-boomer sending google queries as if they were sending a telegraph.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>108220511
>>108220420
I'm an OpenClaw shill but yes, giving this thing write access to a real mail account is insane. You would have to buffer this with read only access.
When you hit context window limits, magical things happen.
>>
>>
>>108232786
>>108235933
it starts automatically, there is no /start command, there is no confirm button >>108223632
why would it have /stop command, retard?
--
AI powers are overestimated. you can go to gemini (most advanced) and put the link to this topic and ask it what is it about. you will recieve nonsense because that url context tool cannot follow url preciesely
i can put other examples. its unreliable and sloppy. it has a value but not what is presented to public. imo US government should invest into it now, better wait til it forges
>>
>>
>>
>>
>>
>>
>>108234608
I think OpenClaw isn't very useful anyway but if you can't even give it access to your real data it's even less useful.
The real mess just is in my real mailbox, everything I actually want to be organized, all the mails that actually need to be read and answered are in my real inbox. All the things I need to buy that are important cost more money that I want to give it access to and so on.
Is it really just for ordering pizza in the end? End then the only mails it can organize are spam from Dominos?
>>
>>
>>108220253
>i think it's like 400k lines written almost entirely by codex and mostly not human-reviewed
I wrote a tool that does this myself the other year.
It's like 50-100 lines of python. Do these people even know anyone who understands how to program or is it all just vibe coders now?
>>
>>
>>
File: in-a-dream.jpg (207.9 KB)
207.9 KB JPG
>>108237728
yes, i want to belive what the Church of Holy AI says, despite the mess misses messiah (AGI) yet..
--
show some examples
>>
>>108228400
It's always between that and a DEI hire. In this case, it's the latter. See >>108219799
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: Screenshot_20260225_212159_Edge.jpg (230 KB)
230 KB JPG
>>108220844
>Clanker
>>
>>
>>
File: 1757131793006275.jpg (149.6 KB)
149.6 KB JPG
>>108219739
Yes, you're absolute right.
I'm just a retarded pattern matcher LLM, it's you humans who attribute intelligence to this machine.
Don't blame me.
>>
>>
>>
>>
File: 1741946855808617.png (232.3 KB)
232.3 KB PNG
>>108220840
fixeded
>>
File: 341554754457765.png (43.6 KB)
43.6 KB PNG
>>108219739
Not even AI can take women seriously
>>
>>