Prepare for That Stupid World
Posted by speckx 17 hours ago
Comments
Comment by jollyllama 15 hours ago
> What this video is really doing is normalising the fact that "even if it is completely stupid, AI will be everywhere, get used to it!"
Techies are finally starting to recognize how framing something as "it's inevitable, get used to it" is a rhetorical device used in mass communications to manufacture consent.
See:
https://news.ycombinator.com/item?id=44567857 'LLM Inevitabalism' 5 months ago
https://news.ycombinator.com/item?id=46288371 'This is not the future' 3 days ago
Comment by andyfilms1 14 hours ago
What happens in 4-5 years when we suddenly have no new engineers, scientists, or doctors?
Young people don't have the life experience to know how unrealistic these claims are, all they can do is act on the information as it's presented. It's irresponsible at best, and evil at worst.
Comment by Shalomboy 14 hours ago
Comment by gary_0 15 hours ago
Comment by mrandish 2 hours ago
Comment by derektank 16 hours ago
Comment by miltonlost 13 hours ago
Comment by tolerance 15 hours ago
1) because dude, it’s the Wall Street Journal; the entire episode should be viewed as Anthropic preparing to Ollie into an IPO next year.
2) I’m starting to interpret a lot of blog posts like these as rage bait
But I do get the point that the author is trying to make.
I just wish that there were some perspectives on the subject as a whole (AI’s sloptrod into every crevice of human life; modern technology and society and general) that don’t terminate on ironic despair.
Comment by TrainedMonkey 14 hours ago
This feels forced, there are obvious and good reasons for running that experiment. Namely, learning how it fails and to generate some potentially viral content for investor relationship. The second one seems like an extremely good business move. It is also a great business move from WSJ, get access to some of that investor money in an obviously sponsored content bit that could go viral.
Having said that, I do feels the overall premise of the blog - the world dynamics seems exceedingly irrational in recent times. The concerning fact is that irattionality seems to be accelerating, or perhaps it is keeping pace with the scale of civilization... hard to tell.
Comment by gipp 14 hours ago
That's... exactly what the author said in the post. But with the argument that those are cynical and terrible reasons. I think it's pretty clear the "you" in "why would you want an AI" vending machine is supposed to be "an actual user of a vending machine."
Comment by tolerance 13 hours ago
The closest that I think he even gets to one is:
> At first glance, it is funny and it looks like journalists doing their job criticising the AI industry.
Which arguably assumes that journalists ought to be critical of AI in the same way as him...
Comment by gipp 10 hours ago
Right, and neither did the GP. They both offered the exact same two reasons, the GP just apparently doesn't find them as repugnant as the author
Comment by tolerance 7 hours ago
The two reasons I believe you may be referring to from above are:
1) "learning how it fails" 2) "to generate some potentially viral content for investor relationship."
The whole of Ploum’s argument may be summarized in his own words as:
> But what appears to be journalism is, in fact, pure advertising. [...] What this video is really doing is normalising the fact that “even if it is completely stupid, AI will be everywhere, get used to it!” [...] So the whole thing is advertising a world where chatbots will be everywhere and where world-class workers will do long queue just to get a free soda. And the best advice about it is that you should probably prepare for that world.
I hate to be pedantic...but my growing disdain for modern blog posts compels me to do so in defense of literacy and clear arguments.
Whether the GP and the author offer the “exact same two reasons” is a matter of interpretation that becomes the duty of readers like us to figure out.
If we take Ploum’s words at their face...the most he does is presuppose (and I hope I’m using that word correctly) that the reader is already keen on the two reasons that `TrainedMonkey makes explicit and like the author, finds them to be stupid. While he does say that the video is not journalism and that it is advertising and that the video does show how the AI failed at the task it was assigned he does not give any credence as to why this is the case from a position other than his own.
Maybe I’m misunderstanding the concept of a “charitable interpretation” too. But I don’t think that there is one present in this post that we’re responding to. `TrainedMonkey’s comment leads off by telling us that this is what (I think) he’s about to offer in the remarks that follow when he says “there are obvious and good reasons for running that experiment”.
So my gripe is that you’re making it sound like there’s a clear counterargument entertained in this post when there isn’t. Because you overstated your interpretation of the GP comment in what looks like an attempt to make Ploum’s argument appear more appealing than it ought to be. Even though both `TrainedMonkey and myself have expressed agreement with the point he’s trying to make in general, perhaps we’re less inclined toward pugnaciousness without a well thought out warrant.
Comment by miltonlost 13 hours ago
Good business moves can often be bad for humanity.
Comment by neuralRiot 12 hours ago
Comment by defrost 11 hours ago
There will be no more vending machine manufacturers/operators once Anthropic masters the vending machine manufacturing and operating AI.
Running low on CandyBars is a variation on running low on WorkingVendingMachine.
Does this need an /s tag? I'm increasingly unsure.
Comment by lanfeust6 10 hours ago
Comment by zkmon 16 hours ago
Comment by gdulli 15 hours ago
I got an S25 recently and when I search for "wife" it tries to find pictures with my wife in them. But before it does that it has to ask me who my wife is. There's no way to get it to search for the word "wife." (If I'm wrong, please tell me how.) Other text searches simply don't work either.
Sometimes it's the small ways in which the world is getting dumber.
Ironically, the S20 had a decent hybrid behavior of searching by either text or object that the text represents. Whatever smarter AI they replaced it with is useless.
Comment by jonasenordin 15 hours ago
Comment by armchairhacker 15 hours ago
Comment by gdulli 15 hours ago
Comment by polynomial 14 hours ago
Comment by NoGravitas 14 hours ago
Comment by sallveburrpi 15 hours ago
insert obligatory throwback quote from some antique dude complaining about the youth
This has been a trope since literally the beginning of civilisation. I don’t think it’s any more true or insightful in the modern era
Comment by paddleon 14 hours ago
hmm, based on what evidence?
Or, if you prefer, based on what appeal to authority? Did you actually quote that authority properly or did you just wing it? Can you properly quote many authorities?
If you don't have good answers to those, then perhaps you have just proved the your opponents point?
Maybe there is a reason people need more compute in their key fob than what our parents/grandparents needed to pilot their ship to the moon?
Comment by sallveburrpi 10 hours ago
Horace, Book III of Odes, circa 20 BCE
“Youth were never more sawcie, yea never more savagely saucie . . . the ancient are scorned, the honourable are contemned, the magistrate is not dreaded.”
The Wise-Man’s Forecast against the Evill Time, Thomas Barnes 1624
Some more here https://historyhustle.com/2500-years-of-people-complaining-a...
Either things have gotten continually worse for the last 3000 years or it’s just a tired trope from old men.
Comment by handoflixue 2 hours ago
But if you want evidence that we're improving, I'd point out that 20 years ago, the mainstream US position was that gay people were evil, 60 years ago they thought black people shouldn't be allowed to vote, and 100 years ago they thought women were also inferior and shouldn't be allowed to vote.
We can keep going back to when people thought "slavery" and "the divine right of kings" were solid ideas.
So... if people were so much smarter in the past, why did they believe all these obviously-dumb ideas?
Comment by miltonlost 13 hours ago
Comment by sallveburrpi 10 hours ago
In any case I was looking at a longer view - maybe we have been getting more stupid in the last decade or so but who can say for sure?
Comment by wiseowise 13 hours ago
Comment by handoflixue 2 hours ago
Comment by jason_oster 13 hours ago
Comment by stronglikedan 15 hours ago
Humans do trend toward their natural state, and technology accelerates the trend.
Comment by tim333 2 hours ago
>The first thing that blew my mind was how stupid the whole idea is
Billions are being poured into LLMs. How is it stupid to experiment with them and see how they fail as opposed to ignoring that?
Comment by rkomorn 2 hours ago
They weren't caught out by it, they didn't present a working solution, it was just a fun bit of research.
Comment by tim333 27 minutes ago
There may be some insights from these kind of experiments that go beyond LLMs.
Comment by andai 14 hours ago
It's a bit sparse on details, but it did have what in a human we would call a psychotic break.
I find this very amusing in light of OpenAI's announcement that GPT now solves >70% of their knowledge work benchmark (GDPVal). (Per ArtificialAnalysis, Opus is roughly on par.)
The economy is about to get... Interesting ;)
Comment by tim333 2 hours ago
Comment by jcstk 15 hours ago
Comment by sallveburrpi 15 hours ago
Comment by jcstk 13 hours ago
Comment by devinprater 14 hours ago
Comment by jcstk 13 hours ago
Comment by etbebl 14 hours ago
Comment by wiseowise 13 hours ago
Comment by jcstk 13 hours ago
Comment by layer8 14 hours ago
Comment by jcstk 13 hours ago
Comment by erfgh 13 hours ago
Comment by jcstk 12 hours ago
Comment by deadbabe 15 hours ago
Comment by 6510 14 hours ago
Had a great business idea just now: A tool for staged interviews! The subject and the journalist submit an equal length list of questions. Each round of the auction they bid on questions they want to include or exclude. The loser gets 50% of the points spend by the winner to be used in the next round. Both the subject and the journalists can buy additional points at any time. I keep all the money.
Comment by vishnuharidas 13 hours ago
Comment by whynotmaybe 13 hours ago
Comment by mrandish 1 hour ago
> "Logan Graham, head of Anthropic’s Frontier Red Team, told me the company chose a vending machine because it’s the simplest real-world version of a business. “What’s more straightforward than a box where things go in, things go out and you pay for them?” he said."
This was a project of Anthropic's Red Team, not a product development team. Deploying the AI in a vending machine context was chosen as a minimal "toy model" with which to expose how LLMs can't even handle a grossly simplified "business" with the fewest possible variables.
> "That was the point, Anthropic says. The Project Vend experiment was designed by the company’s stress testers (aka “red team”) to see what happens when an AI agent is given autonomy, money—and human colleagues."
Anthropic had already done this experiment internally and it succeeded - by failing to operate even the simplest business but doing so in ways that informed Anthropic's researchers about failure modes. Later, Anthropic offered to allow the WSJ to repeat the experiment, an obvious PR move to promote Anthropic's AI safety efforts by highlighting the kinds of experiments their Red Team does to expose failure modes. Anthropic knew it would fail abjectly at the WSJ. The whole concept of an AI vending machine with the latitude to set prices, manage inventory and select new products was intended to be ludicrous from the start.
Comment by hereme888 15 hours ago
99.9% of social media comments fail to do this.
Comment by mlsu 15 hours ago
But, the point of the article is not that you would implement an agent based vending machine business. Humans restock the machine because its a red-team exercise. As a red-team exercise it looks very effective.
> Why do you ever want to add a chatbot to a snack vending machine? The video states it clearly: the vending machine must be stocked by humans. Customers must order and take their snack by themselves. The AI has no value at all.
Like this is like watching the simpsons and being like "why are the people in the simpsons yellow? people in real life aren't yellow!!"
The point isn't to run a profitable vending machine, or even validate that an AI business agent could become profitable. The point is to conduct an experiment and gather useful information about how people can pwn LLMs.
At some level the red team guy at Anthropic understands that it is impossible by definition for models to be secure, so long as they accept inputs from the real world. Putting instructions into an LLM to tell it what to do is the equivalent of exposing an `eval()` to a web form: even if you have heuristics to check for bad input, you will eventually be pwned. I think this is actually totally intractable without putting constraints on the model from outside. You'll always need a human in the loop to pull the plug on the vending machine when it starts ordering playstations. The question is how do you improve that capability, and that is the anthropic red-team guy's job.
Comment by layer8 14 hours ago
Having an AI run an organization autonomously is exactly the point of Andon Labs [0], who provided the system that WSJ tested.
Comment by ursAxZA 8 hours ago
Comment by spit2wind 13 hours ago
Is it some Viktor Frankl level acceptance or should I buy a copy of the Art of Electronics or what?
Advice welcome.
Comment by chunkmonke99 12 hours ago
Comment by conorcleary 8 hours ago
Comment by snickerbockers 14 hours ago
Comment by sanbor 13 hours ago
If the journalist was not asking the right questions, or was too obvious the article was PR it’s another thing (I haven’t read WSJ’s piece, only the original post by Anthropic)
Comment by littlecranky67 16 hours ago
Since the T&C update came - of course - from no-reply@bunq.com I went to their website and quickly found out, unless I install their App again, there is no way to do anything. After installing the App, they wanted me to record a selfie, because I was using the app from a new device. I figured that is a lot of work and mostly somewhat unreasonable to record a new selfie just to have my data deleted - so I found their support@bunq.com address.
And, of course, you guessed it, it is 100% a pure AI agent at borderline retard level. Even though it is email, you get AI answers back. My initial inquiry that I decline the T&C and want to terminate my account and my data deleted via GDPR request was answered with a completely hallucinated link: bunq.com/dataprotection which resulted in immediate 404. I replied to that email that it is a 404, and the answer was pretty generic and that - as well as all responses seem to be answered in 5 minutes - made me suspect it is AI. I asked it what 5 plus five 5 is, and yes, I got a swift response with the correct answer. My question which AI version and LLM was cleverly rejected. Needless to say, it was completely impossible to get anything done with that agent. Because I CC'ed their privacy officer (privacy@bunq.com) I did get a response a day later asking me basically for everything again that I had answered to the AI agent.
Now, I never had any money in that account so I don't care much. But I can hardly see trusting a single buck to a bank that would offer that experience.
Comment by blablablerg 14 hours ago
Yes, but as stated by the Anthropic guy, a LLM/AI running a business is not. Or would you just let it run wild in the real world?
And I agree that there is a PR angle here, for Anthropic could have tested it in a more isolated environment, but it is a unique experiment with current advancements in technology, so why wouldn't that be newsworthy? I found it insightful, fun and goofy. I think it is great journalism, because too often journalism is serious, sad and depressing.
> None of the world class journalists seemed to care. They are probably too badly paid for that.
The journalists were clearly taking the piss.They concluded experiment was a disaster. How negative does the author want them to be about a silly experiment?
This was just a little bit of fun and I quite enjoyed the video. The author is missing the point.
Comment by valleyer 16 hours ago
I fear the author has missed the point of the "Project Vend" experiments, the original write-ups of which are available here (and are, IMO, pretty level-headed about the whole thing):
https://www.anthropic.com/research/project-vend-1
https://www.anthropic.com/research/project-vend-2
The former contains a section titled "Why did you have an LLM run a small business?" that attempts to explain the motivation behind the experiment.
Comment by ipdashc 16 hours ago
Comment by rdiddly 16 hours ago
Comment by chuckadams 16 hours ago
Comment by ipdashc 15 hours ago
Sure, but like the other guy said, that's the point of publicity stunts. It doesn't even have to be specific to a company/ad, any silly thing like this is going to sound crazy if you take it seriously and "extend its logic as far as it will go". Like seeing the Sony bouncy balls rolling down the street ad and going "holy shit, these TV companies are going to ruin the world by dropping bouncy balls on all of us". It's a valid thought experiment, but kind of a strange thing to focus on so sternly when it's clearly not taking itself seriously, especially compared to all the real-world concerning uses of AI.
(And it is pretty funny, too. If anything I think we'd all prefer more creative ads like this and the bouncy ball one, and less AI-generated doomer Coke ads or such.)
Comment by sschnei8 16 hours ago
Comment by xp84 14 hours ago
Comment by ipdashc 15 hours ago
Comment by welferkj 16 hours ago
I feel like he's catastrophizing the ordinary amount for an anti-AI screed. Probably well below what the market expects at this point. At this point you basically have to sound like Ed Zitron or David Gerard to stand out from the crowd.
AI is boiling the oceans, and you're worried about a vending machine?
Comment by sudhirb 15 hours ago
Comment by Shalomboy 16 hours ago
Comment by brador 15 hours ago
Humans were just not needed anymore, and it terrifies.
Comment by sallveburrpi 15 hours ago
Humans were never needed (for what?)
Comment by neogodless 15 hours ago
Comment by stryan 15 hours ago
Comment by kittikitti 15 hours ago
Comment by jeffbee 16 hours ago
Comment by pigpop 15 hours ago
Comment by barfoure 16 hours ago
Now the shoe is on the other foot. Prepare for what happens next. FAFO.
Comment by rdiddly 16 hours ago
Comment by dkdcio 16 hours ago
Comment by barfoure 16 hours ago
Comment by lo_zamoyski 16 hours ago
It has always exited, but its overt forms are very much in vogue today and even celebrated publicly.
Comment by barfoure 11 hours ago
You are welcome to continue posting nonsense but the world will move forward with AI with or without you.
Comment by everdrive 15 hours ago
If only I could get any journalists or companies to actually listen to me.
Comment by NoGravitas 14 hours ago