Vibe coding kills open source
Posted by kgwgk 3 days ago
Comments
Comment by WarmWash 3 days ago
The future will absolutely not be "How things are today + LLMs"
The paradigm now for software is "build a tool shed/garage/barn/warehouse full of as much capability for as many uses possible" but when LLMs can build you a custom(!) hammer or saw in a few minutes, why go to the shed?
Comment by anticorporate 3 days ago
Sure, there will be more personalized apps for those who have a lot of expertise in a domain and gain value from building something that supports their specific workflow. For the vast majority of the population, and the vast majority of use cases, this will not happen. I'm not about to give up the decades of experience I've gained with my tools for something I vibe coded in a weekend.
Comment by tracker1 3 days ago
Comment by anticorporate 3 days ago
This isn't a defense of enterprise applications, though. They're more like a shed fully of rusty tools with a thirty different coping saws blades and not a single handle because corporate policy only allows for you to have a handle if Joe from accounting says you can, but why would he when his VP confidently said you can just hold the blade between your fingers.
Comment by tracker1 3 days ago
Comment by seniorThrowaway 3 days ago
Comment by spamizbad 3 days ago
I've been at this long enough to see that today's best practices are tomorrow's anti-patterns. We have not, in fact, perfected the creation of software. And the your practices will evolve not just with the technology you use but the problem domains you're in.
I don't mean this as an argument against LLMs or vibe coding. Just that you're always going to need a fresh corpus to train them on to keep them current... and if the pool of expertly written code dries up, models will begin to stagnate.
Comment by seniorThrowaway 3 days ago
Comment by thfuran 3 days ago
Comment by suddenlybananas 3 days ago
Comment by seniorThrowaway 3 days ago
Comment by maleldil 3 days ago
Comment by Bishonen88 3 days ago
"Create a copy of xyz. It needs to look and behave similarly. I want these features ... And on top of that ...". Millions decisions not needed. A handful of vague descriptions of what one wants is all it takes today. I think claude and co. can even take in screenshots.
Documentation won't be needed either IMO. Since humans won't write nor read the code. They will simply ask LLM's if they have a question.
I totally am giving up my experience with various paid SaaS this year, which I was paying for last years. Not only am I able to add the features that I was wishing for those tools to have (and would have never made it into the real app because they're niche requests), but am saving money at the same time.
And the above is just whats happening today. Claude Code is younger than 1 year old. Looking forward to come back to this thread in a year and swallow my words... but I'm afraid I won't have to.
Comment by dimitri-vs 3 days ago
"Create a copy of Amazon.com"
ok, how did you want to handle 3pl fulfilment and international red tape?
"No not that complicated, a minimal copy"
How minimal? How many servers should I provision? How vertically integrated should we get?
Etc.
I really want to see someone build an app of any value with minimal decisions made.
Comment by acessoproibido 3 days ago
You could easily replicate the store part of it minimally, at its core its just an index of products, a basket and checkout system. There are other parts that make up the whole thing of course.
There is a lot of room between no value and trillion dollar company
Comment by player1234 3 days ago
Comment by nerdsniper 3 days ago
Comment by generallyjosh 2 days ago
You can ask it about the code, sure, and it'll try to tell you how it works. But, what if there's a bug in the code? Maybe the LLM will guess at how it was supposed to work, or maybe it'll start making stuff up to justify the bug's existence (it's actually a hidden feature!)
The docs say how the code should work. For an LLM that has to go relearn everything about your code base every time you invoke it, that's vitally important
Comment by digiown 3 days ago
As the saying goes, 80% of users only use 20% of the features of your program, but they are different 20% parts. When the user vibecode the program instead, only their specific 20% needs to be implemented.
Comment by theshrike79 1 day ago
It's so standard that the usual paradigm is that your company will adapt itself to the way SAP works, not the other way around. Massive gigantic corporations have tried to adapt SAP and failed. IIRC Lidl had a very expensive high-profile failure in this.
Being standardised isn't always a good thing.
Comment by iknowSFR 3 days ago
Every problem or concern you raise will adapt to the next world because those things are valuable. These concerns are temporary, not permanent.
Comment by blibble 3 days ago
I really, really don't care
I didn't get into programming for the money, it's just been a nice bonus
Comment by frizlab 3 days ago
Exactly the same for me! If kind of feel like an artist whose paintings are worth more more easily than a paint or music artist… But boy would I be poor if this art were worthless!
Comment by shimman 3 days ago
Comment by II2II 3 days ago
Because software developers typically understand how to implement a solution to problem better than the client. If they don't have enough details to implement a solution, they will ask the client for details. If the developer decides to use an LLM to implement a solution, they have the ability to assess the end product.
The problem is software developers cost money. A developer using an LLM may reduce the cost of development, but it is doubtful that the reduction in cost will be sufficient to justify personalized applications in many cases. Most of the cases where it would justify the cost would likely be in domains where custom software is in common use anyhow.
Sure, you will see a few people using LLMs to develop personalized software for themselves. Yet these will be people who understand how to specify the problem they are trying to solve clearly, will have the patience to handle the quirks and bugs in the software they create, and may even enjoy the process. You may even have a few small and medium sized businesses hiring developers who use LLMs to create custom software. But I don't think you're going to see the wholesale adoption of personalized software.
And that only considers the ability of people to specify the problem they are trying to solve. There are other considerations, such as interoperability. We live in a networked world after all, and interoperability was important even before everything was networked.
Comment by aspenmartin 3 days ago
Why do you think agents can’t do that? They can’t do this really well today but if the distance we went in 2025 stays similar it’ll be like a year before this starts getting decent and then like another 1 year before it’s excellent.
> Sure, you will see a few people using LLMs to develop personalized software for themselves. Yet these will be people who understand how to specify the problem they are trying to solve clearly, will have the patience to handle the quirks and bugs in the software they create
Only humans can do this?
Comment by dimitri-vs 3 days ago
You basically need AGI and we are nowhere close to AGI.
Comment by aspenmartin 3 days ago
All of the issues you talk about: they aren’t solved but we’ve made amazing progress on all of them. Continual learning is a big one and labs are likely close to some POCs.
Token costs per unit performance rapidly goes down. GPT4 level perf costs you 10x less today than two years ago. This will continue to be the case as we just continually push efficiency up.
The AGI question “are we close” tbh to me these questions are just rabbit holes and bait for flame wars because no one can decide on what it means and then even if you do (e.g. super human perf on all economically viable tasks is maybe more of a solid staring point) everyone fights about the ecological validity of evals.
All I’m saying is: taking coding in a complete vacuum, we’re very very close to being at a point where it becomes so obviously beneficial and failure rates for many things fall below the critical thresholds that automating even the things people say make engineers unique (working with people to navigate ambiguous issues that they aren’t able to articulate well, making the right tradeoffs, etc) starts looking like less of a research challenge and more of an exercise in deployment
Comment by nicbou 2 days ago
My experience with software development is that a lot of things are learned by asking probing questions informed by your experience. How would an LLM understand the subtle political context behind a requirement or navigate unwritten rules discussed in meetings a year ago?
Comment by pianopatrick 3 days ago
but LLMs are chat bots. Surely eventually someone will set up an LLM based coding system that can ask architecture and design questions before starting to code.
Comment by palmotea 3 days ago
1) Your specific analogy is kinda missing something important: I don't want my tools working differently every time I use them, also it's work to use LLMs. A hammer is kind of a too-simple example, but going with it anyway: when I need a hammer, I don't want my "LLM" generating a plastic one, then having to iterate for 30 minutes to get it right. It takes me far less than 30 minutes to go to my shed. A better example is would be a UI, even if it was perfect, do you want all the buttons and menus to be different every time you use the tool? Because you generate a new one each time instead of "going to the shed"?
2) Then there's the question, can an LLM actually build, or does it just regurgitate? A hammer is an extremely we'll understood tool, that's been refined over centuries, so I think an LLM could do a pretty good job with one. There are lots of examples, but that also means the designs the LLM is referencing are probably better than the LLM's output. And then for things not like that, more unique, can the LLM even do it at all or with a reasonable amount of effort?
I think there's a modern phenomenon where making things "easier" actually results in worse outcomes, a degraded typical state vs. the previous status quo, because it turns what was once a necessity into a question of personal discipline. And it turns out when you remove necessity, a lot of people have a real hard time doing the best thing on discipline alone. LLMs might just enable more of those degenerate outcomes: everyone's using "custom" LLM generated tools all the time, but they all actually suck and are worse than if we just put that effort into designing the tools manually.
Comment by tracker1 3 days ago
Seriously though, you can tell AI what libraries and conventions you want to follow... that's been a lot of what I've done with it recently... I've been relatively pleased with the results.
I've said several times that it's not perfect, but it is an overall force multiplier. It's much like working disconnected with an overseas dev team, but you get turn around in minutes instead of the next morning in your email. The better instructions/specs you give, the better the results. On my best day, I got about 3 weeks of what would take me alone done, after about 3 hours of planning/designing and another 2-3 hours of iteration with Claude Code. On my worst day, it was frustrating and it would have been about the same amount of time doing it myself. On average, I'd say I get close to 5 days of work done in 5-6 hours of AI assisted coding. Purely anecdotally.
That said, I usually have a technical mind for how I want the solution structured as well as features and how those features work... often it clashes with the AI approach and sometimes it swims nicely. I'll also say that not all AI coding is the same or even close in terms of usefulness.
Comment by jerf 3 days ago
A good question but there's a good answer: Debugged and tested code.
And by that, I mean the FULL spectrum of debugging and testing. Not just unit tests, not even just integration tests, but, is there a user that found this useful? At all? How many users? How many use cases? How hard has it been subjected to the blows of the real world?
As AI makes some of the other issues less important, the ones that remain become more important. It is completely impossible to ask an LLM to produce a code base that has been used by millions of people for five years. Such things will still have value.
The idea that the near-future is an AI powered wonderland of everyone getting custom bespoke code that does exactly what they want and everything is peachy is overlooking this problem. Even a (weakly) superhuman AI can't necessarily anticipate what the real world may do to a code base. Even if I can get an AI to make a bespoke photo editor, someone else's AI photo editor that has seen millions of person-years of usage is going to have advantages over my custom one that was just born.
Of course not all code is like this. There is a lot of low-consequence, one-off code, with all the properties we're familiar with on that front, like, there are no security issues because only I will run this, bugs are of no consequence because it's only ever going to be run across this exact data set that never exposes them (e.g., the vast, vast array of bash scripts that will technically do something wrong with spaces in filenames but ran just fine because there weren't any). LLMs are great for that and unquestionably will get better.
However there will still be great value in software that has been tested from top to bottom, for suitability, for solving the problem, not just raw basic unit tests but for surviving contact with the real world for millions/billions/trillions of hours. In fact the value of this may even go up in a world suddenly oversupplied with the little stuff. You can get a custom hammer but you can't get a custom hammer that has been tested in the fire of extensive real-world use, by definition.
Comment by rurp 3 days ago
I simply can't imagine a world where these same people all decide they constantly want to learn a completely unique UX for whatever piece of software they want to use.
Comment by WarmWash 3 days ago
Users will not fumble with the complex web of nested settings that engineers wet dream about.
But they will tell the LLM "I'd really like it if the tool bar only had the hammer and saw tools", and it will be done.
I cannot see software going in any other direction than a blank front end that users prompt LLMs to run scripts on top of.
Picture MS Word where the GUI is just a page and a sidebar for telling an LLM what you want it to do. And if it's not possible, the LLM could even write extensions and plugins that make it possible.
Software is going to completely change.
Comment by charcircuit 3 days ago
Comment by dimitri-vs 3 days ago
Done. And it seems absolutely awful.
"Please bold the text I have selected" instead of a preexisting bold button.
Oh wait I can just tell it all the tools I commonly use and where to put them... Hmmm topbar or side bar. Wow so much fun getting to make all these decisions!
Ok time to change fonts. "Please add a font picker so I can pick a font"
Comment by ryandrake 3 days ago
Comment by 7e 3 days ago
Comment by parineum 3 days ago
Comment by candiddevmike 3 days ago
Comment by technofiend 3 days ago
Not to be argumentative, but I have a concern that whomever I buy my solution from will have vibe coded it instead. I guess that means my support contract entitles me to hassling them about it, but I'm starting to worry it's just LLMs and vibe coded apps all the way down.
Comment by Ravus 3 days ago
When (not if) software breaks in production, you need to be able to debug it effectively. Knowing that external libraries do their base job is really helpful in reducing the search space and in reducing the blast radius of patches.
Note that this is not AI-specific. More generally, in-house implementations of software that is not your core business brings costs that are not limited to that of writing said implementation.
Comment by ryandrake 3 days ago
Also, you don't have to fear breaking updates--you know for sure that the software's UI will not just change out from under you because some designer had to pad their portfolio. Or that you're not going to lose a critical feature because the developer decided to refactor and leave it out.
I'm currently going through and looking at some of the bigger, bloated, crashing slow-moving software I use and working on replacements.
Comment by otikik 3 days ago
Because I thought I needed a hammer for nails (employee payroll) but then I realized I also need it to screw (sales), soldering (inventory management) and cleanup (taxes).
Oh and don't forget that next month the density of iron can lower up to 50%.
Comment by freedomben 3 days ago
Good points. It does feel like that happens quite often
Comment by jayd16 3 days ago
Comment by seniorThrowaway 3 days ago
Comment by LtWorf 2 days ago
Comment by jcarrano 3 days ago
Comment by seniorThrowaway 3 days ago
Comment by YesBox 3 days ago
IMO you only need to look at the 30+ year history of Linux to see how wrong this prediction is. There will be a small group of people who do as you say, but the other 95% will happily pay for someone else to take care of their problem. Convenience is the supreme king of the American market.
Comment by wasmitnetzen 3 days ago
I very much like to use the years of debugging and innovation others spent on that very same problem that I'm having.
Comment by FeloniousHam 3 days ago
- PRO Deployer
- MS Typescript
- Typescript-Go
- a bespoke internal extension to automate a lot of housekeeping when developing against tickets (git checks, branch creation, stash when switching, automatically connecting and updating ticket system)
Comment by skybrian 3 days ago
Comment by marginalia_nu 3 days ago
Comment by wongarsu 3 days ago
Sure, leftpad and python-openai aren't hugely valuable in the age of LLMs, but redis and ffmpeg are still as useful as ever. Probably even more useful now that LLMs can actually know and use all their obscure features
Comment by storystarling 3 days ago
Comment by squigz 3 days ago
Whatever it is, the future will also certainly not be what it was a couple decades ago - that is, every one inventing their own solution to solved problems, resulting in a mess of tools with no standardization. There is a reason libraries/frameworks/etc exist.
Comment by panta 3 days ago
Comment by onraglanroad 3 days ago
Comment by rglover 3 days ago
Comment by tracker1 3 days ago
Comment by rglover 3 days ago
Comment by dimitri-vs 3 days ago
Comment by MagicMoonlight 3 days ago
You go to the shed because you know the hammer has undergone extensive testing and won’t immediately be hacked, costing you hundreds of millions.
Comment by theshrike79 1 day ago
Getting poor software made by Indians is a cycle that takes months.
With AI you get the first poor version within hours. Then you can iterate. In a week of hourly iteration you'll most likely have something usable.
The Indians are still having meetings about when to have meetings and you can see from the logs they haven't accessed any of the documentation you gave them for the issue. Billing is on point though.
Comment by GoatInGrey 3 days ago
I do not suspect that we will stay in an individualized programs Tower of Babel situation, if we ever enter it in the first place.
It's anyone's guess as to what we end up settling on, of course. This is just a guess of mine.
Comment by parliament32 3 days ago
And many businesses do, and it works, until it doesn't. I agree with you; that's the future I'm seeing for ad-hoc LLM apps.
Comment by xacky 2 days ago
Comment by jvanderbot 3 days ago
The danger is not "Nobody uses OSS".
The danger is "building software becomes exponentially more difficult without a commons to build from".
Comment by eikenberry 3 days ago
Comment by exe34 3 days ago
Comment by raphyjake 2 days ago
Comment by pier25 3 days ago
Comment by groundzeros2015 3 days ago
Comment by matkoniecz 3 days ago
once such LLMs exist, this question may be worth considering
they do NOT exist at this moment
Comment by groundzeros2015 3 days ago
Comment by pjmlp 3 days ago
Comment by reactordev 3 days ago
Comment by njhnjhnjhnjh 3 days ago
Comment by MSFT_Edging 3 days ago
Otherwise we'll all individually be burning power via GPUs to reinvent the wheel thousands of times.
Also, look at the level of effort to maintain any bespoke machine. Assembly lines with interchangeable parts were a big deal for a reason. That follows with software via standard libraries and APIs. If you reinvent your software wheel every time, we'll all have careers trying to fix the deluge of slop that has been generated.
Comment by nektro 2 days ago
Comment by estimator7292 3 days ago
Now that LLMs can do it while I make coffee, I can do more useful things with my time like writing business code and posting on HN.
Beyond that, meh.
Comment by HugoDz 3 days ago
Comment by player1234 3 days ago
Comment by computerthings 3 days ago
Comment by draxil 3 days ago
But I think the reality is: LLMs democratise access to coding. In a way this decreases the market for complete solutions, but massively increases the audience for building blocks.
Comment by ipaddr 3 days ago
Comment by croes 3 days ago
Vibe coders don't code, they let code. So LLMs democratise access to coders.
Comment by kibwen 3 days ago
Comment by fragmede 3 days ago
Comment by dns_snek 3 days ago
We all know that the music is going to stop eventually and that the landscape after that is going to look very different. Subsidies will stop and investors will want their trillions in returns. Talking about "democratization" while everyone is just using other people's money is completely premature.
Airbnb "democratized travel" for a while and now they're more expensive than their predecessors.
Comment by fragmede 2 days ago
Comment by dns_snek 2 days ago
I'm not sure where you got this from. They'll jack up the prices at some undetermined time in the future once the weight of reality overpowers the manic hype we've been experiencing for the past few years.
My expectation is that this will start unfolding within the next 5 years and people will start looking back to today as the golden age of AI where everything was cheap, exciting, and ad-free.
Comment by croes 2 days ago
And if all software has been written that can be written what additional software do you want to write?
Comment by pepperball 3 days ago
Comment by InMice 3 days ago
I had a job where in short we had a lot of pain points with software that we had no resources permitted to fix them. With a mix past experience, googling I started writing some internal web based tools to fix these gaps. Everyone was happy. This is where I see vibe coding being really helpful in the higher level stuff like higher level scripting and web based tools. Just my opinion based on my experience.
Comment by nicoburns 3 days ago
It can be mitigated by PR submitters doing a review and edit pass prior to submitting a PR. But a lot of submitters don't currently do this, and in my experience the average quality of PRs generated by AI is definitely significantly lower than those not generated by AI.
Comment by pgroves 3 days ago
[1] https://github.com/anomalyco/opencode/pulls [2] https://github.com/Aider-AI/aider/pulls
Comment by matkoniecz 3 days ago
Some are opening PRs, some are posting comments in issues that repeat what was said already, just in more words.
Comment by trey-jones 3 days ago
Comment by electroly 3 days ago
Comment by trey-jones 3 days ago
Comment by torginus 3 days ago
Comment by electroly 3 days ago
Comment by gjadi 3 days ago
It's like reading, for better learning and understanding, it is advised that you think and question the text before reading it, and then again after just skimming it.
Whereas if you ask first for the answer, you are less prepared for the topic, is harder to form a different opinion.
It's my perception.
Comment by hxugufjfjf 3 days ago
Comment by groundzeros2015 3 days ago
I need to make decisions about how things are implemented. Even if it can pick “a way” that’s not necessarily going to be a coherent design that I want.
In contrast for review I already made the choices and now it’s just providing feedback. More information I can choose to follow or ignore.
Comment by Leynos 3 days ago
Comment by echelon 3 days ago
With the time they save using AI, they can get much more work done. So much that having other engineers learn the codebase is probably not worth it anymore.
Large scale software systems can be maintained by one or two folks now.
Edit: I'm not going to get rate limited replying to everyone, so I'll just link another comment:
Comment by tracker1 3 days ago
In the end, you should not just skip QA/QC and fitness testing. Many things can fit a technical spec and still be absolutely horrible. With AI assisted developmnet, imo it's that much more important to get the UX right. I don't want 10x the apps if they're all half-implemented garbage that look like garbage are hard to use and just painful to install, maintain and use.
Library creation still has a place here... and so far, getting AI code assistants to actually understand and use a given library that may be less popular has been at the very least, interresting.
Comment by wooderson_iv 3 days ago
Comment by j16sdiz 3 days ago
Comment by matkoniecz 3 days ago
> Large scale software systems can be maintained by one or two folks now.
No, LLMs are not so powerful yet.
Comment by erelong 3 days ago
Comment by shafyy 3 days ago
Comment by marginalia_nu 3 days ago
I can't think of even a single example of OSS being monetized through direct user engagement. The bulk of it just isn't monetized at all, and what is monetized (beyond like a tip jar situation where you get some coffee money every once in a while) is primarily sponsored by enterprise users, support license sales, or through grants, or something like that. A few projects like Krita sell binaries on the steam store.
Comment by dfox 3 days ago
Comment by WorldMaker 3 days ago
All of those examples are the "direct user engagement" in question. No one tips a tip jar that they don't see. Enterprise users don't sponsor projects or buy licenses when they don't know they are using projects they should sponsor or buy a better license.
If an LLM is automating your `npm install` you probably don't see the funding requests. Are you running `npm fund` in your spare time?
If an LLM is automating your need to lookup library documentation you probably don't see that the library's own documentation has a Ko-Fi link or GitHub Sponsors request. Would you check library websites and GitHub repos on your own for such things without the need to read their documentation?
Comment by nprateem 3 days ago
Comment by delegate 3 days ago
From the tools which were used to design and develop the models (programming languages, libraries) to the operating systems running them to the databases used for storing training data .. plus of course they were trained mostly on open source code.
If OSS didn't exist, it's highly unlikely that LLMs would have been built.
Comment by jdiff 3 days ago
Comment by blibble 3 days ago
would anyone want SlopHub Copilot if it had been trained exclusively on Microsoft's code?
(rhetorical question)
Comment by devinprater 3 days ago
Now it reads (usually) only newly incoming text, I can feel around the screen to read a line at a time, and cursor tracking works well enough. Then I got Emacs and Emacspeak working, having Gemini build DecTalk (TTS engine) for Termux and get the Emacspeak DecTalk speech server working with that. I'm still amazed that, with a Bluetooth keyboard, I have Linux, and Emacs, in my pocket. I can write Org and Markdown, read EPUB books in Emacs with Nov.el, look at an actual calendar not just a list of events, and even use Gemini CLI and Claude Code, all on my phone! This is proof that phones, with enough freedom, can be workstations. If I can get Orca working on a desktop environment in Termux-GUI. But even with just Emacs and the shell, I can do quite a bit.
Then I decided to go wild and make an MUD client for Emacs/Emacspeak, since accessible ones for Android are limited, and I didn't trust my hacks to Termux to handle Tintin++ very well. So, Emacs with Emacspeak it was, and Elmud [2] was born.
Elmud has a few cool features. First of all, since Emacspeak has voice-lock, like font-lock but for TTS, Ansi colors can be "heard", like red being a deeper voice. Also a few MUD clients have sound packs on Windows, which make them sound more like a modern video game, while still being text-based. I got a few of those working with Elmud. You just load one of the supported MUD's, and the sound pack is downloaded and installed for you. It's easy and simple. And honestly, that's what I want my tools to provide, something I, or anyone else who chooses to use them, that is easy to get the most out of.
None of this would have been possible without AI. None of it would have been done. It would have remained a dream. And yes, it was all vibe-coded, mostly with Codex 5.2 on high thinking. And yes, the code may look awful. But honestly, how many closed-source programs look just as bad or even worse under the covers of compilation?
Comment by cheema33 3 days ago
Comment by Quarrel 3 days ago
Did you read it?
It isn't saying that LLMs will replace major open source software components. It said that the "reward" for providing, maintaining and helping curate these OSS pieces; which is the ecosystem they exist in, just disappears if there is no community around it, just an LLM ingesting open source code and spitting out a solution good or bad.
We've already seen curl buckle under the pressure, as their community minded, good conscious effort to give back to security reports, collapsed under the weight of slop.
This is largely about extending that thesis to the entire ecosystem. No GH issues, no PRs, no interaction. No kudos on HN, no stars on github, no "cheers mate" as you pass them at a conference after they give a great talk.
Where did you get that you needed to see a Linux kernel developed from AI tools, before you think the article's authors have a point?
Comment by volkercraig 3 days ago
Oh... so nothing's gonna change for me then...
Comment by cheema33 3 days ago
It is in the title: "Vibe coding kills open source"
Clickbait titles beget clickbait responses.
Comment by pmarreck 3 days ago
GPL is a dead man walking since you can have any LLM cleanroom a new implementation in a new language from a public spec with verifiable "never looked at the original source" and it can be more permissively-licensed however you wish (MIT, BSD etc).
case in point, check out my current deps on the project I'm currently working on with LLM assist: https://github.com/pmarreck/validate/tree/yolo/deps
"validate" is a project that currently validates over 100 file formats at the byte level; its goal is to validate as many formats as possible, for posterity/all time.
Why did I avoid GPL (which I am normally a fan of) since this is open-source? I have an even-higher-level project I'm working on, implementing automatic light parity protection (which can proactively repair data without a RAID/ZFS setup) which I want to make for sale, whose code will (initially) be private, and which uses this as a dependency (no sense in protecting data that is already corrupted).
Figured I'd give this to the world for free in the meantime. It's already found a bunch of actually-corrupt files in my collection (note that there's still some false-positive risk; I literally released this just yesterday and it's still actively being worked on) including some cherished photos from a Japan trip I took a few years ago that cannot be replaced.
It has Mac, Windows and Linux builds. Check the github actions page.
Comment by romanows 3 days ago
Comment by pmarreck 3 days ago
Comment by natebc 3 days ago
I was under the impression that copyright was only available for works created by people.
Comment by pmarreck 3 days ago
LLMs don’t change that. They’re just tools. The relevant question is whether a human exercised creative judgment in selecting, shaping, or editing the output. Which I most certainly did. Otherwise my month of work here would have just been a minute and a single prompt and then just grabbing a coffee.
We already handle this with cameras, compilers, spell-checkers, and procedural generation. If you push a button and accept whatever falls out, you may not have authorship. If you meaningfully direct and curate the result, you almost certainly do.
Responsibility still lands on the human, which is a good sanity check: there is no such thing as “LLM-authored but human-unaccountable” work. Pretending otherwise is just outsourcing epistemic responsibility to a stochastic tool. Similarly, you can't both assign blame to me when things go wrong, but credit to the LLM when things go right. Or vice versa.
Comment by digiown 3 days ago
Does this make sense at all? ZFS scrubs only reads the data you have, not the whole drive, and repairs data if possible. The more data you have, the more you have to validate regardless of the tools used. The BER is also just a terrible metric and is not reflective of how drives actually behave.
Comment by dahauns 3 days ago
...erm.
To adress the elephant in the room: Who exactly is supposed to be verifiable to never have looked at the original source? You or the LLM?
Comment by pmarreck 3 days ago
Although it would be much harder to prove, of course, that I never saw the original code/implementation.
Comment by tomaytotomato 3 days ago
However trying to get it to do anything other than optimise code or fix small issues it struggles. It struggles with high level abstract issues.
For example I currently have an issue with ambiguity collisions e.g.
Input: "California"
Output: "California, Missouri"
California is a state but also city in Missouri - https://github.com/tomaytotomato/location4j/issues/44
I asked Claude several times to resolve this ambiguity and it suggested various prioritisation strategies etc. however the resulting changes broke other functionality in my library.
In the end I am redesigning my library from scratch with minimal AI input. Why? because I started the project without the help of AI a few years back, I designed it to solve a problem but that problem and nuanced programming decisions seem to not be respected by LLMs (LLMs dont care about the story, they just care about the current state of the code)
Comment by Cthulhu_ 3 days ago
The project, or your brain? I think this is what a lot of LLM coders run into - they have a lot of intrinsic knowledge that is difficult or takes a lot of time and effort to put into words and describe. Vibes, if you will, like "I can't explain it but this code looks wrong"
Comment by tomaytotomato 3 days ago
Essentially I ask an LLM to look at a project and it just sees the current state of the codebase, it doesn't see the iterations and hacks and refactors and reverts.
It also doesn't see the first functionality I wrote for it at v1.
This could indeed be solved by giving the LLM a git log and telling it a story, but that might not solve my issue?
Comment by michaelbuckbee 3 days ago
FWIW - it works a lot better to have it interact via the CLI than the MCP.
Comment by alright2565 3 days ago
Comment by gibspaulding 3 days ago
Comment by alright2565 3 days ago
4.5 Sonnet, but because I've been stuck on 3.7 Sonnet for so long due to corporate policy I wrote the wrong thing.
And yeah corporate policy. Opus is not available. I prefer Codex for my personal coding but I have not needed to look in the Git history here yet.
Comment by azuanrb 3 days ago
Comment by neko-kai 3 days ago
Comment by fragmede 3 days ago
Comment by cpursley 3 days ago
Comment by nevi-me 3 days ago
I suppose a year ago we were talking about prompt engineers, so it's partly about being good at describing problems.
Comment by faxmeyourcode 3 days ago
Comment by epolanski 3 days ago
Document, document, document: your architecture, best practices, preferences (both about code and how you want to work with the LLM and how do you expect it to behave it).
It is time consuming, but it's the only way you can get it to assist you semi-successfully.
Also try to understand that LLM's biggest power for a developer is not in authoring code as much as assistance into understanding it, connecting dots across features, etc.
If your expectation is to launch it in a project and tell it "do X, do Y" without the very much needed scaffolding you'll very quickly start losing the plot and increasing the mess. Sure, it may complete tasks here and there, but at the price of increasing complexity from which it is difficult for both you and it to dig out.
Most AI naysayers can't be bothered with the huge amount of work required to setup a project to be llm-friendly, they fail, and blame the tool.
Even after the scaffolding, the best thing to do, at least for the projects you care (essentially anything that's not a prototype for quickly validating an idea) you should keep reading and following it line by line, and keep updating your scaffolding and documentation as you see it commit the same mistakes over and over. And part of scaffolding requires also to put the source code of your main dependencies. I have a _vendor directory with git subtrees for major dependencies. LLMs can check the code of the dependencies, the tests, and figure out what they are doing wrong much quicker.
Last but not least, LLMs work better with certain patterns, such as TDD. So instead of "implement X", it's better to "I need to implement X, but before we do so, let's setup a way for testing and tracking our progress against". You can build an inspector for a virtual machine, you can setup e2es or other tests, or just dump line by line logs in some file. There's many approaches depending on the use case.
In any case, getting real help for LLMs for authoring code (editing, patching, writing new features) is highly dependent on having good context, good setup (tests, making it write a plan for business requirements and one for implementation) and following and improving all these aspects as you progress.
Comment by tomaytotomato 3 days ago
My project is quite well documented and I created a Prompt a while back along with some mermaid diagrams
https://github.com/tomaytotomato/location4j/tree/master/docs
I can't remember the exact prompt I gave to the LLM but I gave it a Github issue ticket and description.
After several iterations it fixed the issue, but my unit tests failed in other areas. I decided to abort it because I think my opinionated code was clashing with the LLM's solution.
The LLM's solution would probably be more technically correct, but because I don't do l33tcode or memorise how to implement Trie or BST my code does it my way. Maybe I just need to force the LLM to do it my way and ignore the other solutions?
Comment by theshrike79 1 day ago
The role thing doesn't work anymore ("You are a..."), it's just fanfiction.
Also adding MUST NOT just pollutes the context, it's the "don't think of a pink elephant" but for LLMs
It's enough to say "Use Java 21+", no need to add a negative of "don't use below 21"
And this is just weird: "MUST NOT Change the license from MIT" - did an agent change the license in your code or what prompted you to add this?
Comment by eichin 3 days ago
Comment by tomaytotomato 3 days ago
My reworked approach is to return a list of results with a probability or certainty score.
In the situation of someone searching for London, I need to add some sort of priority for London, UK.
My dataset is sourced from an opensource JSON file which I am now pre-processing and identifying all collisions on it.
There are so many collisions!
Could I pick your brains and you could critique my approach? Thanks
Comment by skybrian 3 days ago
Comment by softwaredoug 3 days ago
Make sure there’s a holdout the agent can’t see that it’s measured against. (And make sure it doesn’t cheat)
https://softwaredoug.com/blog/2026/01/17/ai-coding-needs-tes...
Comment by faxmeyourcode 3 days ago
You have to tell it about the backstory. It does not know unless you write about it somewhere and give it as input to the model.
Comment by krona 3 days ago
Comment by krona 3 days ago
Comment by px43 3 days ago
It does not struggle, you struggle. It is a tool you are using, and it is doing exactly what you're telling it to do. Tools take time to learn, and that's fine. Blaming the tools is counterproductive.
If the code is well documented, at a high level and with inline comments, and if your instructions are clear, it'll figure it out. If it makes a mistake, it's up to you to figure out where the communication broke down and figure out how to communicate more clearly and consistently.
Comment by smrq 3 days ago
It's fine to critique your own tools and their strengths and weaknesses. Claiming that any and all failures of AI are an operator skill issue is counterproductive.
Comment by vrighter 11 hours ago
Comment by zeroCalories 3 days ago
Comment by rtp4me 3 days ago
Comment by zeroCalories 3 days ago
Comment by whateveracct 3 days ago
Comment by ktallett 3 days ago
I was very sceptical but I will admit I think vibe coding has a place in society, just what it is yet is still to be determined. It can't help most for sure but it can help some in some situations.
Comment by Cthulhu_ 3 days ago
If they don't exist, AND the author is comitted to maintaining them instead of just putting it online, sure. But one issue I see is that a lot of these tools you describe already exist, so creating another one (using code assist tools or otherwise) just adds noise IMO.
The better choice is to research and plan (as you say in your first sentence) before comitting resources. The barrier to "NIH" is lowered through code assistants, which risks reducing collaboration in open source land in favor of "I'll just write my own".
Granted, "I'll write my own" has always felt like it has a lower barrier to entry than "I'm going to search for this tool and learn to use it".
Comment by data-ottawa 3 days ago
Maybe the best feature of vibe coding is that it makes the regret factor of poor early choices much lower. Its kind of magic to go "you know what, I was wrong, let's try this approach instead" and not having to spend huge amounts of time fixing things or rewriting 80% of the project.
It's made it a lot more fun to try building big projects on my own, where I would go into decision paralysis or prematurely optimize and never start the meat or learning of the core project.
Its also been nice to have agents review my projects for major issues, so I feel more confident sharing them.
Comment by fc417fc802 3 days ago
Setting out to implement a feature only to immediately get bogged down in details that I could probably get away with glossing over. LLMs short circuit that by just spitting something out immediately. Of course it's of questionable quality, but once you get something working you can always come back and improve it.
Comment by cess11 3 days ago
Comment by hayd 3 days ago
I haven't worked out how to do this for my own projects.
Once you've set it up it's not too hard to imagine an AI giving an initial PR assessment... to discard the worst AI slop, offer some stylistic feedback, or suggest performance concerns.
Comment by antirez 3 days ago
Comment by m000 3 days ago
It feels that vibe coding may exacerbate fragmentation (10 different vibe-coded packages for the same thing) and abandonment (made it in a weekend and left it to rot) for open source software.
Comment by antirez 3 days ago
Comment by vrighter 11 hours ago
Software development hasn't been democratized. The term doesn't even make sense. Anyone could learn whatever using the same resources as anyone else. I would (and have, in the past) gladly teach anyone who asks me to. And most of my friends would too. The fact that non-devs couldn't write code was not because devs were gatekeeping, but because they didn't want to learn. And they still don't.
Comment by koakuma-chan 3 days ago
Comment by echelon 3 days ago
I've been a senior engineer doing large scale active-active, five nines distributed systems that process billions of dollars of transactions daily. These are well thought out systems with 20+ folks on design document reviews.
Not all of the work falls into that category, though. There's so much plumbing and maintenance and wiring of new features and requirements.
On that stuff, I'm getting ten times the amount of work done with AI than I was before. I could replace the juniors on my team with just myself if I needed to and still get all of our combined work done.
Engineers using AI are going to replace anyone not using AI.
In fact, now is the time to start a startup and "fire" all of these incumbent SaaS companies. You can make reasonable progress quickly and duplicate much of what many companies do without much effort.
If you haven't tried this stuff, you need to. I'm not kidding. You will easily 10x your productivity.
I'm not saying don't review your own code. Please do.
But Claude emits reasonable Rust and Java and C++. It's not just for JavaScript toys anymore.
- - - - - - - - - - - -
Edit:
Holy hell HN, downvoted to -4 in record time. Y'all don't like what's happening, but it's really happening.
I'm not lying about this.
I provided my background so you'd understand the context of my claims. I have a solid background in tech.
The same thing that happened to illustration and art is happening here, to us and to our career. And these models are quite usable for production code.
I can point Claude to a Rust HTTP handler and say, "using this example [file path], write a new endpoint that handles video file uploads, extracts the metadata, creates a thumbnail, uploads them to the cloud storage, and creates the relevant database records."
And it does it in a minute.
I review the code. It's as if I had written it. Maybe a change here or there.
Real production Rust code, 100 - 500 LOC, one shotted in one minute. It even installs the routes and understands the HTTP framework DSL. It even codegens Swagger API documentation and somehow understands the proc macro DSL that takes Rust five minutes to compile.
This tech is wizardry. It's the sci fi stuff we dreamed of as kids.
I don't get the sour opinions. The only thing to fear is big tech monopolozation.
I suppose the other thing to worry about is what's going to happen to our cushy $400k salaries. But if you make yourself useful, I think it'll work out just fine.
Perhaps more than fine if you're able to leverage this to get ahead and fire your employer. You might not need your employer anymore. If you can do sales and wear many hats, you'll do exceedingly well.
I'm not saying non-engineers will be able to do this. I'm saying engineers are well positioned to leverage this.
Comment by koakuma-chan 3 days ago
There was a submission to a blog post discussing applications of AI but it got killed for some reason.
https://news.ycombinator.com/item?id=46750927
I remain convinced that if you use AI to write code then your product will sooner or later turn into a buggy mess. I think this will remain the case until they figure out how to make a proper memory system. Until then, we still have to use our brains as the memory system.
One strategy I've seen that I like is using AI to prototype, but then write actual code yourself. This is what the Ghostty guy does I believe.
I agree that AI can write decent Rust code, but Rust is not a panacea. From what I heard, Cursor has a lot of vibe-coded Rust code, but it didn't save it from being, as I said, a buggy mess.
Comment by bugglebeetle 3 days ago
FYFY
Comment by volkercraig 3 days ago
Comment by nicoburns 3 days ago
There are lots of people claiming this. Many of whom have a solid background. Every now and then I check out someone's claim (checking the code they've generated). I've yet to find an AI-generated codebase that passed that check so far.
Perhaps yours is the one that does, but as we can't see the code for ourselves, there's no way for us to really know. And it's hard to take your word for it when there are so many people falsely making the same claims.
I expect a lot of HNers have had this experience.
Comment by simonw 3 days ago
Comment by koakuma-chan 3 days ago
Comment by munksbeer 6 hours ago
I went from using copilot autocomplete and mostly just disabling it because it was not great, to now using claude most of my day. I gets stuff wrong. It does generate inefficient code. It definitely has a way to go, but otherwise it just saves so much time.
Comment by koakuma-chan 4 hours ago
Comment by gitaarik 2 days ago
Comment by joks 3 days ago
What are you talking about? Illustrators and artists are not being replaced by AI or required to use AI to "keep up" in the vast majority of environments.
> "I don't get the sour opinions."
The reasoning for folks' "sour opinions" has been very well-documented, especially here on HN. This comment reads like people don't like AI because they think it's slow or something, which is not the case.
Comment by echelon 3 days ago
I don't know what jobs have been impacted yet, but there will likely be pressure for all content creators and knowledge workers to use the tools to get more work done.
We'll probably start seeing this in software development this year. The tools finally feel ready for prime time.
> This comment reads like people don't like AI because they think it's slow or something, which is not the case.
I am familiar with the most common arguments in opposition - stealing training data, hallucinations, not understanding logic (this is why "engineers in the loop" matters), big corps owning the tech (I really agree with this one), power usage, etc.
It feels as though the downvotes are from people that "dislike AI" for any of the aforementioned reasons. In the face of the possibility of losing jobs to engineers that leverage AI to get more quality work done, however, I don't know why HN engineers downvote anecdotes about real world usage. This is vital to know and understand. I would think one would want more evidence to consider about the state of things.
This is a quickly developing story. Your jobs are or will be on the line.
It doesn't matter what your personal misgivings are if your job will soon require the use of AI. You can hate it all you want, but if people are getting 10x more work done than you, you really don't have a choice.
This will be the same in every career sector with AI models that can be deployed to automate work -- marketing, editing, film, animation, VFX, software, music production, 3D modeling, game design, etc.
I don't think the jobs are going away, but I do think they're going to change. Fast.
No sense in sour grapes.
Comment by bloomca 3 days ago
You claimed that it already happened to illustrators and artists, and while I am sure they use it one way or another, I don't think it transformed the industry. Now, I am not saying that it won't amount to anything in software, I just don't think it is ready as of right now outside of greenfield projects, mostly because the scope is limited.
I am pretty positive that at some point we'll have a tool which will automate the generation -> code review -> fixing (multiple loops) -> releasing without people. Currently people are the bottleneck and imo a better way is to exclude people completely outside of initial problem statement and accepting the result. Otherwise it is just too janky, that 10x comes with a huge asterisk that can unironically slow you down after all said and done.
Comment by echelon 3 days ago
I really don't know what else to say.
Comment by bloomca 3 days ago
However, the complexity is still not handled super well, as you need to spend more time in code review and testing to make sure all edge cases are covered and the general module interconnection is decent. Ideally we want to modularize and make the breaking surface very small, but often it is not possible.
I think the next step is to fully remove people as accepting changes manually is just too brittle; I also think it is probably possible to do with the current tools but needs a very different approach from the current meta of highly specific docs.
Comment by matkoniecz 3 days ago
large part of formerly done by humans graphics is now autogenerated
Comment by koakuma-chan 3 days ago
I gave you an upvote FWIW, after all, I mean, my job's codebase is already a buggy mess, so it doesn't hurt to throw AI on it, which is what I do.
> You might not need your employer anymore. If you can do sales and wear many hats, you'll do exceedingly well.
Wasn't this the case before AI as well?
Comment by gegtik 3 days ago
Comment by blibble 3 days ago
so not now, then?
Comment by jen20 3 days ago
Comment by GuinansEyebrows 3 days ago
Comment by avaer 3 days ago
> One incredible thing was the ability to easily merge what was worth merging from forks, for instance
I agree, this is amazing, and really reduces the wasted effort. But it only works if you know what exists and where.
Comment by mg74 3 days ago
Comment by avaer 3 days ago
But IMO the primitives we need are also fundamentally different with AI coding.
Commits kind of don't matter anymore. Maybe PR's don't matter either, except as labels. But CI/hard proof that the code works as advertised is gold, and this is something git doesn't store by default.
Additionally, as most software moves to being built by agents, the "real" git history you want is the chat history with your agent, and its CoT. If you can keep that and your CI runs, you could even throw away your `git` and probably still have a functionally better AI coding system.
If we get a new Github for AI coding I hope it's a bit of a departure from current git workflows. But git is definitely extensible enough that you could build this on git (which is what I think will ultimately happen).
Comment by pietro72ohboy 3 days ago
Comment by fc417fc802 3 days ago
Comment by reactordev 3 days ago
Comment by the__alchemist 3 days ago
Comment by forgotpwd16 3 days ago
Comment by jdiff 3 days ago
Comment by wasmainiac 3 days ago
Comment by barelysapient 3 days ago
The cost improvement for an LLM to emit a feature (with an engineer in the loop) is too much of an improvement. We’ll look at engineers coding in C the same way we look at engineers today who code in assembly. LLM enabled development becomes the new abstraction; probably with a grammar and system for stronger specification formalization.
Comment by delaminator 3 days ago
I have had a lot of conversations with Claude about it and it supports that theory.
Comment by 13rac1 3 days ago
Comment by delaminator 3 days ago
Comment by recursive 3 days ago
Comment by hxugufjfjf 3 days ago
Comment by linuxftw 3 days ago
Comment by ozten 3 days ago
On the whole, I think it is a net gain for civilization, but if we zoom into OSS licensing... not good.
Comment by kode-targz 3 days ago
Comment by finaard 3 days ago
Comment by ozten 3 days ago
Example: We can achieve more climate solutions (and faster) thanks to technological acceleration.
Comment by tracker1 3 days ago
Comment by tracker1 3 days ago
Comment by BoredPositron 3 days ago
Comment by bluejay2387 3 days ago
Comment by korenmiklos 3 days ago
Comment by Sharlin 3 days ago
Comment by contravariant 3 days ago
Comment by tonyedgecombe 3 days ago
Comment by pixl97 3 days ago
Comment by donatj 3 days ago
My guess is instead of Googling "library that does X" people are asking AI to solve the problem and it's regurgitating a solution in place? That's my theory anyway.
Comment by j4coh 3 days ago
Comment by dev_l1x_be 3 days ago
Comment by tracker1 3 days ago
Comment by sanskritical 3 days ago
Vibe coding eventually creates more value for FOSS, not less.
Comment by gegtik 3 days ago
Comment by reustle 3 days ago
Comment by alentred 3 days ago
[1] https://git-scm.com/book/en/v2/Distributed-Git-Distributed-W...
Comment by Sevii 3 days ago
Comment by tracker1 3 days ago
Sometimes having dozens of one-off scripts is easier/simpler than trying to create the uber-flexibly one tool does all solution.
Comment by progx 3 days ago
Comment by program_whiz 3 days ago
Is anyone replacing firefox, chromium, postgres, nginx, git, linux, etc? It would be idiotic to trade git for a vibe coded source control. I can't even imagine the motivations, maybe "merges the way I like it"?
Not sure, but anyone who's saying this stuff hasn't even taken the basic first level glance at what it would entail. By all means, stop paying $10 a month to "JSON validator SaSS", but also don't complain with the little niggling bugs, maintenance and organization that comes with it. But please stop pretending you can just vibe code your own Kafka, Apache, Vulkan, or PostGRES.
Yes, you can probably go faster (possibly not in the right direction if inexperienced), but ultimately, something like that would still require very senior, experienced person, using the tool in a very guided way with heavy review. By why take on the maintenance, the bug hunting, and everything else, unless that is your main business objective?
Even if you can 10x, if you use that to just take on 10x more maintenance, you haven't increased velocity. To really go faster, that 10x must be focused on the right objective -- distinctive business value. If you use that 10x to generate hundreds of small tools you now have to juggle and maintain, that have no docs or support, no searchable history of problems solved, you may have returned yourself to 1x (or worse).
This is the old "we'll write our own inhouse programming language" but leaking out to apps. Sure, java doesn't work _exactly_ the way you want it to, you probably have complaints. But writing your own lang will be a huge hit to whatever it was you actually wanted to use the language for, and you lose all the docs, forums, LSP / debugging tools, ecosystem, etc.
Comment by chfritz 3 days ago
Comment by tinyhouse 3 days ago
Comment by wessorh 3 days ago
Comment by wessorh 3 days ago
Comment by grimmzoww 2 days ago
Comment by verdverm 3 days ago
"most" maintainers make exactly zero dollars. Further, OSS monetization rarely involves developer engagement, it's been all about enterprise feature gating
Comment by Olshansky 3 days ago
---
Concrete example of a no: I set up [1] in such a way that anyone can implement a new blog -> rss feed; docs, agents.md, open-source, free, etc...
Concrete example of a yes: Company spends too much money on simple software.
--- Our Vision ---
I feel the need to share: https://grove.city/
Human Flywheel: Human tips creator <-> Creator engages with audience
Agent Flywheel: Human creates creative content <-> Agent tips human
Yes, it uses crypto, but it's just stablecoins.
This is going to exist in some fashion and all online content creation (OSS and other) will need it.
---
As with everything, it Obvious
Comment by tracker1 3 days ago
But the assertion that everything needs to change is absurd. Articles like this are similar in my mind to arguments for communism because every artist deserves a living wage... that's just not how society can sustain itself in reality. Maybe in a world without scarcity, but I don't see scarcity going away any time soon.
Comment by neko-kai 3 days ago
e.g. Vibe coding defeats GNOME developers' main argument for endlessly deleting features and degrading user experience - that features are ostensibly "hard to maintain".
Well, LLMs are rapidly reducing development costs to 0.
The bottleneck for UI development is now testing, and here desktop Linux has advantage - Linux users have been trained like Pavlov's dogs to test and write detailed upstream bug reports, something Windows and macOS users just don't do.
Comment by Cthulhu_ 3 days ago
Also it's a formal system and process, "vibe" coding is anything but. Call me curmudgeony (?) but I don't think "vibe coding" should be a phrase used to describe LLM assisted software engineering in large / critical systems.
Comment by croes 3 days ago
At some point the investors want to see profit.
Comment by rvz 3 days ago
Oh sweet summer child.
> Well, LLMs are rapidly reducing development costs to 0.
And maintainance costs along with technical debt rapidly goes up.
Comment by rtp4me 3 days ago
Comment by sailfast 3 days ago
Comment by lukan 3 days ago
I think the title is clickbait.
The conclusion is:
"Vibe coding represents a fundamental shift in how software is produced and consumed. The productivity gains are real and large. But so is the threat to the open source ecosystem that underpins modern software infrastructure. The model shows that these gains and threats are not independent: the same technology that lowers costs also erodes the engagement that sustains voluntary contribution."
The dangers I see rather in projects drowning in LLM slop PR's, instead of less engagement.
And the benefits of LLMs to open source in lowering the cost to revive and maintain (abandoned) projects.
Comment by earino 3 days ago
https://bsky.app/profile/gaborbekes.bsky.social/post/3md4rga...
(Note, I receive a thanks in the paper.)
Comment by korenmiklos 3 days ago
> given everything we know about OSS incentives from prior studies and how easy it is to load an OSS library with your AI agent, the demand-reducing effect of vibe coding is larger than the productivity-increasing effect
but that would be a mouthful
Comment by lukan 3 days ago
Comment by positron26 3 days ago
Comment by wolfi1 3 days ago
Comment by lukan 3 days ago
LLM's did help with quickly researching dependencies unknown to me and investigating build errors, but ideally I want to set it up in a way, that the agent can work on its own, change -> try to build -> test it. Once that works half automated, I call it success.
Comment by jorvi 3 days ago
This is also just untrue. There is a study showing that the productivity gain is -20%, developers (and especially managers) just assume it is +25%. And when they are told about this they still feel they are +20% faster. It's the dev equivalent of mounting a cool-looking spoiler to your car.
There are productivity gains, but they're in the fuzzy tasks like generating documentation and breaking up a project into bite-sized tasks. Or finding the right regex or combination of command line flags, and that last one I would triple verify if it was anything difficult to reverse.
Comment by gyanchawdhary 3 days ago
Comment by OrvalWintermute 3 days ago
Comment by matkoniecz 3 days ago
> When OSS is monetized only through direct user engagement (...) Sustaining OSS at its current scale under widespread vibe coding requires major changes in how maintainers are paid.
so it applies to narrow slice of OSS
Comment by geldedus 3 days ago
Comment by phendrenad2 1 day ago
Comment by mellosouls 3 days ago
A recent discussion on a related topic, apparently following the same misguided idea of how OSS is motivated:
https://news.ycombinator.com/item?id=46565281
(All my new code will be closed-source from now on: 93 points, 112 comments)
Comment by p0nce 3 days ago
Comment by pixl97 3 days ago
What's interesting in reading comments like this is reading the same type of message across a bunch of different fields and aspects of life.
"When continents move, not only the weather changes"
If GenAI keeps increasing it's abilities and doesn't bankrupt a number of companies first, I think it's going to make a lot of people bubbles that encompass their entire lives. It's not difficult to imagine little pockets of hyperreality were some peoples lives are only feed by generated content and their existence starts to behave more like a video game than having any grounding in the physical. It's going to be interesting what the fractured remains of society look like in that future.
Comment by maximgeorge 3 days ago
Comment by tug2024 3 days ago
Comment by DarkSource 3 days ago
Comment by dizhn 3 days ago
> In vibe coding, an AI agent builds software by selecting and assembling open-source software (OSS),
Are they talking about indirectly due to prior training of the model? No agent I use is selecting and assembling open source software. That's more of an integration type of job not software development. Are they talking about packages and libraries? If yes, that's exactly how most people use those too.
I mean like this:
> often without users directly reading documentation, reporting bugs, or otherwise engaging with maintainers.
and then,
> Vibe coding raises productivity by lowering the cost of using and building on existing code, but it also weakens the user engagement through which many maintainers earn returns.
Maintainers who earn "returns" must be such a small niche as to be insignificant. Or do they mean things like github stars?
> When OSS is monetized only through direct user engagement, greater adoption of vibe coding lowers entry and sharing, reduces the availability and quality of OSS, and reduces welfare despite higher productivity.
Now the hypothesis is exactly the opposite. Do agents not "select and assamble" OSS anymore? And what does this have to do with how OSS is "monetized"?
> Sustaining OSS at its current scale under widespread vibe coding requires major changes in how maintainers are paid.
Sustaining OSS insofar as maintainers do it for a living requries major changes. Period. I don't see how vibe coding which makes all of this easier and cheaper is changing that equation. Quality is a different matter altogether and can still be achieved.
I am seeing a bunch of disjointed claims taken as truth that I frankly do not agree with in the first place.
What would the result of such a study even explain?
Comment by korenmiklos 3 days ago
AI agents can select and load the appropriate packages and libraries without the user even knowing the name of the library, let alone that of the developer. This reduces the visibility of developers among users, who are now less likely to give a star, sponsor, offer a job, recommend the library to others etc.
Even as a business user, say an agency building websites, I could have been a fedn of certain js frameworks, hosting meetups, buying swags, sponsoring development. I am less likely to do that if I have no idea what framework is powering the websites I build.
Our argument is that rewards fall faster with vibe coding than productivity increases. OSS developers lose motivation, they stop maintaining existing libraries, don't bother sharing new ones (even if they keep writing a lot of code for themselves).
Comment by ipaddr 3 days ago
People (the community and employers) previously were impressed because of the amount of work required. Now that respect is gone as people can't automatically tell on the surface if this is a low effort vibe code or something else.
Community engagement has dropped. Stars aren't being given out as freely. People aren't actively reading your code like they use to.
For projects done before llms you can still link effort and signal but for anything started now.. everyone assumes it's llm created. No one want to read that code and not in the same way you would read other humans. Fewer will download the project.
Many of the reasons why I wrote open source is gone. And knowing the biggest/only engagement will come from llms copying your work giving you no credit.. what's the point?
Comment by 9dev 3 days ago
Without any kind of offence implied: As maintainer of a few open source projects, I'm happy if it stops being an employability optimisation vector. Many of the people who don't code for fun but to get hired by FAANG aren't really bringing joy to others anyway.
If we end up with a small web of enthusiasts who write software for solving challenges, connecting intellectually with likeminded people, and altruism—then I'm fine with that. Let companies pay for writing software! Reduce the giant dependency chains! Have less infrastructure dedicated to distributing all that open source code!
What will remain after that is the actual open source code true to the idea.
Comment by ipaddr 3 days ago
CNC saws use to take pencil draws as input and now they can handle files. People always made handmade furniture while CNCs existed.
Open source projects around a need will continue. Things like youtube downloader fills a need. But many projects were showing off what you as a developer can write to impress a community. Those are dead. Projects that showcased new coding styles or ways to do things are dead.
Faang open source employment was never a thing. Faang filtered by leetcode, referrals, clout and h1 visas.
Comment by 9dev 3 days ago
Good! We like things filling a need.
> But many projects were showing off what you as a developer can write to impress a community. Those are dead.
Good! The world doesn't need more vanity.
> Projects that showcased new coding styles or ways to do things are dead.
That I doubt. People who want to share ideas for the sake of discussing these on an intellectual level will absolutely continue to do so. An LLM doesn't help in any way in this domain.
> Faang open source employment was never a thing. Faang filtered by leetcode, referrals, clout and h1 visas.
Replace FAANG with SMB companies, then. The point is that people were only doing open source as a means to bolster their CV. There is nothing lost to the open source community as a whole if these people leave. Yes, the amount of contributions and projects will shrink. I'm arguing that that's okay: The vital pieces that need to be maintained will be maintained by companies that are forced to pay for that maintenance, and people who enjoy making and participating in open source software development will continue to do so.
Comment by ipaddr 3 days ago
99 percent of developers go to the effort of open sourcing for vanity reasons. They want to show off what they made hope others like it.
The world of solutions to core problems is such a small part of open source
The other thing to consider is you will have mountains of AI slope like tailwind landing pages and alike crowding out original thought.
Comment by em-bee 3 days ago
but for others coding will become an art and craft like woodworking or other hobbies that require mastery.
Comment by ipaddr 3 days ago
Comment by em-bee 2 days ago
Comment by ipaddr 2 days ago
People will still write code for a variety of reasons but why bother sharing if no one is listening.
Comment by Cthulhu_ 3 days ago
Nobody cares if you wrote 5000 LOC, what they care about is what it does, how it does it, how fast and how good it does it, and none of those qualifiers are about volume.
Comment by tosh 3 days ago
more open source, better open source
perhaps also more forking (not only absolute but also relative)
contribution dynamics are also changing
I'm fairly optimistic that generative ai is good for open source and the commons
what I'm also seeing is open source projects that had not so great ergonomics or user interfaces in general are now getting better thanks to generative ai
this might be the most directly noticeable change for users of niche open source
Comment by avaer 3 days ago
Comment by positron26 3 days ago
Comment by avaer 3 days ago
Except it's on Github and it's forks and starts.
Comment by mr_spothawk 3 days ago
Also, it's a scarcity mindset.
I don't agree that sibling to my comment: "make money by getting papers cited". it is not a long-term solution, much as Ad revenue is broken model for free software, also.
I'm hopeful that we see some vibe-coders get some products out that make money, and then pay to support the system they rely on for creating/maintaining their code.
Not sure what else to hope for, in terms of maintaining the public goods.
Comment by positron26 2 days ago