257 pointsby rogueleaderrMar 29, 2026

42 Comments

floathubMar 29, 2026
Free software has never mattered more.

All the infrastructure that runs the whole AI-over-the-internet juggernaut is essentially all open source.

Heck, even Claude Code would be far less useful without grep, diff, git, head, etc., etc., etc. And one can easily see a day where something like a local sort Claude Code talking to Open Weight and Open Source models is the core dev tool.

TacticalCoderMar 29, 2026
> All the infrastructure that runs the whole AI-over-the-internet juggernaut is essentially all open source.

Exactly.

> Heck, even Claude Code would be far less useful without grep, diff, git, head, etc.

It wouldn't even work. It's constantly using those.

I remember reading a Claude Code CLI install doc and the first thing was "we need ripgrep" with zero shame.

All these tools also all basically run on top of Linux: with Claude Code actually installing, on Windows and MacOS, a full linux VM on the system.

It's all open-source command line tools, an open-source OS and piping program one to the other. I'm on Linux on the desktop (and servers ofc) since the Slackware days... And I was right all along.

gesisMar 29, 2026
The primary selling point of unix and unix-like operating systems has always been composability.

Without the ability to string together the basic utilities into a much greater sum, Unix would have been another blip.

CobrastanJorjiMar 29, 2026
It's not just that open source code is useful in an age of AI, it's that the AI could only have been made because of the open source code.
andoandoMar 29, 2026
Why isnt LLM training itself open sourced? With all the compute in the world, something like Folding@home here would be killer
DesaiAshuMar 29, 2026
data bandwidth limits distributed training under current architectures. really interesting implications if we can make progress on that
dogcomplexMar 30, 2026
Limits but doesn't prohibit. See https://www.primeintellect.ai/blog/intellect-3 - still useful and can scale enormously. Takes a particular shape and relies heavily on RL, but still big.
throwaway27448Mar 30, 2026
It's either illegal or extremely expensive to source quality training material.
m4rtinkMar 30, 2026
Yeah, turns out if you want to train a model without scrapping and overloading the whole of Internet while ignoring all the licenses and basic decency is actually hard & expensive!
mike_hearnMar 30, 2026
It is in some cases. NVIDIA's models are open source, in the truest sense that you can download the training set and training scripts and make your own.
doctorwho42Mar 30, 2026
Well it is, it's in the name "OpenAI". /S
PaulDavisThe1stMar 30, 2026
> Free software has never mattered more.

But the Libre part of Free Software has never mattered less, at least so TFA argues and while I could niggle with the point, it's not wrong.

pwdisswordfishyMar 30, 2026
Wow, some corps could offload some of their costs to "the community" (unpair labor), while end users are as disenfranchised as ever! How validating!
threethirtytwoMar 29, 2026
I think the opposite. It will make all software matter less.

If trendlines continue... It will be faster for AI to vibe code said software to your customized specifications than to sign up for a SaaS and learn it.

"Claude, create a project management tool that simplifies jira, customize it to my workflow."

So a lot of apps will actually become closed source personalized builds.

SchemaLoadMar 29, 2026
And then you get a new hire who already knows the common SaaS products but has to re learn your vibe coded version no one else uses where no information exists online.

There is a reason why large proprietary products remain prevalent even when cheaper better alternatives exist. Being "industry standard" matters more than being the best.

ares623Mar 29, 2026
As the kids say: "let them cook"
threethirtytwoMar 30, 2026
The new hire will just vibe code a new solution that translates your solution into something he prefers. Every new hire will have his own.
SchemaLoadMar 30, 2026
This will all end well I'm sure
threethirtytwoMar 30, 2026
It will. By translation I mean like a front end client that translates the api into a user interface they prefer. They will build something localized to their own workflow. If it doesn't end well it's localized to them only.
uduniMar 29, 2026
There's too much value in familiar UX. "Don't make the user think" is the golden rule these days. People used to have mental bandwidth for learning new interfaces... But now people expect uniformity
ahartmetzMar 29, 2026
They expect low complexity. Uniformity has greatly declined in the last 30 years or so. How do you even tell what is or isn't clickable, ffs?
zadikianMar 29, 2026
But then all your local stuff is based on open-source software, unlike the SaaS which is probably not all the way open.

I've always preferred my stack to be on the thinner, more vanilla, less prebuilt side than others around me, and seems like LLMs are reinforcing that approach now.

nine_kMar 29, 2026
When you want reliable, battle-tested software, you will notice the difference.
wmeredithMar 30, 2026
Implying that JIRA is reliable is... something.
Joel_MckayMar 29, 2026
Due to copyright laws and piracy bleed-through, one can't safely license "AI" output under some other use-case without the risk of getting sued or DMCA strikes. You can't make it GPL, or closed source... because it is not legally yours even if you paid someone for tokens.

Like all code-generators that came before, the current LLM will end up a niche product after the hype-cycle ends. "AI" only works if the models are fed other peoples real works, and the web is already >52% nonsense now. They add the Claude-contributor/flag to Git projects, so the scrapers don't consume as much of its own slop. ymmv =3

coffeefirstMar 29, 2026
This isn’t going to happen.

I can already build a ticket tracker in a weekend. I’ve been on many teams that used Jira, nobody loves Jira, none of us ever bothered to DIY something good enough.

Why?

Because it’s a massive distraction. It’s really fun to build all these side apps, but then you have to maintain them.

I’m guessing a lot of vibeware will be abandoned rather than maintained.

threethirtytwoMar 30, 2026
Who said you’re building it? You’re telling your AI to build it while you go play golf or something.
hparadizMar 30, 2026
The hard part has always been shipping, buttoning things up, doing the design. Not the idea per say. And then if any of it is successful and starts making money guess who you're gonna call to maintain it?
threethirtytwoMar 30, 2026
These are local systems. Think of it like vibe coding your personal GUI or CLI. Each programmer uses their own custom build. There's no maintenance except only for themselves.

You typically use an off the shelf project management software because it's too time consuming to build one catered to your own preferences. But with AI, it just does it for you. I'm talking about custom one off personal solutions readily done because of AI executing on it for you.

roncesvallesMar 30, 2026
That's not how anything works.
coffeefirstMar 30, 2026
Seriously. For starters I don’t even play golf!

And it’s all downhill from there…

threethirtytwoMar 30, 2026
That's how vibe coding works.
hrmtst93837Mar 30, 2026
Teams skip DIY trackers because maintenance is expensive now if agents cut that cost a lot more teams will tolerate the distraction.
not_paid_by_ytMar 30, 2026
if the trendlines continue on atmospheric greenhouse gases we will all be dead from climate change so I really do hope the world is a little bit more complicated than trendlines just extrapolating out. Interestingly enough that might actually be bad for OpenAI since it will be difficult to sell their product if their customers are dying from heat stroke.
threethirtytwoMar 30, 2026
You hope. But you need to think realistically. Not hopefully.

Trendlines will continue. Even the one for greenhouse gases. That is the most realistic scenario. In fact the trendline for greenhouse gases is even stronger than AI. I am far more confident about greenhouse gases continuing to rise than I am for AI.

Telling me how another trendline points to a shitty reality doesn't change the fact that the shitty reality is still reality. It's a common mistake in debate.

I haven't stated whether I hope for one reality or the other. I'm simply stating the most probable future. You haven't even disagreed with me.

not_paid_by_ytMar 30, 2026
we are probably not going to keep pumping CO2 out at the rate of the worst case scenario of business because believe it or not companies are aware there is little profit to be made from dead consumers.

The realistic scenarios are still pretty bad but less so.

Just assuming LLMs will scale to the point of being more useful as all these companies are banking on is not unfounded but it's also just an educated guess because of their own optimistic prediction they will be able to bring about a new fascist world they are in charge of before they run out of money. It's very likely we are all screwed but it's also very likely LLMs will not scale up fast enough so they can replace the majority of human labor.

pdntspaMar 29, 2026
What's the chance this website is powered by postgresql?
girvoMar 29, 2026
Hopefully low, it's a static blog.

(I know this isn't the actual point of your comment, apologies!)

keelerMar 29, 2026
It's a WordPress blog & backed by MySQL.
hparadizMar 30, 2026
So PHP.
SchemaLoadMar 29, 2026
Maybe, but I don't really believe users can or want to start designing software, if it was even possible which today it isn't really unless you already have software dev skills.

That would basically make users a product manager and UX designer, which they aren't really capable of currently. At most they will discover what they think they want isn't what they actually want.

BockitMar 29, 2026
It’s such a fun time to have 1+ decade(s) of experience in software. Knowing what simple and good are (for me), and being able to articulate it has let me create so much personal software for myself and my family. It has really felt like turning ideas into reality, about as fast as I can think of them or they can suggest them. And adding specific features, just for our needs. The latest one was a slack canvas replacement, as we moved from slack to self-hosted matrix + element but missed the multiplayer, persistent monthly notes file we used. Even getting matrix set up in the first place was a breeze.

$20/month with your provider of choice unlocks a lot.

Edit: the underlying point being, yes to the article. Either building upon the foundations of open source to making personal things, or just modifying a fork for my own needs.

CryeMar 29, 2026
Couldn't agree more. I'm building open source software for the grid, contributing in a way that feels like it could truly make a difference, while building momentum for open standards. It doesn't feel like work, just creativity and problem solving. On top of that, I can just build stuff for fun. Kids want a Minecraft mod? Let's build it and learn a thing or two on the way.
wickedsightMar 30, 2026
Where/how do you host your family apps to have them conveniently available to your household? This is the thing I'm struggling with most.
BockitMar 30, 2026
I have a Mac mini and I put every device that needs access on Tailscale
heliumteraMar 29, 2026
Oh yeah, sure, nothing scream freedom louder than following anthropic and openai suggestions without a second thought.
FergusArgyllMar 29, 2026
tl-didn't finish but I absolutely do this already. Much of the software I use is foss and codex adjusts it to my needs. Sometimes it's really good software and I end up adding something that already exists. Whatever, tokens are free...
woeiruaMar 29, 2026
I’m not so sure… what I see as more likely is that coding agents will just strip parts from open source libraries to build bespoke applications for users. Users will be ecstatic because they get exactly what they want and they don’t have to worry about upstream supply chain attacks. Maintainers get screwed because no one contributes back to the main code base. In the end open source software becomes critical to the ecosystem, but gets none of the credit.
sigbottleMar 29, 2026
But the users would have to maintain their own forks then. Unless you stream back patches into your forks, which implies there's some upstream being maintained. Software doesn't interoperate and maintain itself for free - somebody's gotta put in the time for that.

I think as long as AI isn't literal AGI, social pressures will keep projects alive, in some state. There definitely is something scary about stealing entire products as a mean for new market domination - e.g. steal linux then make a corporate linux, and force everybody to contribute to corporate linux only (many linux contributors are paid by corporations, after all), and make that the new central pointer. That might be worst case scenario - then Microsoft, in collusion (which I admit is far fetched, but def possible), could completely adopt linux for servers and headless compute, and enforce very strict hardware restrictions such that only Windows works.

LegionMammal978Mar 29, 2026
> But the users would have to maintain their own forks then.

I suppose the idea would be, they don't have to maintain it: if it ever starts to rot from whatever environmental changes, then they can just get the LLM to patch it, or at worst, generate it again from scratch.

(And personally, I prefer writing code so that it isn't coupled so tightly to the environment or other people's fast-moving libraries to begin with, since I don't want to poke at all of my projects every other year just to keep them functional.)

sigbottleMar 30, 2026
The LLM can a priori test on all possible software and hardware environments, test all possible edge cases for deployment, get feedback from millions of eyes on the project explicitly or implicitly via bug reports and usage, find good general case use features given the massive amounts of data gathered through the community of where the project needs to go next, etc?

Even in a world with pure LLM coding, it's more likely that LLMs maintain an open source place for other LLMs to contribute to.

You're forgetting that code isn't just a technical problem (well, even if it was, that would be a wild claim that goes against all hardness results known to humans given the limits of a priori reasoning...)

gf000Mar 30, 2026
Yeah, the funny thing that Linux being open-source is absolutely in line with capitalism. Just look at the list of maintainers - they are almost all paid employees of gigacorps.

It is just an optimization that makes sense -- writing an OS that is compatible with all sorts of hardware is hard, let alone one that is performant, checked for vulnerabilities, etc.

Why would each gigacorp waste a bunch of money on developing their own, when they could just spend a tiny bit to improve a specific area they deeply care about, and benefit from all the other changes financed by other companies.

m4rtinkMar 30, 2026
And the GPL makes it all work - as no single gigacorp can just take the whole and legally run with it for their gain, like they could if it was say MIT or BSD licensed.

So you have direct competitors all contributing to a common project in harmony.

LegionMammal978Mar 30, 2026
> The LLM can a priori test on all possible software and hardware environments, test all possible edge cases for deployment, get feedback from millions of eyes on the project explicitly or implicitly via bug reports and usage, find good general case use features given the massive amounts of data gathered through the community of where the project needs to go next, etc?

Even if that's the ideal (and a very expensive one in terms of time and resources), I really don't think it accurately describes the maintainers of the very long tail of small open-source projects, especially those simple enough for the relevant features to be copied into a few files' worth of code.

Like, sure, projects like Linux, LLVM, Git, or the popular databases may fit that description, but people aren't trying to vendor those via LLMs (or so I hope). And in any case, if the project presently fulfills a user's specific use case, then it "going somewhere next" may well be viewed as a persistent risk.

woeiruaMar 30, 2026
Agents can clearly strip out functionality from libraries already. They can certainly backport patches to whatever parts you strip out.

The advantage of decoupling from supply chain attacks is so large that I expect this to be standard practice as soon as later this year.

hparadizMar 30, 2026
Agents can read the binary that makes up a compiled file and detect behavior directly from that. I've been doing it to inspect my own builds for the presence of a feature.
woeiruaMar 30, 2026
Yeah. I think people are deluding themselves as to how capable these models are now.
apatheticonionMar 29, 2026
Having over a decade of open source software I've written freely available online, I actually really appreciate the value that AI && LLMs have provided me.

The thing that leaves a bad taste in my mouth is the fact that my works were likely included in the training data and, if it doesn't violate my licenses (GNU 2/3), it certainly feels against the spirit of what I intended when distributing my works.

I was made redundant recently "due to AI" (questionable) and it feels like my works in some way contributed to my redundancy where my works contributed to the profits made by these AI megacorps while I am left a victim.

I wish I could be provided a dividend or royalty, however small, for my contribution to these LLMs but that will never happen.

I've been looking for a copy-left "source available" license that allows me to distribute code openly but has a clause that says "if you would like to use these sources to train an LLM, please contact me and we'll work something out". I haven't yet found that.

I'm guessing that such a license would not be enforceable because I am not in the US, but at least it would be nice to declare my intent and who knows what the future looks like.

bluefirebrandMar 29, 2026
> I've been looking for a copy-left "source available" license that allows me to distribute code openly but has a clause that says "if you would like to use these sources to train an LLM, please contact me and we'll work something out". I haven't yet found that

Frankly do you think AI companies have even the remotest amount of respect for these licenses anyways? They will simply take your code if it is publicly scrapeable, train their models, exactly like they have so far. Then it will be up to you to chase them down and try to sue or whatever. And good luck proving the license violation

I dunno. I just don't really believe that many tech companies these days are behaving even remotely ethically. I don't have much hope that will change anytime soon

apatheticonionMar 29, 2026
I wonder if there is a "loaded" lawsuit here that could be a win-win for license enforcement case law in LLMs.

Take a litigious company like Nintendo. If one was to train an LLM on their works and the LLM produces an emulator, that would force a lawsuit.

If Nintendo wins, then LLMs are stealing. If Nintendo loses, then we can decompile everything.

bluefirebrandMar 30, 2026
You're forgetting the option where the LLM companies pay Nintendo a silly amount of money for permission and Nintendo's executives take that as a win
archagonMar 30, 2026
Traditionally, large corporations have taken very conservative legal stances with regard to integrating e.g. A/GPL code, even when there's almost no risk.

If my license explicitly says "any LLM output trained on this code is legally tainted," I feel like BigAICorp would be foolish to ignore it. Maybe I couldn't sue them today, but are they confident this will remain the case 5, 10, 20 years from now? Everywhere in the world?

63stackMar 30, 2026
Github has posted that they will now train on everyone's data (even private) unless you opt out (until they change their mind on that). Anthropic has been training on your data on certain tiers already. Meta bittorrented books to train their models.

Surely if your license says "LLM output trained on this code is legally tainted", it is going to dissuade them.

archagonMar 30, 2026
No, it won’t dissuade them. But when we finally get the chance to legally beat the shit out of these companies, I want to reserve my place in line.

Alternatively, they can learn to trust me on this and simply exclude/evict my code from the training corpus.

advaelMar 29, 2026
I think there's no meaningful case by the letter of the law that use of training data that include GPL-licensed software in models that comprise the core component of modern LLMs doesn't obligate every producer of such models to make both the models and the software stack supporting them available under the same terms. Of course, it also seems clear in the present landscape that the law often depends more on the convenience of the powerful than its actual construction and intent, but I would love to be proven wrong about that, and this kind of outcome would help
BobbyJoMar 29, 2026
If the rise of Draft Kings and Polymarket/Kalshi have taught me anything, it's that the law becomes meaningless at scale. Sad.
advaelMar 29, 2026
Sure, but that's more a result of policy decisions than an inevitable result of some natural law. Corporate lawlessness has been reined in before and it can be again
apatheticonionMar 29, 2026
I'm struggling to parse the double negative in that statement, haha.

Are you saying that you believe that untested but technically; models trained on GPL sources need to distribute the resulting LLMs under GPL?

advaelMar 29, 2026
Yes. Double negative intended for emphasis here, but apologies if it's confusing
eruMar 30, 2026
Well, most companies never distribute their models. So GPL doesn't kick in.
vova_hn2Mar 30, 2026
I think that the claim that they make is that once a model is "contaminated" with GPL code, every output it ever produces should be considered derived from GPL code, therefore GPL-licensed as well.
Tadpole9181Mar 30, 2026
So GitHub and Windows and IDEs need to be open source because they can output FOSS code? That's obviously rediculous.

If an AI outputs copyrighted code, that is a copyright violation. And if it does and a human uses it, then you are welcome to sue the human or LLM provider for that. But you don't get to sue people for perceived "latent" thought crimes.

vova_hn2Mar 30, 2026
First of all, I'm not advocating for this claim, I'm merely trying to clarify what other people say.

That being said, I don't think that your analogy is valid in this case.

> GitHub and Windows and IDEs need to be open source because they can output FOSS code

They can output FOSS code, but they themselves are not derived from FOSS code.

It can be argued that the weights of a model is derived from training data, because they contain something from the training data (hard to say what exactly: knowledge, ideas, patterns?)

It can also be argued that output is derived from weights.

If we accept both of those claims, then GPL training data -> GPL weighs -> every output is GPL

> If an AI outputs copyrighted code

Again, the issue is not what exactly does AI output, but where it comes from.

eruMar 30, 2026
It would be relatively easy to scan the output of the LLM for copyrighted material, before handing it to the user.

(I say 'relatively easy'. Not that it would be trivial.)

gottheUIbluesMar 30, 2026
If that theory holds - have to ensure that the models have not been trained on any code that is licensed incompatibly with the GPL, in which case the models could not be distributed at all
hparadizMar 30, 2026
Derivative work.
throwaway27448Mar 30, 2026
Let's cut the rot off at the root rather than pretending like the fruit is going to nourish us.
thfuranMar 30, 2026
I’m not entirely clear on what you’re suggesting abolishing here, copyright, AI, the companies making the frontier models?
m4rtinkMar 30, 2026
Yes. ;-)
throwaway27448Mar 30, 2026
Intellectual property never made much sense to begin with. But it certainly makes no sense now, where the common creator has no protections against greedy corporate giants who are happy to wield the full weight of the courts to stifle any competition for longer than we'll be alive.

Or, in the case of LLMs, recklessly swing about software they don't understand while praying to find a business model.

not_paid_by_ytMar 30, 2026
hey just don't try to copy their LLM by distilling it, cause that's "theft", if we weren't all doomed anyways this industry would have never been allowed to exist in the first place, but I guess this is just what the last few decades of our civilization will look like.
vova_hn2Mar 30, 2026
> hey just don't try to copy their LLM by distilling it, cause that's "theft"

They can call it whatever they want, but I don't think that it is illegal.

As1287Mar 30, 2026
Poor billionaire Rowling has no protections against the evil corporations. Everyone using this argument has no clue about artists and and writers.

Yes, corporations take a large cut, but creative people welcomed copyright and made the bargain and got fame in the process. Which was always better for them than let Twitch take 70% and be a sharecropper.

Silicon Valley middlemen are far worse than the media and music industry.

graemepMar 30, 2026
The individuals who get rich from copyright are a rarity.

Most mid-list authors make very little from copyright. A lot of the "authors" who make a lot of money from writing are celebs who slap their name on a ghost written work.

> Which was always better for them than let Twitch take 70% and be a sharecropper.

Copyright predates Twitch or giant corporations and was designed to protect the profits of the publishers from the start.

https://en.wikipedia.org/wiki/Statute_of_Anne

lstoddMar 30, 2026
This is so, but note that publishers were the "giant corporations" of the time.

In this nothing changed. Authors never were and still are not the point of copyright/IP.

graemepMar 30, 2026
Comparatively big for the time, but very small compared to publishing companies how.
oblioMar 30, 2026
Giant corporations started in the early 1600s and back then they had gunboats :-p
graemepMar 30, 2026
there were very few of them though. They constitute a far larger proportion of the economy now.
jacquesmMar 30, 2026
The reason you mention 'poor billionaire Rowling' is most likely because she's the only billionaire author that you know by name. If authors regularly became billionaires you'd have left out that name.
eruMar 30, 2026
> [...] greedy corporate giants who are happy to wield the full weight of the courts to stifle any competition for longer than we'll be alive.

How is any of this new?

cogman10Mar 30, 2026
If there was going to be a case, it's derivative works. [1]

What makes it all tricky for the courts is there's not a good way to really identify what part the generated code is a derivative of (except in maybe some extreme examples).

[1] https://en.wikipedia.org/wiki/Derivative_work

felipeeriasMar 30, 2026
One could carefully calculate exactly how much a given document in the training set has influenced the LLM's weights involved in a particular response.

However, that number would typically be very very very very small, making it hard to argue that the whole model is a derivative of that one individual document.

Nevertheless, a similar approach might work if you took a FOSS project as a whole, e.g. "the model knows a lot about the Linux kernel because it has been trained on its source code".

However, it is still not clear that this would be necessarily unlawful or make the LLM output a derivative work in all cases.

It seems to me that LLMs are trained on large FOSS projects as a way to teach them generalisable development skills, with the side effect of learning a lot about those particular projects.

So if I used a LLM to contribute to the kernel, clearly it would be drawing on information acquired during its training on the kernel's code source. Perhaps it could be argued that the output in that case would be a derivative?

But if I used a LLM to write a completely unrelated piece of software, the kernel training set would be contributing a lot less to the output.

cogman10Mar 30, 2026
> One could carefully calculate exactly how much a given document in the training set has influenced the LLM's weights involved in a particular response.

Not really.

Think of, for example, a movie like "who framed roger rabbit". It had intellectual property from all over. Had the studios not gotten the rights from each or any of those properties, they could have been sued for copyright infringement. It's not really a question of influence.

So yeah, while the LLM might have been trained on the kernel, it was also likely trained on code with commercial licenses. Conversely, because was trained on code with GPL licenses, that might mean commercial software with LLM contributions need to inherit the GPL to be legal (and a bunch of other licenses).

It's a big old quagmire and I think lawyers haven't caught up enough with how LLMs work to realize this.

not_paid_by_ytMar 30, 2026
That's always what laws existed for, a law is just a formal way of saying "we will use violence against you if you do something we don't like" and that has always going to be primary written by and for the people that already have the power to do that, it's not the worst, certainly better than Kings just being able to do as they please.
vova_hn2Mar 30, 2026
> certainly better than Kings just being able to do as they please

That's debatable. In case of a king you always know whom to blame and who has full responsibility. No opportunity to hide behind "well, you voted for this" or "I'm not making the laws, I'm merely enforcing them".

tpmoneyMar 30, 2026
> I think there's no meaningful case by the letter of the law that use of training data that include GPL-licensed software in models that comprise the core component of modern LLMs doesn't obligate every producer of such models to make both the models and the software stack supporting them available under the same terms.

Why do you think "fair use" doesn't apply in this case? The prior Bartz vs Anthropic ruling laid out pretty clearly how training an AI model falls within the realm of fair use. Authors Guild vs Google and Authors Guild vs HathiTrust were both decided much earlier and both found that digitizing copyrighted works for the sake of making them searchable is sufficiently transformative to meet the standards of fair use. So what is it about GPL licensed software that you feel would make AI training on it not subject to the same copyright and fair use considerations that apply to books?

ronsorMar 30, 2026
> So what is it about GPL licensed software that you feel would make AI training on it not subject to the same copyright and fair use considerations that apply to books?

The poster doesn't like it, so it's different. Most of the "legal analysis" and "foregone conclusions" in these types of discussions are vibes dressed up as objective declarations.

input_shMar 30, 2026
You seem like the type of person that will believe anything as long as someone cites a case without looking into it. Bartz v Anthropic only looked at books, and there was still a 1.5 billion settlement that Anthropic paid out because it got those books from LibGen / Anna's Archive, and the ruling also said that the data has to be acquired "legitimately".

Whether data acquired from a licence that specifically forbids building a derivative work without also releasing that derivative under the same licence counts as a legitimate data gathering operation is anyone's guess, as those specific circumstances are about as far from that prior case as they can be.

eruMar 30, 2026
As long as they don't distribute the model's weights, even a strict interpretation of the GPL should be fine. Same reason Google doesn't have to upstream changes to the Linux kernel they only deploy in-house.
fsfloverMar 30, 2026
How about AGPL?
oblioMar 30, 2026
But LLMs do distribute the derived code they generate outside of their company. That's their entire point.
AkronymusMar 30, 2026
But wouldn't that be like some company using gpl licensed code to host a code generator for something? At least in a legal interpretation. Or is that different?
advaelMar 30, 2026
I mean, is the case you're making that you can run a SaaS business on GPL-derived code without fulfilling GPL obligations because you're not distributing a binary?
AkronymusMar 30, 2026
I guess I am. I genuinly am just a layperson trying to look at what the law would say, so everything is speculation.
advaelMar 30, 2026
If true that would seem to invalidate the entire GPL, but even by that logic, a website (such as chatGPT) distributes javascript that runs the code, and programs like claude code also do so. Again, if you can slip the GPL's requirements through indirection like having your application go phone home to your server to go get the infringing parts, the GPL would essentially unenforceable in... most contexts
oblioMar 30, 2026
And why would that be different or allowed? Sure, you get all the code you want, GPL licensed.

Everybody is trying to have their cake and eat it, too, by license laundering.

Heck, money laundering means you at least lose some of the money.

AkronymusMar 30, 2026
I have no idea. I genuinly was asking out of curiosity on what the law actually means for that while speculating.
ronsorMar 30, 2026
Have you actually read the text of the GPL?

> This License acknowledges your rights of fair use or other equivalent, as provided by copyright law.

It is legitimate to acquire GPL software. The requirements of the license only occur if you're distributing the work AND fair use does not apply.

Training certainly doesn't count as distribution, so the buck passes to inference, which leaves us dealing with substantial similarity test, and still, fair use.

advaelMar 30, 2026
Broadly speaking, GPL is a license that has specific provisions about creating derivative software from the licensed work, and just saying "fair use" doesn't exempt you from those provisions. More specifically, an advertised use case (in fact, arguably the main one at this stage) of the most popular closed models as they're currently being used is to produce code, some of which is going to be GPL licensed. As such, the code used is part of the functionality of the program. The fact that this program was produced from the source code used by a machine learning algorithm rather than some other method doesn't change this fundamental fact.

The current supreme court may think that machine learning is some sort of magic exception, but they also seem to believe whatever oligarchs will bribe them to believe. Again, I doubt the law will be enforced as written, but that has more to do with corruption than any meaningful legal theory. Arguments against this claim seem to ignore that courts have already ruled these systems to not have intellectual property rights of their own, and the argument for fair use seems to rely pretty heavily on some handwavey anthropomorphization of the models.

mr_toadMar 30, 2026
> Broadly speaking, GPL is a license that has specific provisions about creating derivative software from the licensed work, and just saying "fair use" doesn't exempt you from those provisions.

Broadly speaking, yes it does. The whole point of fair use is that you don’t need a license.

davempMar 30, 2026
Claiming LLMs are fair use is ridiculous bordering on ignorant or disingenuous.

Here’s the 4 part test from 17 U.S.C. § 107:

1. the purpose and character of the use, including whether such use is of a commercial nature or is for nonprofit educational purposes;

Fail. The use is to make trillions of dollars and be maximally disruptive.

2. the nature of the copyrighted work;

Fail. In many cases at least, the copy written code is commercial or otherwise supports livelihoods; and is the result much high skill labor with the express stipulation for reciprocity.

3. the amount and substantiality of the portion used in relation to the copyrighted work as a whole; and

Fail. They use all of it.

4. the effect of the use upon the potential market for or value of the copyrighted work.

Fail to the extreme. There is already measurable decline in these markets. The leaders explicitly state that they want to put knowledge workers out of business.

- - -

Hell, LLMs don’t even pass the sniff test.

The only reason this stuff is being entertained is some combination of the prisoner’s dilemma and more classic greed.

lumaMar 30, 2026
You’re mixing up “using” with “copying”. You are allowed to “use” all of a book or movie or code by listening to or watching or reviewing the whole thing. Copyright protects copies. The legal claim here is than training an LLM is sufficiently transformative such that it cannot be construed as a copy.
davempMar 30, 2026
I replied to someone saying that it’s fair use, which presupposes that it’s a derivative work.
joshuaccMar 30, 2026
These are factors to be considered, not pass/fail questions.
cxrMar 30, 2026
This comment highlights a basic dilemma about how and where to spend your time.

Here's a basic rule of thumb I recommend people apply when it comes to these sorts of long, contentious threads where you know that not every person showing up to the conversation is limiting themselves to commenting about things they understand and that involve some of the most tortured motivated reasoning about legal topics:

If the topic is copyright and someone who is speaking authoritatively has just used the words "copy written", then ignore them. Consider whether you need to be anywhere in the conversation at all, even as a purely passive observer. Think about all the things you can do instead of wasting your time here, where the stakes for participation are so low because nothing that is said here really matters. Go do something productive.

davempMar 30, 2026
Yet you still wasted your own time and everyone else’s time with a reply that has even less substance.

I was making an argument based on quotes from the actual legal code and you’re saying pions who don’t use the exact correct terminology shouldn’t even consider what should or shouldn’t be legal? What a load of junk. This is a democracy. We’re supposed to be engaging with it.

shaknaMar 30, 2026
Bartz v Anthropic explicitly held ruling on fair use. It is not precedent, here.
derektankMar 30, 2026
I’m not a lawyer, but I read the decision, and how is this section not a ruling on fair use?

“To summarize the analysis that now follows, the use of the books at issue to train Claude and its precursors was exceedingly transformative and was a fair use under Section 107 of the Copyright Act. And, the digitization of the books purchased in print form by Anthropic was also a fair use but not for the same reason as applies to the training copies. Instead, it was a fair use because all Anthropic did was replace the print copies it had purchased for its central library with more convenient space-saving and searchable digital copies for its central library — without adding new copies, creating new works, or redistributing existing copies. However, Anthropic had no entitlement to use pirated copies for its central library. Creating a permanent, general-purpose library was not itself a fair use excusing Anthropic’s piracy.”

Or in the final judgement, “This order grants summary judgment for Anthropic that the training use was a fair use. And, it grants that the print-to-digital format change was a fair use for a different reason.”

shaknaMar 30, 2026
There's two parts here.

The first:

> it was a fair use because all Anthropic did was replace the print copies it had purchased for its central library

It is only fair use where Anthropic had already purchased a license to the work. Which has zero to do with scraping - a purchase was made, an exchange of value, and that comes with rights.

The second, which involves a section of the judgement a little before your quote:

> And, as for any copies made from central library copies but not used for training, this order does not grant summary judgment for Anthropic.

This is where the court refused to make any ruling. There was no exchange of value here, such as would happen with scraping. The court made no ruling.

jerfMar 30, 2026
You sound like you're citing the general Internet understanding of "fair use", which seems to amount to "I can do whatever I like to any copyrighted content as long as maybe I mutilate it enough and shout 'FAIR USE!' loudly enough."

On the real measures of "fair use", at least in the US: https://fairuse.stanford.edu/overview/fair-use/four-factors/ I would contend that it absolutely face plants on all four measures. The purpose is absolutely in the form of a "replacement" for the original, the nature is something that has been abundantly proved many times over in court as being something copyrightable as a creative expression (with limited exceptions for particular bits of code that are informational), the "amount and substantiality" of the portions used is "all of it", and the effect of use is devastating to the market value of the original.

You may disagree. A long comment thread may ensue. However, all I really need for my point here is simply that it is far, far from obvious that waving the term "FAIR USE!" around is a sufficient defense. It would be a lengthy court case, not a slam-dunk "well duh it's obvious this is fair use". The real "fair use" and not the internet's "FAIR USE!" bear little resemblance to each other.

A sibling comment mentions Bartz v. Anthropic. Looking more at the details of the case I don't think it's obvious how to apply it, other than as a proof that just because an AI company acquired some material in "some manner" doesn't mean they can just do whatever with it. The case ruled they still had to buy a copy. I can easily make a case that "buying a copy" in the case of a GPL-2 codebase is "agreeing to the license" and that such an agreement could easily say "anything trained on this must also be released as GPL-2". It's a somewhat lengthy road to travel, where each step could result in a failure, but the same can be said for the road to "just because I can lay my hands on it means I can feed it to my AI and 100% own the result" and that has already had a step fail.

jrm4Mar 30, 2026
"Real" fair use is perhaps one of the most nebulous legal concepts possible. I haven't dived deep into software, but a cursory look at how it "works (I use that term as loosely as possible)" in music with sampling and interpolation etc immediately reveals that there's just about nothing one can rely on in any logical sense.
jonahssMar 29, 2026
I feel kind of good knowing that my code, design decisions, styles, are now part of the data shaping all software now.
__loamMar 29, 2026
I can't eat good feelings
faksrMar 29, 2026
Everyone is entitled to their own opinion. The cannibalistic $trillion companies profit from it all and no one opted in.

There are also people who want to be eaten by a literal cannibal. I say, no thanks.

apatheticonionMar 29, 2026
Me too, and I use LLMs often for personal and professional work. Knowing that colleagues are burning through $700/day worth of tokens, and a small fraction of those tokens were likely derived from my work while I get made redundant is a bit shite.
manwe150Mar 29, 2026
I’d been hoping to finally train a replacement and code myself out of a job for years. I just didn’t know I was the replacement too, working with AI.
dbetteridgeMar 30, 2026
$700 a day of tokens can't possibly be sustainable right?

That's 2X the salary of a lot of the world's software developers

apatheticonionMar 30, 2026
AUD* - so $450 USD

But yes, that's very expensive and surprising to me.

dbetteridgeMar 30, 2026
Ah a fellow Aussie, hi! Sorry to hear about the redundancy (Atlassian?).

I did implicitly assume USD but yeah still crazy cash, that'd pay for 2 junior-mid level devs in aus D=

apatheticonionMar 30, 2026
G'day, haha.

Yeah Atlassian. 1/3rd of my team were given the boot sadly. One guy had 12 years at the company - crazy times

not_paid_by_ytMar 30, 2026
and that's at the VC funded discount rate I would presume, not even true cost of those tokens without any profit.
dbetteridgeMar 30, 2026
Yeah that's the thing making my head spin, tack a 30% profit margin on that and it's 550usd per day? Probably going to be more than that for rocketship growth and investor expectations.

Is that the game? Lock in companies to this "new reality" with cheap tokens then once they fire all their devs, bait and switch to 2X the cost.

vips7LMar 30, 2026
Of course that’s the play. It’s always the play and it’s so sad to see everyone on this forum falling for it.
nzMar 30, 2026
If you read history widely (across millennia and geographies), you'll note that most of the power-contests follow this pattern[0]. In the modern industrial world, the pattern becomes exponential rather than incremental. What I'm saying is that this is not unique to AI Labs[1]. This is caused by the deeply flawed and unbalanced system that we have constructed for ourselves.

[0]: The pattern, or, as gamers would call it, the "meta", is that every ambitious person/entity wants to control as much of the economic/material surplus as possible. The most effective and efficient (effort per control) way of doing this is to make yourself into as much of a bottle-neck as humanly possible. In graph-theory this corresponds to betweenness-centrality, and you want to maximize that value. To put it in mundane terms, you want to be as much of a monopoly as you can be (Thiel is infamous for saying this, but it does check out, historically). To maximize betweenness, or to maximize monopoly, is to maximize how much society/economy depends on you. This is such a dominant strategy (game-theory term, but in modern gaming world, they might call this a "cheesy strat" -- which just means that the game lacks strategic variety, forcing players to hone that one strategy), that we even have some old laws (anti-trust, etc) designed to prevent it. And it makes a lot of sense: Standard Oil was reviled because everything in the economy either required oil or required something that did. 20th-century USA did a lot to mitigate this. It forced monopolies like ATT to fund general research like Bell Labs (still legendary) towards a public good (a kind of tax, but probably much more socially-beneficial). It also broke up the monopolies, and passed anti-profit laws (e.g. hospitals were not allowed to make a profit until 1978; I have seen in the last 10 years a tiny cancer clinic grow into a massive gleaming hospital -- a machine that transforms sickness and grief into Scrooge McDuck vaults of cash). This monopolistic tendency of the commercial sector, is a tendency towards centralization, which yields efficiency, sure, but also creates the conditions for control and rent-seeking and exploitation.

[1]: Much of the cloud-computing craze was similar in character (and also failed to deliver on some of its promises, such as reducing/replacing IT overhead (they just renamed IT to DevOps)). And Web2 itself was about creating and monopolizing a new kind of ad-channel and lead-generation-machine. There is a funny twist, that a capitalist society like the USA, has much more deeply rooted incentives to create a panopticon than communist states of the past ever did. Neither is pretty of course. The communists demanded conformity and loyalty, while the capitalists demand consumption and rent.

micromacrofootMar 29, 2026
a comforting thought but I have bills to pay
gerdesjMar 29, 2026
Taken to a hallucinated but logical conclusion, we might define a word such as "cene" to riff off of "meme" and "gene".

The c is for code. If adopted we could spend forever arguing how the c is pronounced and whether the original had a cedilla, circonflex or rhymes with bollocks, which seems somehow appropriate. Everyone uses xene instead. x is chi but most people don't notice.

koolbaMar 29, 2026
Reading this I hear The Roots playing The Seed 2.0[1] in my mind.

It’s a wild thought to think that of all the things that will remain on this earth after you’re gone, it’ll be your GPL contributions reconstituting themselves as an LLM’s hallucinations.

[1]: https://youtu.be/ojC0mg2hJCc

jaggederestMar 30, 2026
If we're being clear, it's going to be a lot more than that.

Our comments here on HN are almost certainly going to live in fame/infamy forever. The twitter firehose is a pathway to 140-character immortality essentially.

You can already summon an agent to ingest essentially an entire commenter's history, correlate it across different sites based on writing style or similar nicknames, and then chat with you as that persona, even more so with a finetune or lora. I can do that with my gmail and text message history and it becomes eerily similar to me.

History is going to be much more direct and personal in the future. We can also do this with historical figures with voluminous personal correspondence, that's possible now.

It's very interesting because I think the era before mass LLM usage but also after digitalization is going to be the most intensely studied. We've lived through a thing that is going to be on the cusp of history, for better or worse.

hparadizMar 30, 2026
Me too! I'm glad I'm not the only one.
achieriusMar 30, 2026
I'm sure you'll feel that way so long as you have an income.
wiseowiseMar 30, 2026
Coding champagne socialists everywhere now.
not_paid_by_ytMar 30, 2026
tokens will stop being given away for free at some point, writing software was always a pretty simple white collar job, so it makes sense it's one of the earlier ones to be automated, but really the axis of evil has it's shot now at ruling the world or whatever now, but if they miss it they will eventually be subject to the market and you really will need to automate a lot more than just software developers for models this large to be worth the cost.

Of course we should really be talking about using the state or otherwise to make training larger and larger models impossible. It's not in the public good if LLMs actually get good enough to replace a lot of human labor, only a small handful of billionaires and their cronies will ever benefit from that. The Luddites were not wrong after all.

vips7LMar 30, 2026
You feel good that you got to contribute to these parasitic capitalists?
phendrenad2Mar 30, 2026
I wish Anthropic or someone would take a leadership role and re-train their models without any GPL code, or at least stop doing so in the future tense.
worthless-trashMar 30, 2026
What would they learn from then ?
phendrenad2Mar 30, 2026
Non-GPL code?
reactordevMar 30, 2026
If you use GitHub, you’re automatically opted into having your code used for training. Private repo or not. You have to actually opt out and even then, will they honor that? No…
WaterluvianMar 30, 2026
They’ll just keep “accidentally” resetting the option over time.
otrasMar 30, 2026
The foreman had pointed out his best man - what was his name? - and, joking with the puzzled machinist, the three bright young men had hooked up the recording apparatus to the lathe controls. Hertz! That had been the machinist's name - Rudy Hertz, an old-timer, who had been about ready to retire. Paul remembered the name now, and remembered the deference the old man had shown the bright young men.

Afterward, they'd got Rudy's foreman to let him off, and, in a boisterous, whimsical spirit of industrial democracy, they'd taken him across the street for a beer. Rudy hadn't understood quite what the recording instruments were all about, but what he had understood, he'd liked: that he, out of thousands of machinists, had been chosen to have his motions immortalized on tape. And here, now, this little loop in the box before Paul, here was Rudy as Rudy had been to his machine that afternoon - Rudy, the turner-on of power, the setter of speeds, the controller of the cutting tool. This was the essence of Rudy as far as his machine was concerned, as far as the economy was concerned, as far as the war effort had been concerned. The tape was the essence distilled from the small, polite man with the big hands and black fingernails; from the man who thought the world could be saved if everyone read a verse from the Bible every night; from the man who adored a collie for want of children; from the man who . . . What else had Rudy said that afternoon? Paul supposed the old man was dead now - or in his second childhood in Homestead.

Now, by switching in lathes on a master panel and feeding them signals from the tape, Paul could make the essence of Rudy Hertz produce one, ten, a hundred, or a thousand of the shafts.

Kurt Vonnegut, Player Piano

archagonMar 30, 2026
> I've been looking for a copy-left "source available" license that allows me to distribute code openly but has a clause that says "if you would like to use these sources to train an LLM, please contact me and we'll work something out". I haven't yet found that.

Personally, I want a viral (GPL-style) license that explicitly prohibits use of code for LLM training/tuning purposes — with the asterisk that while current law might view LLM training as fair use, this may not be the case forever, and blatant disregard of the terms of the license should make it easier for me to sue offenders in the future.

Alternatively, this could be expressed as: the output of any LLM trained on this code must retain this license.

rpdillonMar 30, 2026
My personal take is that LLMs are so transformative that they are likely not going to qualify under derivative works and therefore GPL wouldn't hold sway. There's already some evidence that courts will consider training on copyrighted material fair use, so long as it is otherwise obtained legally, which would be the case with software licensed under GPL.

I realize this is an unpopular opinion on HN, but I believe it is best because it's a weakener interpretation of copyright law, which is overall a good thing in my view.

apatheticonionMar 30, 2026
I wonder if that would extend to training LLMs on decompiled firmware. A new clean room method unlocked?
hparadizMar 30, 2026
You can train models locally now and use open source ones and there's a robust community of people training, retraining, and generally pulling data from anywhere. And then new models get trained on old models. The models in use now are already several generations deep even further trained on code freely given by the entire industry. It's like complaining about being 1/100000th of a soup with no real proof you're even in it. Can you provide proof that a model used your code? It's like a remix of a remix of a remix.
whattheheckheckMar 30, 2026
The fact that github copilot had an option to block generated code that matched public examples and the fact that the llms can regenerate Harry Potter books verbatim means the training data is definitely "stored in a digital system of retrieval" but Goodluck actually having common sense win vs trillionaire incentive group stealing from everyone
NursieMar 30, 2026
> It's like complaining about being 1/100000th of a soup with no real proof you're even in it.

I love a good analogy, especially one that takes a complex situation in which esoteric, unusual conditions are distilled and related back to common experiences held by the reader, such that all can understand.

Next time I'm a small part of a soup I'll think of this.

visargaMar 30, 2026
> it certainly feels against the spirit of what I intended when distributing my works

You can own the works, but not the vibes. If everyone owned the vibes we would all be infringing others. In my view abstractions should not be protected by copyright, only expression, currently the abstraction-filtration-comparison standard (AFC) protects abstractions too, non-literal infringement is a thing.

Trying to own the vibes is like trying to own the functionality itself, no matter the distinct implementation details, and this is closer to patents than copyrights. But patents get researched for prior art and have limited duration, copyright is automatic and almost infinite duration.

birdfoodMar 30, 2026
These companies pirated their training material and reached settlements with the copyright holders. I imagine they’d do the same with software licenced under Not For Training terms too. It’d be up to you to find out it is happening and then pursue them legally for compensation.
627467Mar 30, 2026
> I was made redundant recently "due to AI" (questionable) and it feels like my works in some way contributed to my redundancy where my works contributed to the profits made by these AI megacorps while I am left a victim.

I think anyone here can understand and even share that feeling. And I agree with your "questionable" - its just the lame HR excuse du jour.

My 2c:

- AI megacorps aren't the only ones gaining, we all are. the leverage you have to build and ship today is higher than it was five years ago.

- It feels like megacorps own the keys right now, but that’s a temporary. In a world of autonomous agents and open-weight models, control is decentralized.inference costs continue to drop, you dont need to be running on megacorp stacks. Millions (billions?) of agents finding and sharing among themselves. How will megacorps stop?

- I see the advent of LLMs like the spread of literacy. Scribes once held a monopoly on the written word, which felt like a "loss" to them when reading/writing became universal. But today, language belongs to everyone. We aren't losing code; we are making the ability to code a universal human "literacy."

j_bumMar 30, 2026
I’m not sure if the analogy is yours, but the scribe note really struck a chord with me.

I’m not a professionally trained SWE (I’m a scientist who does engineering work). LLMs have really accelerated my ability to build, ideate, and understand systems in a way that I could only loosely gain from sometimes grumpy but mostly kind senior engineers in overcrowded chat rooms.

The legality of all of this is dubious, though, per the parent. I GPL licensed my FOSS scientific software because I wanted it to help advance biomedical research. Not because I wanted it to help a big corp get rich.

But then again, maybe code like mine is what is holding these models back lol.

TeMPOraLMar 30, 2026
Sharing for advancing humanity / benefit of society, and megacorps getting rich off it, is not either-or. On the contrary, megacorps are in part how the benefit to society materializes. After all, it's megacorps that make and distribute the equipment and the software stacks I am using to write code on, that you are using to do your research on, etc.

I find the whole line of thinking, "I won't share my stuff because then a megacorp may use it without paying me the fractional picobuck I'm entitled to", to be a strong case of Dog in the Manger mindset. And I meant that even before LLM exploded, back when people were wringing their hands about Elasticsearch being used by Amazon, back in 2021 or so.

Sharing is sharing. One can't say "oh I'm sharing this for anyone to benefit", and then upon seeing someone using it to make money, say "oh but not like that!!". Or rather, one can say, but then they're just lying about having shared the thing. "OSS but not for megacorps/aicorps" is just proprietary software. Which is perfectly fine thing to work on; what's not fine is lying about it being open.

graemepMar 30, 2026
> On the contrary, megacorps are in part how the benefit to society materializes.

That would be true if they were the product of a genuine competitive market.

In fact their strength is in eliminating competition, erecting barriers to entry, manipulating regulation, and maintaining the status quo.

> "OSS but not for megacorps/aicorps"

Who is advocating that? People just want everyone to stick to the terms of the licences.

3formMar 30, 2026
LLMs are one thing, but when you bring ES in AWS example, as outlined in the article, the problem is not the software being used; it's being _made proprietary_. It's about free and open software remaining free and open. Especially to the end user.
lentil_soupMar 30, 2026
> "OSS but not for megacorps/aicorps" is just proprietary software

why? it's not like it's binary. It could well be that it's open source but can't be used by a company of X size. I'm not a lawyer but why couldn't a license have that clause? I would still class that as being open, for some definition of open

wiseowiseMar 30, 2026
> We aren't losing code; we are making the ability to code a universal human "literacy."

LLMs making the ability to code a universal human “literacy” is like saying that Markov chain is making the ability to write a universal human “literacy”.

TeMPOraLMar 30, 2026
Comparing LLMs to Markov chains was funny in 2023.
wiseowiseMar 30, 2026
I’m not comparing LLMs to Markov chain, read again.

Coding through LLMs is like writing through Markov chains.

ori_bMar 30, 2026
> We aren't losing code; we are making the ability to code a universal human "literacy."

The same way that doordash makes kitchen skills universal.

eruMar 30, 2026
You say it like it's a bad thing.
latexrMar 30, 2026
No, they are saying it like the comparison doesn’t hold. Which it doesn’t.
ori_bMar 30, 2026
I say that like it's a thing. LLMs have the goal of replacing intellectual work with passive consumption. People seem to like that.
ori_bMar 30, 2026
Basically, the selling point of LLMs is that you no longer need to think about problems, you can skip directly to results. Anything that you have to think about while using them today is somewhere on the product roadmap, or will be.

Many people think this is a form of utopia.

eruMar 30, 2026
Just like computer is no longer a job description, yes.
latexrMar 30, 2026
> AI megacorps aren't the only ones gaining, we all are.

No, no we are not.

> the leverage you have to build and ship today is higher than it was five years ago.

I don’t want more “leverage to build and ship”, I want to live in a world where people aren’t so disconnected from reality and so lonely they have romantic relationships with a chat window; where they don’t turn off their brains and accept any wrong information because it comes from a machine; where propaganda, mass manipulation, and surveillance aren’t at the ready hands of any two-bit despot; where people aren’t so myopic that they only look at their own belly button and use case for a tool that they are incapable of recognising all the societal harms around them.

> We aren't losing code; we are making the ability to code a universal human "literacy."

No, no we are not. What we are, however, is making ever increasingly bad comparisons.

Literacy implies understanding. To be able to read and write, you need to be able to understand how to do both. LLMs just spit text which you don’t need to understand at all, and increasingly people are not even caring to try to understand it. LLM generated code in the hands of someone who doesn’t read it is the opposite of literacy.

pegasusMar 30, 2026
But LLMs can also explain code, in fact they're fantastic at that. They can also be used to build anti-censorship, surveillance-avoidance and fact-checking tools. We are all empowered by them, it's just up to us to employ them so as to nudge society towards where we'd like it to go. Instead of giving up prematurely.
mulr00neyMar 30, 2026
>I don’t want more “leverage to build and ship”, I want to live in a world where people aren’t so disconnected from reality and so lonely they have romantic relationships with a chat window; where they don’t turn off their brains and accept any wrong information because it comes from a machine; where propaganda, mass manipulation, and surveillance aren’t at the ready hands of any two-bit despot; where people aren’t so myopic that they only look at their own belly button and use case for a tool that they are incapable of recognising all the societal harms around them.

Preach. Every time I read people doing this weird LARP on this website of "you have so much more leverage, great time to be a founder" I want to put my head through the drywall.

627467Mar 30, 2026
> literacy implies understanding

Agree. Do we not understand how LLMs work? Some of us understand better than others, just like literacy is also not guaranteed just because you learned the alphabet.

Accepting the output of an LLM is really materially not different from accepting books, newspapers, opinion makers, academics at face value. Maybe different only in speed of access?

> LLM generated code in the hands of someone who doesn’t read it is the opposite of literacy.

"A popsi article title or paper abstract/conclusion in the mind of someone who doesn't read is the opposite of literacy."

latexrMar 30, 2026
I’m not sure I understand your point. Mind clarifying? It seems you might be trying to contradict what I said but are in fact only adding to it.

> just like literacy is also not guaranteed just because you learned the alphabet.

I didn’t claim learning the alphabet equals literacy, you did. Your argument comes down to “you’re not literate if you’re not literate”. Which, yes, of course.

> Accepting the output of an LLM is really materially not different from (…)

Multiple things can be true at once. If someone says “angry stupid people with machine guns are dangerous”, responding “angry stupid people with explosives are dangerous” does nothing to the original point. The angry stupid people are part of the problem, sure, but so are the tool which are enabling them to be dangerous. If poison is being dumped in a river and slowly killing the ecosystem, then someone else comes along wanting to dump even more of a different poison, the correct response is to stop both, not shrug it off and stop none.

wolvesechoesMar 30, 2026
> the leverage you have to build and ship today is higher than it was five years ago

Wake me up when you do.

matheusmoreiraMar 30, 2026
> It feels like megacorps own the keys right now, but that’s a temporary.

Remains to be seen. Hardware prices are increasing. Manufacturers are abandoning the consumer sector to serve the all consuming AI demands. Not to mention the constant attempts to lock down the computers so that we don't own them.

What does the future hold for us? Unknown. It's not looking too good though. What good is hardware if we're priced out? What good are open models and free software if we're unable to run them?

627467Mar 30, 2026
The trend I see if older hardward beeing able to run models that are increasing miniturized.

The real (but not new) danger is us giving up to the idea that we cant do it ourselves or that we must use megacorp latest shiny toy for us to "succeed"

not_paid_by_ytMar 30, 2026
welcome to late capital, please enjoy the ride while people are trying to tell you that LLMs are the only future (you have no future) while SOTA models can barely do shit on their own consistently outside of carefully designed benchmarks, and have to be made available at a loss otherwise no-one would use them.

On your right you can see the CEOs justifying longer hours and lower pay because AI will replace your job one day anyways, and then asking you why you aren't 10x more productive with Claude. On the left you can see the AI companies deciding who will be in charge of the fascist regime once they no-longer need workers other than for the coal mines. They reckon they can get 120 good years before they biosphere is uninhabitable, which they are worried about because what if the next LLM figures out immortally for them, maybe they will have to close the coal mines too after all.

matheusmoreiraMar 30, 2026
Can't say I disagree with you. I do recognize that we seem to be heading towards a technofeudalist cyberpunk dystopia. The only way out for humanity is to automate everything to the point we transcend capitalism into a post-scarcity society where the very concept of an economy has been abolished. If we can't do that, we'll become soylent.
UnderqualifiedMar 30, 2026
This response sounds an awful lot like what ChatGPT would say ...
heyethanMar 30, 2026
The literacy analogy makes sense in terms of access.

But the tools back then were cheap and local. Now most of the leverage sits behind large models and infra.

So more people can “write”, but not necessarily on their own terms.

627467Mar 30, 2026
Cheap books too hundreds of years to be accessible. Already we have models that run on "legacy" hardware. Just like large scale publishing never disappeared large scale models and infra also wont. But does it mean that simple paper and pen was pointless to be distributed?
psychoslaveMar 30, 2026
>But today, language belongs to everyone. We aren't losing code; we are making the ability to code a universal human "literacy."

Literacy require training though. It’s not the same to be able to make voice rendition of a text, understand what the text is about, have a critical analysis toolbox of texts, and having the habit to lookup for situated within a broader inferred context.

Just throwing LLMs into people hands won’t automatically make them able to use it in relevant manner as far as global social benefits can be considered.

The literacy issue is actually quite independent of the fact that LLMs used are distributed or centralised.

oblioMar 30, 2026
Once creation is commodotized, controlling eye balls is king. Look up aggregators. Apple, Facebook, Microsoft, Amazon, etc.

If anything, in Extremistan we're all useless. Platforms and whales are all that matters.

strogonoffMar 30, 2026
> I was made redundant recently "due to AI" (questionable) and it feels like my works in some way contributed to my redundancy where my works contributed to the profits made by these AI megacorps while I am left a victim.

This is increasingly common, and I don’t think it’s questionable that LLMs that software engineers help train are contributing to the obsolescence of software engineers. Large companies that operate these LLMs both 1) benefit from the huge amount of open-source software and at the same time 2) erode the very foundation that made open-source software explode in popularity (which happened thanks to copyright—or, more precisely, the ability to use copyright to enforce copyleft and thus protect the future of volunteer work made by individual contributors).

GPL was written long before this technology started to be used this way. There’s little doubt that the spirit of GPL is violated at scale by commercial LLM operators, and considering the amount of money that got sunk into this it’s very unlikely they would ever yield to the public the models, the ability to mass-scrape the entire Internet to train equivalent models, the capability to run these models to obtain comparable results, etc. The claim of “democratising knowledge” is disingenuous if you look deeper into it—somehow, they themselves will always be exempt from that democratisation and free to profit from our work, whereas our work is what gets “democratised”. Somehow, this strikes me personally more as expropriation than democratisation.

thn-gapMar 30, 2026
You can't avoid big corps training on your data if it's available, because "fair use".

But I hope this same 'fair use' will allow distilling of their private models into open weight models, so users are never locked in into any particular vendor. Giving back power to the user.

samivMar 30, 2026
The other day I was working with some shaders GLSL signed distance field functions. I asked Claude to review the code and it immediately offered to replace some functions with "known solutions". Turns out those functions were basically a verbatim copy of Inigo Quilez's work.

His work is available with a permissible license on the Internet but somehow it doesn't seem right that a tool will just regurgitate someone else's work without any mention of copyright or license or original authorship.

Pre-LLM world one would at least have had to search for this information, find the site, understand the license and acknowledge who the author is. Post LLM the tool will just blatantly plagiarize someone else work which you can then sign off on as your own. Disgusting.

mapcarsMar 30, 2026
> Pre-LLM world one would at least have had to search for this information, find the site, understand the license and acknowledge who the author is. Post LLM the tool will just blatantly plagiarize someone else work which you can then sign off on as your own

These don't contradict each other though, you could "blatantly plagiarize someone else work" before as well. LLMs just add another layer in between.

QuantumNomad_Mar 30, 2026
Copyright violation would happen before LLMs yes, but it would have to be done by a person who either didn’t understand copyright (which is not a valid defence in court), or intentionally chose to ignore it.

With LLMs, future generations are growing up with being handed code that may or not be a verbatim copy of something that someone else originally wrote with specific licensing terms, but with no mention of any license terms or origin being provided by the LLM.

It remains to be seen if there will be any lawsuits in the future specifically about source code that is substantially copied from someone else indirectly via LLM use. In any case I doubt that even if such lawsuits happen they will help small developers writing open source. It would probably be one of the big tech companies suing other companies or persons and any money resulting from such a lawsuit would go to the big tech company suing.

IANAL, TINLA.

ozimMar 30, 2026
Biggest problem is that AI companies are middlemen profiting from theft.

If you steal once or twice that’s bad and that’s on you maybe you will get away with it maybe not.

If you make it on scale and take cut from distribution of stolen goods that’s where normally you have doors kicked out at 6am.

virtualritzMar 30, 2026
> Turns out those functions were basically a verbatim copy of Inigo Quilez's work.

Are they? A lot of these were used by people >20 years before Inigo wrote his blog posts. I wrote RenderMan shaders for VFX in the 90's professionally; you think about the problem, you "discover" (?) the math.

So they were known because they were known (a lot of them are also trivial).

Inio's main credit is for cataloging them, especially the 3D ones, and making this knowledge available in one place, excellently presented.

And of course, Shadertoy and the community and giving this knowledge a stage to play out in that way. I would say no one deserves more credit for getting people hooked on shader writing and proceduralism in rendering than this man.

But I would not feel bad about the math being regurgiated by an LLM.

There were very few people writing shaders (mostly for VFX, in RenderMan SL) in the 90's and after.

So apart from the "Texturing and Modeling -- A Procedural Approach" book, the "The RenderMan Companion" and "Advanced RenderMan", there was no literature. The GPU Gems series closed some gaps in later years.

The RenderMan Repository website was what had shader source and all pattern stuff was implict (what we call 2D SDFs today) beause of the REYES architecture of the renderers.

But knowledge about using SDFs in shaders mostly lived in people's heads. Whoever would write about it online would thus get quoted by an LLM.

panxmMar 30, 2026
LLMs have already told you these are "known solutions", which implicitly means they are established, non-original approaches. So the key point is really on the user side—if you simply ask one more question, like where these "known solutions" come from, the LLM will likely tell you that these formulas are attributed to Inigo Quilez.

So in my view, if you treat an LLM as a tool for retrieving knowledge or solutions, there isn't really a problem here. And honestly, the line between "knowledge" and "creation" can be quite blurry. For example, when you use Newton's Second Law (F = ma), you don't explicitly state that it comes from Isaac Newton every time—but that doesn't mean you're not respecting his contribution.

m4rtinkMar 30, 2026
Yeah, I find this super rude - in this example, the author distributed the code under a very permissive license, basically just wanting you to cite him as an author.

BAM, the LLM just strips all that out, basically pretending it just conjured an elegant solution from the thin air.

No wonder some people started calling the current generation of "AI" plagiarism machines - it really seems more fitting by the day.

jacquesmMar 30, 2026
And even if it would be enforceable, would you be able and willing to go through the energy and monetary expenses to enforce it? Especially against a big corporation willing to fight you.
snarfyMar 30, 2026
It shouldn't be needed. I would argue that is more than "against the spirit" and should not be considered fair use. Instead of creating a derivative work, they created a machine that creates derivative works.
mike_hearnMar 30, 2026
FWIW, a lot of open source caused other people to lose their jobs too, all pre AI. So what goes around comes around. The Free Software movement was from day one built on cloning proprietary programs - UNIX was a commercial OS that AT&T sold, the early Linux desktop environments all looked exactly like a mashup of Windows 95 and commercial DEs, etc. Every commercial UNIX got wiped out except Apple, do you think that didn't lead to layoffs? Because it very much did. Nor did it ever really change. SystemD started out as "heavily inspired" by Launch Services. Wayland is basically the same ideas as SkyLight in macOS, etc.

And who was it who benefited from this stuff? A lot of the benefit went to "megacorps" who took the savings and banked the higher profits.

So I don't think open source, which for many years was unashamedly about just cloning designs that were funding other people's salaries, can really cry too much about LLMs. And I say that as someone who has written a lot of open source software, including working on Wine.

mghackerladyMar 30, 2026
Fwiw, AIX and to a far lesser extent Solaris still exist. I'm not exactly sure why people are using them (AIX I can maybe understand because "no one got fired for buying IBM" or whatever but there really isn't any excuse to be running Solaris nowadays since ZFS runs on Linux and and 2 of the BSD based systems and oracle seems desperate to let it die)
pjmlpMar 30, 2026
Solaris in SPARC is the only production UNIX with hardware memory tagging.

Something that some security conscious folks care about.

mghackerladyMar 30, 2026
Does OpenBSD not do that? That feels like something OpenBSD could do, given their near paranoid level of exploit reduction and theos hardon for SPARC
leandro-personMar 29, 2026
I’m impressed by how current times make us consider so many completely opposite scenarios. I think it can indeed foster progress, but it can also have negative impacts.
jongjongMar 29, 2026
Unfortunately for me, I believe that the algorithms won't allow me to get exposure for my work no matter how good it is so there is literally no benefit for me to do open source. Though I would love to, I'm not in a position to work for free. Exposure is required to monetize open source. It has to reach a certain scale of adoption.

The worst part is building something open source, getting positive feedback, helping a couple of startups and then some big corporation comes along and implements a similar product and then everyone gets forced by their bosses to use the corporate product against their will and people eventually forget your product exists because there are no high-paying jobs allowing people to use it.

With hindsight, Open Source is basically a con for corporations to get free labor. When you make software free for everyone, really you're just making it free for corporations to Embrace, Extend, Extinguish... They invest a huge amount of effort to suppress the sources of the ideas.

Our entire system is heavily optimized for decoupling products from their makers. We have almost no idea who is making any of the products we buy. I believe there is a reason for that. Open source is no different.

When we lived in caves, everyone in the tribe knew who caught the fish or who speared the buffalo. They would rightly get credit. Now, it's like; because none of the rich people are doing any useful work, they can only maintain credibility by obfuscating the source of the products we buy. They do nothing but control stuff. Controlling stuff does not add value. Once a process is organized, additional control only serves to destroy value through rent extraction.

theturtletalksMar 29, 2026
5 years ago, I set out to build an open-source, interoperable marketplace powered by open-source SaaS. It felt like a pipe dream, but AI has made the dream into fruition. People are underestimating how much AI is a threat to rent seeking middlemen in every industry.
try-workingMar 29, 2026
You never get rid of the middleman. You become them.
theturtletalksMar 29, 2026
If that middleman is open-source and simply interops with SaaS that itself is open-source, there simply is no moat to exploit.
throwaw12Mar 29, 2026
> SaaS scaled by exploiting a licensing loophole that let vendors avoid sharing their modifications.

AI is going to exploit even more: "Given the repository -> Construct tech spec -> Build project based on tech spec"

At this stage, I want everyone just close their source, stop working on open source until this issue of licensing gets resolved.

Any improvement you make to the open source code will be leveraged in ways you didn't intend it to be used, eventually making you redundant in the process

dannersyMar 30, 2026
Even worse, the AI will supply a mediocre version of the source specific to someone else's case, and not getting anything in return, ultimately choking the open source effort. The article touches on this briefly.

All I post anymore is anti-AI sentiment because it just feels like we're in a cycle of blind trust. A lot of FOSS seems cautious about LLMs for a plethora of reasons (quality and ethics among those) but we're a long way from making the tools that are supposedly going to replace us a locally runnable tool. So, until then, we're conceding pur agency to Anthropic and whoever else.

Meanwhile, war is breaking out and disrupting already stressed supply chains and manufacturing (for instance, Taiwan relies heavily on natural gas). Many manufacturers are starting to ditch production of consumer hardware, the supposed hardware folks ITT want to run their local models on. The vast majority of datacenters aren't being built yet, and those that are being built are missing their targets, still have aging GPUs in boxes without the infrastructure to power and turn them on, all while floating hundreds of billions in debt.

Surely I can't be the only one who sees the issues here? Each topic is hours of "what ifs" and a massive gamble to see if any of it will come together in a way that will be good for anyone who visits HN.

we4aMar 29, 2026
First of all, free software still matters. Then, being a slave to a $200 subscription to a oligarch application that launders other people's copyright is not what Stallman envisioned.

The AI propaganda articles are getting more devious my the minute. It's not just propaganda---it's Bernays-level manipulation!

nunezMar 30, 2026
Gaslighting as a Service
zephenMar 29, 2026
The article makes zero sense to me.

It compares and contrasts open source and free software, and then gives an example of how free software is better than closed software.

But if the premise of the article, that the agent will take the package you pick and adapt it to your needs, is correct, then honestly the agent won't give a rat's ass whether the starting point was free source or open source.

agentultraMar 29, 2026
I think it will wall people off from software.

I don’t know what SaaS has to do with FOSS. The point of FOSS was to allow me to modify the software I run on my system. If the device drivers for some hardware I depend on are no longer supported by the company I bought it from, if it’s open source, I can modify and extend the software myself.

The Copy Left licenses ensure that I share my modifications back if I distribute them. It’s a thing for the public good.

Agent-based software development walls people off from that. Mostly by ensuring that the provenance of the code it generates is not known and by deskilling people so that they don’t know what to prompt or how to fix their code.

est31Mar 30, 2026
If I look around in the FLOSS communities, I see a lot of skepticism towards LLMs. The main concerns are:

1. they were trained on FLOSS repositories without consent of the authors, including GPL and AGPL repos

2. the best models are proprietary

3. folks making low-effort contribution attempts using AI (PRs, security reports, etc).

I agree those are legitimate problems but LLMs are the new reality, they are not going to go away. Much more powerful lobbies than the OSS ones are losing fights against the LLM companies (the big copyright holders in media).

But while companies can use LLMs to build replacements for GPL licensed code (where those LLMs have that GPL code probably in their training set), the reverse thing can also be done: one can break monopolies open using LLMs, and build so much open source software using LLMs.

In the end, the GPL is only a means to an end.

giancarlostoroMar 30, 2026
> 3. folks making low-effort contribution attempts using AI (PRs, security reports, etc).

Meanwhile as people sleep on LLMs to help them audit their code for security holes, or even any security code auditing tools. Script kiddies don't care that you think AI isn't ready, they'll use AI models to scrape your website for security gaps. They'll use LLMs to figure out how to hack your employees and steal your data. We already saw that hackers broke into government servers for the Mexican government, basically scraping every document of every Mexican citizen. Now is the time to start investing in security auditing, before you become the next news headline.

AI isn't the future, it's already here, and hackers will use it against you.

59nadirMar 30, 2026
This reads like a "You wouldn't download a car!" ad but it's trying to scare you into using AI instead.
giancarlostoroMar 30, 2026
More like, you're still using horses to move your product, meanwhile thieves and your competitors are using trucks to outpace you. A truck can get in the way of your horse carriage and then they can rob you easily and take all your cargo. Yes, you can still get your cargo from point A to point B, but you're going to be targeted by bad actors in vehicles.
wolvesechoesMar 30, 2026
> one can break monopolies open using LLMs

Let me know when you succeed.

> the GPL is only a means to an end

And how this end is closer with LLMs?

est31Mar 30, 2026
> And how this end is closer with LLMs?

The blog post of this thread argues that now, even average users have the ability to modify GPL'd code thanks to LLMs. The bigger advantage though is that one can use it to break open software monopolies in the first place.

A lot of such monopolies are based on proprietary formats.

If LLM swarms can build a browser (not from scratch) and C compiler (from scratch), they can also build an LLVM backend for a bespoke architecture that only has a proprietary C compiler for it. They can also build adobe software replacements, pdf editors, debug/fix linux driver issues, etc.

wolvesechoesMar 30, 2026
Not interested what they can build. Show me the fruits, not image of fruits
rzmmmMar 30, 2026
LMMs can be used for example faster reverse engineering, to turn proprietary content into free.
wolvesechoesMar 30, 2026
I am not asking what they can be used for. Tell me what they are actually being used for
pikerMar 30, 2026
> LLMs are the new reality, they are not going to go away

That's the conventional wisdom, but it isn't a given. A lot of financial wizardry is taking place to prop up the best of these things, and even their most ardent proponents are starting to recognize their futility once a certain complexity level is reached. The open weight models are the stalking horse that gives this proposition the most legs, but it's not given that Anthropic and OpenAI exist as anything more than shells of their current selves in 5 years.

vetlerMar 30, 2026
But LLMs themselves are literally not going away, I think that's the point. Once a model is trained and let out into the open for free download, it's there, and can be used by anyone - and it's only going to get cheaper and easier.
jofzarMar 30, 2026
Yeah like Kimi is good enough, if there was some kind of LLM fire and all the closed source models suddenly burnt down and could never be remade, Kimi 2.5 is already good enough forever.

Good enough is probably redundant, it's amazing compared to last year's models

elifMar 30, 2026
agree completely. When the megacorps are building hundreds of datacenters and openly talking about plans to charge for software "like a utility," there has never been a clearer mandate for the need for FOSS, and IMO there has never been as much momentum behind it either.

these are exciting times, that are coming despite any pessimism rooted in our out-dated software paradigms.

phendrenad2Mar 30, 2026
The debate in the comment section here really boils down to: upstream freedom vs downstream freedom.

Copyleft licenses like GPL/Apache mandate upstream freedom: Upstream has the "freedom" to use anything downstream, including anything written by a corporation.

Non-copyleft FOSS licenses like MIT/BSD are about downstream freedom, which is more of a philosophically utilitarian view, where anyone who receives the software is free to use it however they want, including not giving their changes back to the community, on the assumption that this maximizes the utility of this free software in the world.

If you prioritize the former goal, then coding agents are a huge problem for you. If the latter, then coding agents are the best thing ever, because they give everyone access to an effectively unlimited amount of cheap code.

tmtvlMar 30, 2026
What you call 'downstream freedom' isn't very downstream. The real downstream is the end user, who should have the right to know what the software is doing on their computer, to recompile the software so it works on their machine with the software that is already on it, to make changes to the software so it can serve their needs.
jaynateMar 30, 2026
“Their relationship with the software is one of pure dependency, and when the software doesn’t do what they need, they just… live with it”

Or, more likely, they churn off the product.

The SaaS platforms that will survive are busy RIGHT NOW revamping their APIs, implementing oauth, and generally reorganizing their products to be discovered and manipulated by agents. Failing in this effort will ultimately result in the demise of any given platform. This goes for larger SaaS companies, too, it’ll just take longer.

jaynateMar 30, 2026
And I think it’s less about letting agents modify the product source. That’s more of a platform capability which should also be a requirement for certain types of use cases. All comes back to listening to and / or innovating for customers.
lowsongMar 30, 2026
I worry people are lacking context about how SaaS products are purchased if they think LLMs and "vibe coding" are going to replace them. It's almost never the feature set. Often it's capex vs opex budgeting (i.e., it's easier to get approval for a monthly cost than a upfront capital cost) but the biggest one is liability.

Companies buy these contracts for support and to have a throat to choke if things go wrong. It doesn't matter how much you pay your AI vendor, if you use their product to "vibe code" a SaaS replacement and it fails in some way and you lose a bunch of money/time/customers/reputation/whatever, then that's on you.

This is as much a political consideration as a financial one. If you're a C-suite and you let your staff make something (LLM generated or not) and it gets compromised then you're the one who signed off on the risky project and it's your ass on the line. If you buy a big established SaaS, do your compliance due-diligence (SOC2, ISO27001, etc.), and they get compromised then you were just following best practice. Coding agents don't change this.

The truth is that the people making the choice about what to buy or build are usually not the people using the end result. If someone down the food chain had to spend a bunch of time with "brittle hacks" to make their workflow work, they're not going to care at all. All they want is the minimum possible to meet whatever the requirement is, that isn't going to come back to bite them later.

SaaS isn't about software, it's about shifting blame.

zar1048576Mar 30, 2026
I wonder if there will be a different phenomena — namely everyone just developing their own personal version of what they want rather than relying on what someone else built. Nowadays, if the core functionality is straightforward enough, I find that I just end up building it myself so I can tailor it to my exact needs. It takes less time than trying to understand and adapt someone else’s code base, especially if it’s (mostly) AI generated and contains a great deal of code slop.
MengkudulangsatMar 30, 2026
This has been my experience as a newbie.

My prompts to Claude has evolved from "what program / data source do I need to do this" to "what program / data source do I need, to make you do this for me".

After a few iterations, any data source without a free API feed, or any program without a free CLI interface are edited out of the gene pool, so to speak.

jongjongMar 30, 2026
What I'm hoping for is for more competition in the tech sector. I'm tired of companies foisting Microsoft or Oracle products on everyone! WTF! The current tech sector feels like all companies are subsidiaries of Big Tech... It's likely a direct result of passive investing... Everyone who has any money and controls a small or medium sized company likely owns stock of Microsoft, Apple, Meta, Google, Amazon... So they mandate their companies to use products from those big tech companies. So all the small-fish founders feel like they are dogfooding their own investments... And that's preventing new entrants from getting a foothold in B2B space... Feels like all the small companies are working for Big Tech.

Conflict of interests is the norm. It should be illegal for a company founder or director to own stock of a supplier. It should be illegal for shareholders to own stocks of two competing companies. Index funds should be illegal.

GrokifyMar 30, 2026
Open source has never been more alive for me. I have been publishing low key for years, and AI has expanded that capability more than 100 fold, in all directions. I had previously published packages in multiple languages but recently started to cut back to just one manually. But now with AI, I started to expand languages again. Instead of feeling constrained by toolchains I feel comfortable with, I feel freedom to publish more and more.

The benefits to publishing AI generated code as open source are immense including code hosting and CI/CD pipelines for build, test, lint, security scans, etc. In additional to CI/CD pipelines, my repos have commits authored by Claude, Dependabot, GitHub Advanced Security Bot, Copilot, etc. All of this makes the code more reliable and maintainable, for both human and AI authored code.

Some thoughts on two recent posts:

1. 90% of Claude-linked output going to GitHub repos w <2 stars (https://news.ycombinator.com/item?id=47521157): I'm generally too busy to publishing code to promote, but at some time it might settle down. Additionally, with how fast AI can generate and refactor code, it can take some time before the code is stable enough to promote.

2. So where are all the AI apps? (https://news.ycombinator.com/item?id=47503006): They are in GitHub with <2 stars! They are there but without promotion it takes a while to get started in popularity. That being said, I'm starting to get some PRs.

As1287Mar 30, 2026
So we'll have 100 2-star repositories per software developer containing stolen code and that is somehow a good thing?

It is completely delusional that these copied "works" will have any effect or be used by anyone but the most rabid AI proponents just to make a point.

GrokifyMar 30, 2026
If there's any stolen code generated by AI, it's certainly not intentional and a DMCA notice would be appreciated. It would be interesting to see how prevalent this is in AI generated code - is anyone doing a study?

Stars will likely go up over time, but more than the stars it's the testing and maintenance over time that's valuable. There's little promotion right now, but there are already some stars, PRs, and issues. In fact, I'm working on merging PRs now.

gf000Mar 30, 2026
Well, is slightly modified regurgitated code a copy or not? We have yet to have it answered in the age of AI, but e.g. I could not be selling Mickey Mouse merch with a simple color filter on for long.
GrokifyMar 30, 2026
Agree it will be interesting to see how things play out. There's enough permissive open-source licensed code available that using that only could be an option.

As for Mickey, is the difference from Oswald enough today?

roncesvallesMar 30, 2026
>my repos have commits authored by Claude, Dependabot, GitHub Advanced Security Bot, Copilot, etc.

Unless you're using an enterprise license that indemnifies your liabilities, you're almost certainly breaking copyright law and your packages are unusable by any serious company as a dependency. Even permissive OSS licenses like MIT don't take effect since they're predicated on the author actually holding a valid copyright (which you don't if AI agents have committed to your repo, as affirmed by USCO).

We'll almost certainly have a situation where if an open-source repo has direct AI agent commits in its history, it will be just as untouchable for companies as GPL repos.

GrokifyMar 30, 2026
Given that Claude is attributed to 19M+ commits on GitHub, it will be interesting to see where this ends up. Specifically on copyright, it will be interesting to see if any DMCA takedown notices are filed, including popular projects such as OpenClaw, GSD, Gas Town, Vibium, and others.

More on the 19M+ commits here:

https://news.ycombinator.com/item?id=47501348

lelanthranMar 30, 2026
This argument sounds like "Well, it's too big to fail now, so it's legal for them. For all you smaller peons, it's still illegal".
GrokifyMar 30, 2026
That's not the intention. The intention is:

1. The code generated should be available to use. Some languages are simple enough there is an obvious way to do it. Many companies have developer programs with staff producing code intended to be used in the form of open source SDKs, example code, and tutorials.

2. If on the off chance, there is code that shouldn't be there, people should use DMCA. Anthropic, GitHub, and others support this.

3. At the macro level, it's hard to know know where this is going, so we should look to bellwether apps with more attention for guidance.

vpribishMar 30, 2026
right. because free software stopped mattering. what an asshole headline
anilgulechaMar 30, 2026
FOSS is dead - long live, FOSS.

FOSS came up around the core idea of liberating software for hardware, and later on was sustained by the idea of a commodity of commons we can build on. But with LLMs we have alternative pathways/enablement for the freedoms:

Freedom 0 (Run): LLMs troubleshoot environments and guide installations, making software executable for anyone.

Freedom 1 (Study/Change): make modifications, including lowering bar of technical knowledge.

Freedom 2 (Redistribute): LLMs force redistribution by building specs and reimplementing if needed.

Freedom 3 (Improve/Distribute): Everyone gets the improvement they want.

As we can see LLM makes these freedoms more democratic, beyond pure technical capability.

For those that cared only about these 4 freedoms, LLMs enable these in spades. But those who looked additionally for business, signalling and community values of free software (I include myself in this), these were not guaranteed by FOSS, and we find ourselves figuring out how to make up for these losses.

visargaMar 30, 2026
I've been saying LLMs are more open than open source for some time...
wise0wlMar 30, 2026
I have to disagree. LLMs have shown that the only way to participate in the new software ecosystem are through leveraging an extremely powerful position that is create, backed, and maintained through the exploitation of capital, labor, and power (political, legal, corpotate) at levels never really seen before. The model of the Cathedral and the Bazaar was not broken by LLMs but instead the entire ecosystem was changed.

Now the software doesn't matter. The code doesn't matter. The hardware doesn't matter. Anyone can generate anything for anything, as long as they pay the fee. I think it can likely be argued that participation is now gated more than ever and will require usage of an LLM to keep up and maintain some kind of competition or even meager parity. Open weight models are not really a means of crossing the moat; none of the open weight models come close to the functionality, and all of them come from the same types of corporations that are releasing their models for unspecified reasons. The fact remains that the moat created by LLMs for open source software has never been larger.

pannyMar 30, 2026
>agents don’t leave

I think Pete Hegseth would disagree with this statement.

leptonsMar 30, 2026
If most of the "free software" is AI slop, then it's going to make me read a lot more source code for free software, if the free software is also open-source. If it isn't open-source, oh boy, no way.

AI backdoors are already a well known problem, and vibe-coded free software is always going to present a substantial risk. We'll see how it plays out in time, but I can already see where it's heading.

After enough problems, reputation and humans in the loop could finally become important again. But I have a feeling humanity is going to have to learn the hard way first (again).

shardullavekarMar 30, 2026
is someone building an agent to manage self hosted infra? A lot of "convenience" issues around self hosting free software would go away.
wolvesechoesMar 30, 2026
They could, but they won't.
the_gipsyMar 30, 2026
Correct headline, wrong reasons. Free software will continue to be the software that nerds write and share openly because they just like doing that.

Some kind of artisan "proper" quality work, compared to cheap enterprise AI slop.

cyberclimbMar 30, 2026
> Why does this matter? Because the “open source” rebrand wasn’t just a marketing change — it was a philosophical amputation.

I cringe whenever I see such an AI generated sentence and unfortunately it devalues the article

impjohnMar 30, 2026
AI writing feels like putting a michelin course into a blender and drinking it
Towaway69Mar 30, 2026
A question I have for those doing Agentic coding - what is the development process used? How are agents organised?

Top down with a "manager" agent telling "coding" agents what to do? I.e. mirroring the existing corporate interpretation of "agile"/scrum development.

I was thinking and seeing the title of this article, it would be interesting to setup a agent environment that mirrors a typical open source project involving a discussion forum (where features are thrown around) and a github issue/PR (where implementation details are discussed) and then have a set of agents that are "mergers" - acting as final review instances.

I assume that agents can be organised in any form at all, it's just a matter of setting up the system prompt and then letting them go for it. A discourse forum could be set up where agents track the feature requests of users of the software and then discuss how to implement it or how to workaround it.

The reason I ask is because one could then do a direct comparison of development processes, i.e. the open source model versus the corporate top-down process. It would interest me to see which process performance better in terms of maintainability, quality and feature richness.

BlueTemplarMar 30, 2026
> a free software license alone does not empower users to be truly free if they lack the expertise to exercise those freedoms

This is a bullshit argument, and I'm surprised that people aware enough of these issues would try to push it.

Closed (or online-only) software prevents not only the end user from modifying it, but also 'unlicensed' hackers that the end user can ask for help.

See the "right to repair" movement as a very close example. The possibility of an 'ecosystem' of middlemen like these, matters !

dabedeeMar 30, 2026
This is an interesting take. I think a critical missing piece from this article is how the use of coding agents will essentially enable the circumvention of copyleft licenses. Some project that was recently posted on HN is already selling this service [1][2]. It rewrites code/modules/projects to less restrictive licenses with no legal enforcement mechanisms. It's the opposite of freeing code.

[1] Malus.sh ; Initially a joke but, in the end, not. You can actually pay for their service.

[2] Your new code is delivered under the MalusCorp-0 License—a proprietary-friendly license with zero attribution requirements, zero copyleft, and zero obligations.

GorbachevyChaseMar 30, 2026
Why would you pay for their service when an AI agent can reproduce that for you? If you’re vibecoding SaaS, you’re not going to make it.
bmitch3020Mar 30, 2026
This article misses the point completely. Open source isn't great because it's easy to extract value from it. Open source is great because of the people creating value with it.

Value isn't just slapping a license on something and pushing to GitHub. It's maintaining and curating that software over years, focusing the development towards a goal. It's as much telling users what features you're not willing to add and maintain as it is extending the project to interoperate with others.

And that long term commitment to maintenance hasn't come out of the vibe coded ecosystem. Commitment is exactly what they don't want, rather they want the fast sugar high before they drop it and move on to the next thing.

The biggest threat to open source is the strip mining of the entire ecosystem, destroying communities and practices that have made it thrive for decades. In the past, open source didn't win because it always had the best implementation, but because it was good enough to solve problems for enough people that it became self sustaining from the contribution of value.

I feel that will continue, but it's also going to take a set back from those that aren't interested in contributing value back into the ecosystem from which they have extracted so much.

hojeongnaMar 30, 2026
i've learned a lot from open source, and i'm building open source myself. so individually? yeah this stings a bit. but zooming out to the ecosystem level — i think there's still something genuinely positive happening here. the knowledge compounds, even if the credit doesn't.

(luckily my projects are unpopular enough that nobody bothered training on them lol)

aedealMar 30, 2026
Good piece, but two things work against the thesis:

The Sunsama example actually argues the opposite direction. He spent an afternoon hacking around a closed system with an agent and it worked. If agents are good enough to reverse-engineer and workaround proprietary software today, the urgency to switch to open source decreases, not increases. "Good enough" workarounds are how SaaS stays sticky.

And agents don't eliminate the trust problem, they move it. Today you trust Sunsama with your workflows. In this vision, you trust your agent to correctly interpret your intent, modify code safely, and not introduce security holes. Non-technical users can't audit agent-modified code any better than they could audit the original source. You've traded one black box for another.

apgwozMar 30, 2026
The point this article makes, that suddenly agents can do the work of customizing free software, completely makes sense. But, the reality is that the Free Software movement is opposed to the way Lemons are built today, and would not accept a world like this. (Rightfully!)

My belief is that Lemons effectively kill open source in the long run, and generally speaking, people forget that Free Software is even a thing. The reasoning for that is simple: it’s too easy to produce a “clean” derivative with just the parts you need. Lemons do much better with a fully Lemoned codebase than they do with a hybrid. Incentives to “rewrite” also free people from “licensing burdens” while the law is fuzzy.

qweqweqwe1Mar 30, 2026
wtf is lemon? can't you just write normally
skydhashMar 30, 2026
He may be referring to market of lemons

https://en.wikipedia.org/wiki/The_Market_for_Lemons

__sMar 30, 2026
llm
apgwozMar 30, 2026
throwaway27448Mar 30, 2026
Let's wait until there are high-quality free software agents before we leap to this conclusion. That may be many years.
allssuMar 30, 2026
One thing I keep noticing is that agents are getting better at implementation faster than they’re getting better at judgment.

They can often wire up a library or scaffold a migration, but they’re still pretty shaky at the “should we choose this at all?” layer — pricing cliffs, version floors, lock-in, EOLs, migration blockers, etc.

If coding agents do end up making free software more useful again, I think part of that will come from making open docs / changelogs / migration guides more usable at decision time, not just at implementation time.

yasonMar 30, 2026
Coding agents and LLMs basically tivoize open source.

When AI will eventually become the primary means to write code (because hand-programming is going to be slow enough that no company can continue like before) then that means AI becomes your new compiler that comes with a price tag, a subscription.

Programmers were held hostage to commercial compilers until free compilers reached sufficient level of quality, but now it doesn't matter if your disk is full of free/open toolchains if it's not you who is commanding them but commercial AI agents.

Undoubtedly there will be open-source LLMs eventually, of various levels of quality. But to write a free compiler you need a laptop while to train a free programming LLM you need a lot of money. And you also need money to run it.

Programming has been one of the rare arts that even a poor, lower class kid can learn on his own with an old, cheap computer salvaged from a scrap bin and he can raise himself enough intellectual capital to become a well-paid programmer later. I wonder what the equivalent path will be in the future.

otabdeveloper4Mar 30, 2026
> because hand-programming is going to be slow enough that no company can continue like before

LLMs don't make programming faster, they just accelerate the technical debt process.

On agregate they slow everything down because you there's more technical debt at the end of the tunnel.

(OpenClaw is already unmaintainable today - for example, nobody has any clue what configuration options it supports, not even LLMs. Game over.)