Hacker Newsnew | past | comments | ask | show | jobs | submit | philipov's commentslogin

GPL-2.0-only is the name of a license. One word. It is an alternative to GPL-2.0-or-later.

Right, the final hyphen changes the meaning of the sentence.

"GPL-2.0-only" "GPL-2.0 only"


You take responsibility. That means if the AI messes up, you get punished. No pushing blame onto the stupid computer. If you're not comfortable with that, don't use the AI.

There’s no reasonable way for you to use AI generated code and guarantee it doesn’t infringe.

The whole use it but if it behaves as expected, it’s your fault is a ridiculous stance.


If you think it's an unacceptable risk to use a tool you can't trust when your own head is on the line, you're right, and you shouldn't use it. You don't have to guarantee anything. You just have to accept punishment.

That’s just it though it’s not just your head. The liability could very likely also fall on the Linux foundation.

You can’t say “you can do this thing that we know will cause problems that you have no way to mitigate, but if it does we’re not liable”. The infringement was a foreseeable consequence of the policy.


This policy effectively punts on the question of what tools were used to create the contribution, and states that regardless of how the code was made, only humans may be considered authors.

From the foundation's point of view, humans are just as capable of submitting infringing code as AI is. If your argument is sound, then how can Linux accept contributors at all?

EDIT: To answer my own question:

    Instead of a signed legal contract, a DCO is an affirmation that a certain person confirms that it is (s)he who holds legal liability for the act of sending of the code, that makes it easier to shift liability to the sender of the code in the case of any legal litigation, which serves as a deterrent of sending any code that can cause legal issues.
This is how the Foundation protects itself, and the policy is that a contribution must have a human as the person who will accept the liability if the foundation comes under fire. The effectiveness of this policy (or not) doesn't depend on how the code was created.

Anyone distributing copyrighted material can be liable that DCO isn’t going to stop anyone.

If that worked any corporation that wanted to use code they legally couldn’t could just use a fork from someone who assumed responsibility and worst case they’d have to stop using it if someone found out.


> liability could very likely also fall on the Linux foundation.

It’s just the same as if I copy-paste proprietary code into the kernel and lie about it being GPL.

Is the Linux foundation liable there?


Maybe. DCOs haven’t been tested. But you can at least say that the person who did this committed fraud and that you had no reasonable way to know they would do that.

LLMs can and do regurgitate code without the user’s knowledge. That’s the problem, the user has no way to mitigate against it. You’re telling contributors “use this thing that has a random chance of creating infringing code”. You should have foreseen that would result in infringing code making its way into the kernel.


If someone sent you some code and said “it’s all good bro, you can put it in the kernel with your name on it”, would you?

If you don’t feel comfortable about where some code has come from, don’t sign your name.

The fact LLMs exist and can generate code doesn’t change how you would behave and sign your name to guarantee something.


Are you being purposely obtuse?

Not at all.

Linus and the rules have always been very clear. If you don’t know where code came from, don’t submit it.


That’s like a speed limit sign that says “whatever speed you think is reasonable” but in small print “as long as that doesn’t exceed 45mph”.

Yes it’s technically correct, but it won’t hold up I court and it’s a ridiculous statement.

What Linus’ statement is actually saying is that: we want to benefit from AI tooling, but we don’t want to accept any liability.


The only lawsuits so far have been over training on open source software. You're inventing a liability problem that essentially does not exist.

OpenAI and Anthropic added an indemnity clause to their enterprise contracts specifically to cover this scenario because companies wouldn’t adopt otherwise.

Yeah, but that's not a useful thing to do because not everybody thinks about that or considers it a problem. If somebody's careless and contributes copyrighted code, that's a problem for linux too, not only the author.

For comparison, you wouldn't say, "you're free to use a pair of dice to decide what material to build the bridge out of, as long as you take responsibility if it falls down", because then of course somebody would be careless enough to build a bridge that falls down.

Preventing the problem from the beginning is better than ensuring you have somebody to blame for the problem when it happens.


It was already necessary to solve the problem of humans contributing infringing code. It was solved by having contributors assume liability with a DCO. The policy being discussed today asserts that, because AI may not be held legally liable for its contributions, AI may not sign a DCO. A human signature is required. This puts the situation back to what it was with human contributors. What you are proposing goes beyond maintaining the status quo.

It’s not solved. It hasn’t been tested in court to my knowledge and in my opinion is unlikely to hold up to serious challenge. You can be held liable for just distributing copyrighted code even if the whole “the Linux foundation doesn’t own anything” holds up.

> Preventing the problem from the beginning is better than ensuring you have somebody to blame for the problem when it happens.

that's assuming that the problems and incentives are the same for everyone. Someone whose uncle happens to own a bridge repair company would absolutely be incentivized to say

> "you're free to use a pair of dice to decide what material to build the bridge out of, as long as you take responsibility if it falls down"


Their position is probably that LLM technology itself does not require training on code with incompatible licenses, and they probably also tend to avoid engaging in the philosophical debate over whether LLM-generated output is a derivative copy or an original creation (like how humans produce similar code without copying after being exposed to code). I think that even if they view it as derivative, they're being pragmatic - they don't want to block LLM use across the board, since in principle you can train on properly licensed, GPL-compatible data.

>There’s no reasonable way for you to use AI generated code and guarantee it doesn’t infringe.

I guess we’ll need to reevaluate what copy rights mean when derivatives grow on trees?


If they merge it in despite it having the model version in the commit, then they're arguably taking a position on it too - that it's fine to use code from an AI that was trained like that.

> That means if the AI messes up

I'm not talking about maintainability or reliability. I'm talking about legal culpability.


Unless one wishes to control the entire ecosystem rather than simply participate in it.

You are falling into the trap of thinking there's a single monolithic being called Software Developers that has inconsistent opinions. In fact, you're observing different people with conflicting values.

Yeah yeah. But LLMs certainly have been embraced by a large number of developers. Many of whom I've observed react with disgust when they see "not X, but Y" or emdashes in an article. But when it comes to code, "wow this is so awesome!"

I always understood it as advice for school children who don't know how to express themselves in writing and end up with a blank page. You are encouraged to imagine you're talking to someone to overcome the hurdle of translating thoughts into words.

Along similar lines, in elementary school we were taught the introduction/body/conclusion pattern to help organize our thoughts in essays, but actually using those words as headers makes you sound like a child. "Conclusion" is especially common. It just reads like you're blindly following that pattern and don't actually know how to tie your thoughts together.

I remember editing my then girlfriend’s (now wife) papers in university, when she was struggling to explain something succinctly. I’d ask her: well what are you trying to say? And she would explain it to me, clear as day. “Just write that.”

I learned the opposite lesson and ended up talking like I write.

This is largely where I'm ending up, but I started at the other end.

There is value in prose that carries your literal voice when the audience is _people who know you_. There is negative value in writing prose that requires the audience to _read it in your voice_ in order for it to make sense, avoid offense, or convey intent.

My prose changed first: it became plain spoken, as devoid of contextual subtlety as I could make it. My career benefitted. My spoken interactions followed.

The only thing that bothers me about it is the nagging sense that I've become so fucking boring.


I also started at the other end, in the sense that from an early age and for the rest of my life I have spent a lot more time chatting over text than with spoken word.

Perhaps there is beauty in 'negative value'. Art may have 'negative value' by that definition.

I would not argue with this, but I regret to confess that I am merely an engineer. I no longer aspire to artistry.

That's how you get flagged as an AI!

Yes, perl is considered write-only because it is a mess of features that allow unhygienic programming habits to flourish - it is full of hard-to-trace magical behavior. Completely different than APL, which has had perl's write-only label applied to it by programmers not used to reading terse mathematical notation.

Not to be confused with b programming language, which is not its succesor, but is the predecessor to c.

The more people use chargebacks to get around black hole customer service the better, because it is difficult for companies to blacklist everyone. If they don't want to pay the mediation fee, they should provide customer service in the first place.

There's a misunderstanding here. I'll make it clearer.

The "Unless you're big cheese" is the company you're doing the charge back against.

If a company, such as Anthropic has too many chargebacks? Visa/MC can ban them from their network. It happens to smaller companies all the time, mostly because it costs Visa/MC + the banks involved to deal with each chargeback, and also, it's typically a sign of fraudulent behaviour.

Visa/MC are not a charity, or are payment processors. They need profit. Take it away by creating all this extra work, chargeback work, and they're not making money any more.

The "big cheese" part is, if you're amazon or google, things can be negotiated at that scale. Maybe they pay a larger settlement fee, whatever. And of course Google Play, or Amazon utterly dwarfs Anthropic CC activity at this point, even though they have a large valuation and potential future ahead.

Still, I have no idea what the background metrics and profit points are for Visa/MC, only that I've seen even medium sized companies have issues with too many chargebacks. And, we've all seen Visa/MC decide they don't like gambling, or porn sites and just drop them. Some of those companies were quite large and had a lot of flow for them.

So I don't think many companies will just use chargebacks as a support mechanism. That is, unless they're just completely incompetent.


at this point Anthropic/OpenAI have enough equity to buy out the entire Visa and Mastercard ecosystems.

It's all fake venture capital money, but they are the big cheese.


Having equity doesn't mean they can buy it, and regardless, that doesn't mean Visa/MC will work for free, or the banks/payment processors. Too many charge backs from an account, and that's the result.

It's unclear how large their retail business is, which is why I mentioned that, and that's where you see most CC payments. Companies with any serious usage are going to pay via wire or bill payment via banks directly. McDonalds, for example, likely has a larger daily spend on cards.


I have that same book. It was a major source of inspiration when I was a kid.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: