by wgx on 9/7/23, 3:56 PM with 377 comments
by tremon on 9/7/23, 8:47 PM
by satvikpendem on 9/7/23, 6:19 PM
Google Books literally copied and pasted books to add to their online database and that was deemed fair use, so something much more transformative like generative AI will likely fall under much broader consideration for fair use. Google Books was, yes, non-commercial, but the courts generally have the provision that the more transformative something is, the less it needs to adhere to the guidelines laid out for determining such fair use.
by StewardMcOy on 9/7/23, 8:46 PM
Is this blog post a legally enforceable contract? Is Microsoft specifically indemnifying all users of Copilot against claims of copyright infringement that arise from use of Copilot?
The blog post says that "there are important conditions to this program", and it lists a few, but are those conditions exhaustive, or are there more that the blog post doesn't cover? For example, is it only in specific countries, or does it apply to every legal system worldwide?
What guarantees do users have that Microsoft won't discontinue this program? If Microsoft gets kicked in the teeth repeatedly by courts ruling against them, and they realize that even they can't afford to pay out every time Copilot license-launders large chunks of copyrighted code, what means to users have to keep Microsoft to its promises?
by jtchang on 9/7/23, 5:34 PM
The way AI is going I'm sure we'll see some landmark cases very soon. It is very much in Microsoft's interest to grow this market as fast as possible and be at the center of it. This removes one of the key impediments to adopting generated code for smaller orgs: "Will I get sued if this product generates code that is copyrighted?".
by fsdavcaa on 9/7/23, 5:28 PM
It hinges on what *Microsoft* decides "attempting to generate infringing materials" means. You'd like it to mean that it only excludes use when you're doing something you know would infringe copyright, like "reproduce the entire half life 2 source code." But who knows.
by jacquesm on 9/7/23, 6:50 PM
This is the key bit:
"Specifically, if a third party sues a commercial customer for copyright infringement for using Microsoft’s Copilots or the output they generate, we will defend the customer and pay the amount of any adverse judgments or settlements that result from the lawsuit, as long as the customer used the guardrails and content filters we have built into our products."
The 'we will defend' is one important part, I assume that means that you will be using their lawyers rather than your own (which they have in house and so are cheaper to use than the ones that bill you, the would be defendant by the hour).
The second part that matters is that there are conditions on how you are supposed to use the product and crucially: you will have to document that this is how you used it.
But: interesting development, clearly enterprise customers are a bit wary of accidentally engaging in copyright infringement by using the tool and that may well have slowed down adoption.
by lijok on 9/7/23, 11:01 PM
We tested copilot with those guardrails enabled and it completely lobotomizes it.
This by the way is not a change. They already had this “Microsoft will assume liability if you get sued” clause in Copilot Product Specific Terms: https://github.com/customer-terms/github-copilot-product-spe...
by whitfieldsdad on 9/7/23, 7:37 PM
Is it "stealing" to have a working understanding of the next best token, or even simply the token that shows up the most often (e.g. on GitHub)?
I'm sure that the argument could be made that all AI should be illegal as all ideas worth having have already been had, and all text worth writing has already been written, but, where would that leave us?
(e.g. your function for converting a string from uppercase to lowercase will probably look like a function that someone else on Earth has written, and the same goes for your error handling code, your state of the art technique for centering a div, etc.)
by littlestymaar on 9/7/23, 7:42 PM
by scj on 9/7/23, 10:29 PM
I don't know what case history is like for damages with open source projects, but I suspect it wouldn't be that big of a concern for Microsoft.
Otherwise stated, Microsoft's downside to this is committing their lawyers. And the upside is to improve their code generation tools.
IANAL though.
by lewhoo on 9/7/23, 8:54 PM
4.the effect of the use upon the potential market for or value of the copyrighted work (wiki)
I don't know if this particular case is good for exploring all angles of fair use, but to me this certainly is a greater hurdle for commercial generative ai.
by dataflow on 9/7/23, 11:12 PM
by indymike on 9/7/23, 9:24 PM
by tpmx on 9/7/23, 7:16 PM
by tboyd47 on 9/7/23, 6:35 PM
by bobobob420 on 9/7/23, 7:02 PM
by coding123 on 9/7/23, 9:01 PM
Microsoft just became a code copyright insurance company. The premium is paid for with individual copilot accounts for each developer. And the policy has its exceptions of course.
This is interesting.
by soultrees on 9/7/23, 10:52 PM
In any case, super annoying to have that happen so consistently these days that I just use chatgpt to fix my tailwind styling now.
by aldousd666 on 9/8/23, 1:28 AM
by alberth on 9/8/23, 1:20 AM
by elzbardico on 9/8/23, 1:52 PM
by treprinum on 9/7/23, 10:24 PM
by Havoc on 9/7/23, 10:36 PM
by PeterStuer on 9/7/23, 7:06 PM
by matt3210 on 9/8/23, 1:26 AM
by dirtyid on 9/8/23, 5:56 AM
by heavyset_go on 9/7/23, 5:25 PM
by thesuperbigfrog on 9/7/23, 7:10 PM
Now it is "Train, Task, Transform, and Transfer":
Train - Feed copyrighted works into machine learning model or similar system
Task - Machine learning model is tasked with an input prompt
Transform - Machine learning model generates hybrid output derived from copyrighted works, but usually not directly traceable to a given work in the training set
Transfer - Generated output provides the essence of the copyrighted works, but is legally untraceable to the originals
by baz00 on 9/7/23, 5:22 PM
by tyingq on 9/7/23, 5:34 PM
by jacquesm on 9/7/23, 6:58 PM
GitHub Copilot and open source laundering
https://drewdevault.com/2022/06/23/Copilot-GPL-washing.html
Previously on HN, in case you missed it:
by CameronNemo on 9/7/23, 5:25 PM
by sublinear on 9/7/23, 6:34 PM
Copilot is such a flawed product from the start. It's not even a matter of its ability to write "good" code. The concept is just dumb.
Code is necessarily consumed by people first before it's executed by a computer in a production environment. There are many ways to get a computer to do something, but the approval process by experienced humans is vastly more important than the drafting of it. Software dev is already incredibly cheap and the last place to cut costs.
There is no AI threat other than the one posed by grifters trying to convince you that there is.
by hulitu on 9/7/23, 4:34 PM
Extinguish.
by naikrovek on 9/7/23, 6:15 PM
How dare they? amirite?