Ebooks go on sale a lot. You can often get them for less than $3.00, but it takes effort to track what you like and keep an eye out for sales. The trade off wasn't worth it for me years ago and I paid full price for what I wanted to read... lately there are very few books I pay full price for.
I don't think GitHub even set a precedent for this. My understanding is that they don't train on private repositories per se, though if you access a private repository through copilot, the data flow through copilot can be trained on, which pulls in data from the repo.
So a private repo should be safe, as long as you don't use copilot. While Atlassian wants to pull in data from private issue trackers/wikis.
Listen to yourself. Take a moment and try to unpack the mental gymnastics wrangling you just did. Ask yourself, why does the fact that you have a Copilot subscription make it okay to train on all your private repos?
GitHub does not have any of its own models. It routes to partners like OpenAI. Just because some data is from private repos, doesn’t mean all data is flowing nor does it mean it should be trained on just because it’s being inferences on, and there is a difference on the data that was used vs. all the data from that repo, and difference between just that repo vs. all private repos. And they made it all opted in as default. Draconian.
So yes, they did set a precedent and you’re here arguing why it’s okay.
It should be possible to comply with LGPL without publishing the source code of the whole application. Either by running the application and ffmpeg in different isolates (wasm processes), or by offering a way to merge (link) the wasm code of the closed-source application with a user compiled FFmpeg wasm build.
Different isolates might even be enough to satisfy GPL, similar to how you can invoke FFmpeg as a command line tool from a closed application. Though that feels like legally shaky ground.
Also if you want to keep a secret a secret forever, encrypted but saved data may be easily decrypted in the future. Most secrets though in reality are less useful in X years time.
I mean, sure, that can happen, but that obviously depends on what the test is testing, it's not like it's bad in all cases to say "now plus 1 year". In the case in question it's really just "cookie is far enough in the future so it hasn't expired", so "expire X years in the future from now" is fine.
reply