• 362 Posts
  • 28 Comments
Joined 2 years ago
cake
Cake day: May 11th, 2022

help-circle




















  • I don’t really know how I feel about this. On one hand the algorithm is presumably spitting out something unique based on other work rather than regurgitating other people’s work. On the other hand, they are making use of a huge body of work to create that new unique work. Is that acceptable? I don’t know.

    The other side of this is, can you really copyright code that has been produced by an ai? If something has been created by a mechanism, my very limited input from a human, can you really call that a creative work? In the monkey photo case, it was determined that the photograph that was taken by the monkey could not be copyrighted by the photographer because the photographer did not take the photo. If you have a mechanical monkey spitting out code for you, can you copyright the equivalent of a mechanical monkey pressing a button?

    There are several issues with the generated content from AI systems and copyright aspects. In the USA, someone already filed a law­suit with a class action on the most relevant issues related to Ai generated content concerning art representations. See https://stablediffusionlitigation.com







  • I think there are some real dangers of having non-humans involved with court proceedings.

    First there’s the obvious slippery slope of first your lawyer is an AI, then the prosecutor is an AI, then the judge is an AI and suddenly we’re living entirely off the dictates of an AI system arguing with itself.

    Second, there’s the fact that no AI is a human. This might not seem important, but there’s a lot of truth that a human can perceive that an AI can’t. The law isn’t computer code, it’s extremely squishy and that fact is important to it being just but it’s also important because you can’t just enter text into a prompt and expect to get the results out of the system you want. There’s a big difference between the same question asked by a judge who appears to be convinced by your argument and a judge who appears to be skeptical of your argument.

    You might make an argument that it’s just traffic violations, but there’s a slippery slope there as well. First it’s traffic violations, eventually you might have poor people making use of the AI for serious crimes because through degrees you go “oh, it’s just a traffic violation, oh it’s just a low level possession charge, oh it’s just for crimes with a guilty plea anyway, oh it’s just a tort claim, oh it’s just a real estate case…”

    Another thing is as AI expands, suddenly you get a potential risk with hackers. If you have a really important court case, it might be justifiable to pay someone to break into the AI and sabotage it so you win the case.

    I agree with you. The topic is complex ad would deserve much more space to be deepened. Some issues are related, for example, to biases; there are several misdefined cases due to AI biases, especially in the USA.




  • It has not escaped your notice. I usually talk about app-related issues. The choice for one or the other solution is based on trust, and personally, after several trials with different solutions, I trust Apple. I am certainly aware that Apple is one of the biggies and that it is not exempt from criticism, but the policy adopted in recent years is user-friendly. It is only worth mentioning that in 2018, during the international conference of Data Protection and Privacy Commissioners, Tim Cook wished that the U.S. had a privacy regulation like the GDPR. This is not the appropriate venue, but your comment will allow me to post something on the point you arise.