A rising movement of artists and authors are suing tech companies for training AI on their work without credit or payment

  • Lil' Bobby Tables@programming.dev
    link
    fedilink
    English
    arrow-up
    0
    arrow-down
    1
    ·
    1 year ago

    Sorry if this comes off as offensive, but this isn’t new news, and I think we all need a reality check here. I’ll also be forward, the artifice of doubt cast on this makes me pretty angry and is a threat to every one of us.

    So, OK. Speaking as a thirty-year programmer, a neurobiology minor, and a hobby animator, you know what these objections sound like? They sound like somebody who just learned to program, with JavaScript, at the top of that initial peak. You know the one, when you feel like you can do anything and present as such. You’ve never had to worry about stack overflows or memory exceptions or, quite possibly, even fundamental networking utilities, but you got a web page to do something which is, in fact, legitimately cool, and you’re proud of yourself—as you friggin’ should be. However, you don’t know how much you don’t know, and are way further from the top than you suspect you are.

    Until you learn something like C, and really walk away from it with a sore ass and a sense of perspective, you don’t know what you don’t know. This isn’t because we all aren’t rooting for you, it’s just an expected rite of passage; when you really begin to learn. We’re totally rooting for you, we used to be you! And this is also true of visual art and programming; they have nothing to do with each other, and as far as I can see this case is open and shut, and we’re still mulling over the details.

    Artists don’t copy, they analyze. It’s difference between reading all of the answers on a Stack Exchange site, considering them, discussing them, putting them in the context of your own life, and applying them in an organized and personal fashion; versus simply copying the code verbatim and jamming things in until it arguably works, which is literally what an LLM neural network is designed to do. We’ve all seen this with extremely questionable code output by GPT-3 (and, yes, GPT-4, it still happens, just not as frequently).

    The art neural networks are the same—an artist charges for that inspiration, because it was a lot of physical and emotional toil for them. It was a lot of feedback and self-critique. “Prompt engineering” is neither art nor, if we’re honest with ourselves, engineering.

    I maintain that this is basically a slightly obfuscated recap of the Napster trials from twenty years ago, who, ironically, fell back on almost the same “fair use” defense. It’s going to falter just as hard, as there’s a massive difference between showing smaller, low-resolution images on a search page; and using meaningful elements of those images to produce brand new, and competing, works; which happen to be flawed, but look the same to someone who browses JPEGs on Google like a channel surfer.

    The last thing I’m going to say is that we need to stop describing LLMs as “AI”. “AI” doesn’t mean anything. It could refer to a neural network, machine learning, A* path finding, collective intelligence, and any number of other things, the colloquial definition being technology that “performs a function which was previously only possible for a human”, and come on, once upon a time you could describe a hammer that way. AI is not a formal industry term, it’s marketing flak. LLMs are effectively a database of connections browsed with the use of a neural network.

    You want to keep using Midjourney and Stable Diffusion? Great! Go for it. You want to use ChatGPT to help you understand some multivariate calculus? Go nuts, I do it all the time, most mathematicians are terrible at articulating. However, they should, without question, have to pay for the art that they used, or cease using it if the sale won’t be completed. Any other outcome is absolutely going to lead to an economic collapse.