• PupBiru@kbin.social
    link
    fedilink
    arrow-up
    2
    arrow-down
    1
    ·
    1 year ago

    we output nothing other than what we’re trained on; the only difference is that we’re allowed to roam the world freely and consume whatever information we stumble on

    what you say would be true if the LLM were only trained on content by the author seeking to say that their works had been infringed, however these LLMs include a lot of other data from public domain sources

    one could consider these public domain sources and our experience of the world to be synonymous (and if you don’t i’d love to hear the distinction), in which case there’s some kind of a line that you seem to be drawing, and again i’d love to hear where you think that line is

    is it just ratio? there’s precedent to that for sure: current law has fair use rules which stipulate things like “amount and substantiality”. in that case the question becomes one of defining the ratio. certainly the ratio of content that the author is referring to vs the content not trained by the author is minuscule

    • walrusintraining@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      I agree with what you’re saying, and a model that is only trained on public domain would be fine. I think the very obvious line is that it’s a computer program. There seems to be a want for computers to be human but they aren’t. They don’t consume media for their own enjoyment, they are forced to do it so someone can sell the output as a product. You can’t compare the public domain to life.

      • PupBiru@kbin.social
        link
        fedilink
        arrow-up
        2
        ·
        1 year ago

        i think the distinction that either side is seeing here is that you think humans are inherently different to a neural network, where i think that the only difference is in the complexity: that if we had a neural network at the same scale as the human brain, that there’s nothing stopping those electronic neurons from connecting and responding in a way that’s indistinguishable from a human

        the fact that we’re not there yet i don’t see as particularly relevant, because we’re talking about concepts rather than specifics… of course a LLM doesn’t display the same characteristics as a human: it’s not of the same scale, and the training is different but functionally there’s nothing different between chemical neurons firing and neurons made of transistors firing

        we learn in the same way: by reinforcing connections between our neurons

        • walrusintraining@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 year ago

          A few points:

          • Humans are more than just a brain. There’s the entire experience of ego, individualism, and body

          • Another massive distinction is autonomy and liberty, which no AI models currently possess.

          • We don’t know all there is to know about the human brain. We can’t say it is functionally equivalent to a neural network.

          • If complexity is irrelevant, then the simplest neural network trained on a single work of writing is equivalent to the most advanced models for the purposes of this discussion. Such a network would, again, output a copy of the work it was trained on

          When we’ve developed a true self-aware AI that can move and think freely, the idea that there is little difference will have more weight to it.