• Hangglide@lemmy.world
    link
    fedilink
    English
    arrow-up
    21
    arrow-down
    8
    ·
    1 year ago

    Bullshit. If I learn engineering from a textbook, or a website, and then go on to design a cool new widget that makes millions, the copyright holder of the textbook or website should get zero dollars from me.

    It should be no different for an AI.

      • Marxine@lemmy.ml
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        3
        ·
        1 year ago

        While I agree, corporations shouldn’t make bucks on knowledge(sorta) they basically eavesdropped and violated the privacy of millions of people for.

        AI solutions are made from people’s ideas, and should be freely accessible by the people by definition. It not being sustainable as a business model is also a feature in this case, since there’d be no intrinsic incentive to steal data and violate privacy.

    • Treczoks@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      3
      ·
      1 year ago

      Yes, but what about you going into teaching engineering, and writing a text book for it that is awfully close to the ones you have used? Current AI is at a stage where it just “remixes” content it gobbled in, and not (yet) advanced enough to actually learn and derive from it.

    • Mouselemming@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      3
      ·
      1 year ago

      Last time I looked, textbooks were fucking expensive. You might be able to borrow one from the library, of course. But most people who study something pay up front for the information they’re studying on

    • Shazbot@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      7
      ·
      1 year ago

      Every time I see this argument it reminds me of how little people understand how copyright works.

      • When you buy that book the monetary amount is fair compensation for the contents inside. What you do afterwards is your own business so long as it does not violate the terms within the fine print of the book (no unauthorized reproductions, etc.)
      • When someone is contracted for an ad campaign there will be usage rights in the contract detailing the time frame and scope for fair compensation (the creative fee + expenses). If the campaign does well, they can negotiate residuals (if not already included) because the scope now exceeds the initial offer of fair compensation.
      • When you watch a movie on TV, the copyright holder(s) of that movie are given fair compensation for the number of times played. From the copyright holders, every artist is paid a royalty. Jackie Chan and Chris Tucker still get royalty checks whenever Rush Hour 2 airs or is streamed, as do all the other obscure actors and contributing artists.
      • Deviant Art and ArtStation provide free hosting for artists in exchange for a license that lets them distribute images to visitors. The artists have agreed to fair compensation in the form of free hosting and potential promotion should their work start trending, reaching all front page visitors of the site. Similarly, when the artists use the printing services of these sites they provide a license to reproduce and ship their works, as fair compensation the sites receive a portion of the artists’ asking price.

      The crux is fair compensation. The rights holder has to agree to the usage, with clear terms and conditions for their creative works, in exchange for a monetary sum (single or reoccurring) and/or a service of similar or equal value with a designated party. That’s why AI continues to be in hot water. Just because you can suck up the data does not mean the data is public domain. Nor does it mean the license used between interested parties transfers to an AI company during collection. If AI companies want to monetize their services, they’re going to have to provide fair compensation for the non-public domain works used.

    • VonCesaw@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      10
      ·
      1 year ago

      Human experience considers context, experience, and relation to previous works

      ‘AI’ has the words verbatim in it’s database and will occasionally spit them out verbatim

      • 📛Maven@lemmy.sdf.org
        link
        fedilink
        English
        arrow-up
        5
        ·
        1 year ago

        It doesn’t. The original data is nowhere in its dataset. Words are nowhere in its dataset. It stores how often certain tokens (numbers computationally equivalent to language fragments; not even words, but just a few letters or punctuation, often chunks of words) are found together in sentences written by humans, and uses that to generate human-sounding sentences. The sentences it returns are thereby a massaged average of what it predicts a human would say in that situation.

        If you say “It was the best of times,” and it returns “it was the worst of times.”, it’s not because “it was the best of times, it was the worst of times.” is literally in its dataset, it’s because after converting what you said to tokens, its dataset shows that the latter almost always follows the former. From the AI’s perspective, it’s like you said the token string (03)(153)(3181)(359)(939)(3)(10)(108), and it found that the most common response to that by far is (03)(153)(3181)(359)(61013)(12)(10)(108).

        • Hello Hotel@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          2
          ·
          edit-2
          1 year ago

          Impressioning and memorization, it memorised the impression (“sensation”) of what it’s like to have the text in the buffer: “It was the best of times,” and “instinctively” outputs it’s impression “it was the worst of times.” Knowing each letter it added was the most “correct” rewarding.