Thousands of authors demand payment from AI companies for use of copyrighted works::Thousands of published authors are requesting payment from tech companies for the use of their copyrighted works in training artificial intelligence tools, marking the latest intellectual property critique to target AI development.

  • linearchaos@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 year ago

    I don’t know how I feel about this honestly. AI took a look at the book and added the statistics of all of its words into its giant statistic database. It doesn’t have a copy of the book. It’s not capable of rewriting the book word for word.

    This is basically what humans do. A person reads 10 books on a subject, studies become somewhat of a subject matter expert and writes their own book.

    Artists use reference art all the time. As long as they don’t get too close to the original reference nobody calls any flags.

    These people are scared for their viability in their user space and they should be, but I don’t think trying to put this genie back in the bottle or extra charging people for reading their stuff for reference is going to make much difference.

    • BartsBigBugBag@lemmy.tf
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      It’s not at all like what humans do. It has no understanding of any concepts whatsoever, it learns nothing. It doesn’t know that it doesn’t know anything even. It’s literally incapable of basic reasoning. It’s essentially taken words and converted them to numbers, and then it examines which string is likely to follow each previous string. When people are writing, they aren’t looking at a huge database of information and determining the most likely word to come next, they’re synthesizing concepts together to create new ones, or building a narrative based on their notes. They understand concepts, they understand definitions. An AI doesn’t, it doesn’t have any conceptual framework, it doesn’t even know what a word is, much less the definition of any of them.

      • Buttons@programming.dev
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 year ago

        I think you underestimate the reasoning power of these AIs. They can write code, they can teach math, they can even learn math.

        I’ve been using GPT4 as a math tutor while learning linear algebra, and I also use a text book. The text book told me that (to write it out) “the column space of matrix A is equal to the column space of matrix A times its own transpose”. So I asked GPT4 if that was true and it said no, GPT disagreed with the text book. This was apparently something that GPT did not memorize and it was not just regurgitating sentences. I told GPT I saw it in a text book, the AI said “sorry, the textbook must be wrong”. I then explained the mathematical proof to the AI, and the AI apologized, admitted it had been wrong, and agreed with the proof. Only after hearing the proof did the AI agree with the text book. This is some pretty advanced reasoning.

        I performed that experiment a few times and it played out mostly the same. I experimented with giving the AI a flawed proof (I purposely made mistakes in the mathematical proofs), and the AI would call out my mistakes and would not be convinced by faulty proofs.

        A standard that judged this AI to have “no understanding of any concepts whatsoever”, would also conclude the same thing if applied to most humans.

        • unlimitedolm_sjw@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          1 year ago

          That doesn’t prove that GPT is reasoning, its model predicts that those responses are the most likely given the messages your sending it. It’'s read thousands of actual conversations with people stating something incorrect, then having it explained to them and them coming around and admitting they were wrong.

          I’ve seen other similar cases where the AI is wrong about something, and when it’s explained, it just doubles down. Because humans do that type of thing too, refusing to admit their wrong.

          The way it’s designed means that it cannot reason in the same way humans experience it. It can simulate a likely conversation someone would have if they could reason.

      • oce 🐆@jlai.lu
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        1 year ago

        How can you tell that our thoughts don’t come from a biological LLM? Maybe what we conceive as “understanding” is just a feeling emerging from a more fondamental mechanism like temperature emerges from the movement of particles.

        • Telodzrum@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          Because we have biological, developmental, and psychological science telling us that’s not how higher-level thinking works. Human brains have the ability to function on a sort of autopilot similar to “AI”, but that is not what we are describing when we speak of creative substance.