Authors using a new tool to search a list of 183,000 books used to train AI are furious to find their works on the list.

  • adriaan@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    10
    arrow-down
    9
    ·
    9 months ago

    That would be a much better comparison if it was artificial intelligence, but these are just reinforcement learning models. They do not get inspired.

    • Shurimal@kbin.social
      link
      fedilink
      arrow-up
      9
      arrow-down
      4
      ·
      9 months ago

      just reinforcement learning models

      …like the naturally occuring neural networks are.

      • Khalic@kbin.social
        link
        fedilink
        arrow-up
        6
        arrow-down
        4
        ·
        9 months ago

        The brain does not work the way you think… (I work in the field, bio-informatics). What you call “neural networks” come from an early misunderstanding of how the brain stores information. It’s a LOT more complicated and frankly, barely understood.

        • FaceDeer@kbin.social
          link
          fedilink
          arrow-up
          5
          arrow-down
          5
          ·
          9 months ago

          It’s a LOT more complicated and frankly, barely understood.

          Yet you confidently state that the brain doesn’t work the way LLMs do?

          Obviously it doesn’t work exactly the same way that LLMs do, if only because of the completely different substrates. But when you get to more nebulous concepts like “creativity” and “inspiration” it’s not so clear.

          • lloram239@feddit.de
            link
            fedilink
            English
            arrow-up
            5
            ·
            9 months ago

            The part where brain and neural net differ is in the learning via backpropagation, that seem to be done different in the brain, as there is no mechanism to go backwards through the network and jiggle the weights.

            That aside, they seem to work very similar once they are trained, as the knowledge they are able to extract from data ends up being basically the same that a human would be able to extract. There is surprisingly little weirdness in AI and a surprising amount of human-like capabilities.

      • lemmyvore@feddit.nl
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        7
        ·
        9 months ago

        Tell you what, you get a landmark legal decision classifying LLM as people and then we’ll talk.

        Until then it’s software being fed content in a way not permitted by its license i.e. the makers of that software committing copyright infringement.