Authors using a new tool to search a list of 183,000 books used to train AI are furious to find their works on the list.

  • Gibdos@feddit.de
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    1
    ·
    1 year ago

    I certainly hope that none of these authors have ever read a book before or have been inspired by something written by another author.

    • Wander@kbin.social
      link
      fedilink
      arrow-up
      1
      arrow-down
      2
      ·
      1 year ago

      Are you saying the writers of these programs have read all these books, and were inspired by them so much they wrote millions of books? And all this software is doing is outputting the result of someone being inspired by other books?

      • Grimy@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        arrow-down
        1
        ·
        1 year ago

        Clearly not. He’s saying that other authors have done the same as the software does. The software creators implemented the same principle into their llm. You are being daft on purpose.

    • adriaan@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      4
      ·
      1 year ago

      That would be a much better comparison if it was artificial intelligence, but these are just reinforcement learning models. They do not get inspired.

      • Shurimal@kbin.social
        link
        fedilink
        arrow-up
        3
        arrow-down
        1
        ·
        1 year ago

        just reinforcement learning models

        …like the naturally occuring neural networks are.

        • Khalic@kbin.social
          link
          fedilink
          arrow-up
          1
          arrow-down
          3
          ·
          1 year ago

          The brain does not work the way you think… (I work in the field, bio-informatics). What you call “neural networks” come from an early misunderstanding of how the brain stores information. It’s a LOT more complicated and frankly, barely understood.

          • FaceDeer@kbin.social
            link
            fedilink
            arrow-up
            2
            arrow-down
            1
            ·
            1 year ago

            It’s a LOT more complicated and frankly, barely understood.

            Yet you confidently state that the brain doesn’t work the way LLMs do?

            Obviously it doesn’t work exactly the same way that LLMs do, if only because of the completely different substrates. But when you get to more nebulous concepts like “creativity” and “inspiration” it’s not so clear.

        • lemmyvore@feddit.nl
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          4
          ·
          1 year ago

          Tell you what, you get a landmark legal decision classifying LLM as people and then we’ll talk.

          Until then it’s software being fed content in a way not permitted by its license i.e. the makers of that software committing copyright infringement.

  • Pyr_Pressure@lemmy.ca
    link
    fedilink
    English
    arrow-up
    3
    ·
    1 year ago

    Curious if the AI company actually bought those books or if they just came across them by pirating.

  • LEX@lemm.ee
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    1
    ·
    edit-2
    1 year ago

    Any AI model that uses publically available information for training should be open source by law.

    This business where corporations (that includes authors, who are published by huge corporations) fight over who “owns” ideas is assinine. When it comes down to it, this is a fight about money being wrapped in an argument about “ideas.”

    AI models were developed with the collective knowledge and wisdom of society. They’re like libraries and should be public like libraries. OpenAI, Google, all those fucks should be forced to open source their models, end of story.

    • kibiz0r@midwest.social
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      I’d say they should have to follow the most-restrictive license of all of their training data, and that existing CC/FOSS licenses don’t count because they were designed for use in a pre-LLM world.

      It seems like a pretty reasonable request. But people like free stuff, and when they think about who will get screwed by this they like to imagine that they’re sticking it to the biggest publishers of mass media.

      But IRL, those publishers are giddy with the idea that instead of scouting artists and bullying them into signing over their IP, they can just summon IP on demand.

      The people who will suffer are the independents who refused to sign over their IP. They never got their payday, and now they never will either.

  • 👁️👄👁️@lemm.ee
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    1 year ago

    Here’s an idea, legally force companies like OpenAI to rely on opt-in data, rather then build their entire company on stealing massive amounts of data. That includes requiring to retrain from scratch. Sam Altman was crying for regulations for scary AI, right?

    • FaceDeer@kbin.social
      link
      fedilink
      arrow-up
      2
      arrow-down
      1
      ·
      1 year ago

      Would search engines only be allowed to show search results for sources that had opted in? They “train” their search engine on public data too, after all.

      • kibiz0r@midwest.social
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        1 year ago

        First: There are mechanisms to opt out (robots.txt and meta noindex)

        Second: There is some foreknowledge on the part of the web author. Even in the early days of the web — before you could’ve predicted the concept of search engines — in order to distribute anything you had to understand the basics of hypermedia, among which is the idea that anything can link to anything else and clients can be users or machines alike.

        Third: Even though you are correct that search engines are tokenizing text and doing statistical analysis to recombine the tokens into novel forms in order to rank against queries, those novel forms are never presented to the user. Only direct quotes. So a user never gets a false reference to the supposed content of a page (unless the page itself lies to crawler requests).

        Fourth: All of the technical points above are pretty much meaningless, because we are social creatures and our norms don’t stem from a mechanical flow chart divorced from real-world context.

        Creators are generally okay with their content being copied into search DBs, because they know it’s going to lead to users finding the true author of those words, which will advance their creative pursuits either through collaboration or monetary support.

        Creators are complaining about content being copied into LLMs, because their work will be presented out of context, often cited incorrectly, keep people away from the author of those words, and undermine the lifeblood of their creative pursuits – be it attracting new collaborators or making sales.

        Whether it technically counts as IP infringement or not under current law? Who really cares? Current IP law is a fucking scam, designed to bully creators out of their own creations and assign full control to holding companies who see culture as nothing more than a financial instrument to be optimized. We desperately need to change IP law anyway – something that I think even many strident “AI” supporters agree with – so using it as a justification for the ethics of LLMs reveals just how weak the group’s position truly is.

        LLM vendors see an opportunity for profit, if they can get away with it. They are offering consumers a utopian vision of infinite access to content while creating an IP chokepoint that they can enshittify once it blows past critical mass. It’s the same tactics the social media companies used 15 years ago, and it weighs heavy on my heart that so many Lemmy users are falling for it once again while the lesson is still so fresh.

      • 👁️👄👁️@lemm.ee
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        They aren’t reselling their information, they’re linking you to the source which then the website decides what to do with your traffic. Which they usually want your traffic, that’s the point of a public site.

        That’s like trying to say it’s bad to point to where a book store is so someone can buy from it. Whereas the LLM is stealing from that bookstore and selling it to you in a back alley.

      • just another dev@lemmy.my-box.dev
        link
        fedilink
        English
        arrow-up
        0
        arrow-down
        1
        ·
        1 year ago

        Fair use is any copying of copyrighted material done for a limited and “transformative” purpose, such as to comment upon, criticize, or parody a copyrighted work.

        I don’t see why it should.

        • FaceDeer@kbin.social
          link
          fedilink
          arrow-up
          1
          ·
          1 year ago

          The creation of the AI model is transformative. The AI’s model does not contain a literal copy of the copyrighted work.

          • just another dev@lemmy.my-box.dev
            link
            fedilink
            English
            arrow-up
            0
            arrow-down
            1
            ·
            1 year ago

            No, but the training data does contain a copy. And making a model is not criticising, commenting upon, or creating a parody of it.

            • FaceDeer@kbin.social
              link
              fedilink
              arrow-up
              1
              ·
              1 year ago

              That list is not exclusive, it’s just a list of examples of fair use.

              The training data is not distributed with the AI model.

  • leaky_shower_thought@feddit.nl
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    1
    ·
    1 year ago

    There’s an idea by Barath Raghavan about an AI dividend that companies pay each netizen a share for the data they use to train these models.

    I am into this idea if companies can’t even do a simple opt-in mechanism.

  • originalucifer@moist.catsweat.com
    link
    fedilink
    arrow-up
    1
    arrow-down
    3
    ·
    1 year ago

    do they also complain when their books are used to train wet networks in public schools? those networks are also later exploited by corporations who dont give back the writers. hmmmmmmm