• dustycups
    link
    fedilink
    English
    649 months ago

    I assume AI is training off the content here for free.

      • deweydecibel
        link
        fedilink
        English
        27
        edit-2
        9 months ago

        And ya know what? Frankly, if AI is going to harvest all this shit, I’d rather fuckers like spez couldn’t get rich off it in the process. Granted I’m not happy the tech bros running these AI companies are getting rich with these fucking things, but I can at least take solace that, for Lemmy at least, there isn’t some asshole middle man making bank off the work and words of users they never paid a dime to.

        Genuinely, why does Sepz and Reddit deserve to make money off anything we posted? Why does any social media site? They make the site, pay for the servers, maintain the apps, sure, and they can get compensation for that, I don’t see a problem there. But why does any social media company deserve to get rich when the only thing that makes their platform valuable is the people that post to it? Reddit didn’t even have paid mods, the community did all the work on the content of that site, why in the fuck do we tolerate these assholes making profit off it like this?

        • @[email protected]
          link
          fedilink
          English
          29 months ago

          This is sad to read because I agree with all of it (except the casual sexism).

          why in the fuck do we tolerate these assholes making profit off it like this?

          Look at this thread. People delete their posts on Reddit. Which means that they can no longer be scraped for free. Which means they are now exclusively available in Reddit’s archive. It’s not that people tolerate it. It’s that the first instinct of people who don’t tolerate it, is to make it worse. What can you do?

    • @[email protected]
      link
      fedilink
      English
      169 months ago

      I was curious if a robots.txt equivalent exists for AI training data, and there was some solid points here:

      If I go to your writing, I read it & learn from it. Your writing influences my future writing. We’ve been okay with this as long as it’s not a blatant forgery.

      If a computer goes to your writing, it reads it & learns from it. Your writing influences its future writing. It seems we are not okay with this, even if it isn’t blatant forgery.

      [AI at the moment is] different because the company is re-using your material to create a product they are going to sell. I’m not sure if I believe that is so different than a human employee doing the same thing.

      https://news.ycombinator.com/item?id=34324208

      I still think we should have the ability to opt out like we do with search engines and webcrawlers, but if the algorithm works ideally and learns but does not recycle content, is it truly any different from a factory of workers pumping out clones of popular series on Amazon? I honestly don’t know the answer to that.

      • deweydecibel
        link
        fedilink
        English
        10
        edit-2
        9 months ago

        The problem is not the technology, the problem is the businesses and the people behind them.

        These tools were made with the explicit purpose of taking the content that they did not create, repurposing them, and creating a product. Throw all these conversation about intelligence and learning out the fucking window, what matters is what the thing does, and why it was created to do that thing.

        Until we reach a point where there is some sort of AI out there that has any semblance of free will, and can choose not to learn if fed certain information, and choose not to respond to input given to it without being programmed to do not respond, then we are not talking about intelligence, we are talking about a tool. No matter how they dress it up.

        Stop arguing about this on their terms, because they’re gaslighting the fuck out of you.

      • @[email protected]
        link
        fedilink
        English
        69 months ago

        Afaik the OpenAI bot may choose to ignore it? At least that’s what another user claimed it does.

        • JohnEdwa
          link
          fedilink
          English
          129 months ago

          Robots.txt has been always ignored by some bots, it’s just a guideline originally meant to prevent excessive bandwidth usage by search indexing bots and is entirely voluntary.

          Archive.org bot for example has completely ignored it since 2017.

      • Mossy Feathers (She/They)
        link
        fedilink
        English
        39 months ago

        This is kinda my take on it. However, the way I see it is that the AI isn’t intelligent enough yet to truly create something original. As such, right now AI is closer to being a tool than a being. Because of that, it somewhat bothers me that I’m being used to teach a tool. If I thought that companies like OpenAI were truly trying to create beings and not tools, then I’d feel differently.

        It’s kinda nuanced, but a being can voluntarily determine whether or not something is copyright infringing, understand why that might be an issue, and then decide whether or not to continue writing based on that. A tool can’t really do that. You can try and add filters to a tool to avoid writing copy written text, but that will have flaws and holes in it. A being who understands what it’s writing and what makes it plagiarism vs reference vs homage/inspiration/whatever is less likely to have those issues.