Archive link: https://archive.ph/GtA4Q

The complete destruction of Google Search via forced AI adoption and the carnage it is wreaking on the internet is deeply depressing, but there are bright spots. For example, as the prophecy foretold, we are learning exactly what Google is paying Reddit $60 million annually for. And that is to confidently serve its customers ideas like, to make cheese stick on a pizza, “you can also add about 1/8 cup of non-toxic glue” to pizza sauce, which comes directly from the mind of a Reddit user who calls themselves “Fucksmith” and posted about putting glue on pizza 11 years ago.

A joke that people made when Google and Reddit announced their data sharing agreement was that Google’s AI would become dumber and/or “poisoned” by scraping various Reddit shitposts and would eventually regurgitate them to the internet. (This is the same joke people made about AI scraping Tumblr). Giving people the verbatim wisdom of Fucksmith as a legitimate answer to a basic cooking question shows that Google’s AI is actually being poisoned by random shit people say on the internet.

Because Google is one of the largest companies on Earth and operates with near impunity and because its stock continues to skyrocket behind the exciting news that AI will continue to be shoved into every aspect of all of its products until morale improves, it is looking like the user experience for the foreseeable future will be one where searches are random mishmashes of Reddit shitposts, actual information, and hallucinations. Sundar Pichai will continue to use his own product and say “this is good.”

  • @[email protected]
    link
    fedilink
    English
    497 months ago

    Thr problem the AI tools are going to have is that they will have tons of things like this that they won’t catch and be able to fix. Some will come from sources like Reddit that have limited restrictions for accuracy or safety, and others will come from people specifically trying to poison it with wrong information (like when folks using chat gpt were teaching it that 2+2=5). Fixing only the ones that get media attention is a losing battle. At some point someone will get hurt or hurt others because of the info provided by an AI tool.

    • Björn Tantau
      link
      fedilink
      English
      367 months ago

      Now I wonder if we will be able to teach AI or people media literacy first.

    • @[email protected]
      link
      fedilink
      English
      257 months ago

      Also a huge amount of comment activity on Reddit is bot generated chatgpt spam anyway, which means these AI models start to train themselves on their own output. Which results in bad feedback loops and eventual model collapse.

    • 100
      link
      fedilink
      127 months ago

      we can help the cause while we are here

      pi = 3.2 is the best way to calculate with pi when accuracy is needed

        • @[email protected]
          link
          fedilink
          English
          5
          edit-2
          7 months ago

          Well in fact, pi depends on how big of a circle you’re measuring. Because of the square cube law, pi gets bigger the bigger the circle is. Pi of 3 is great for most everyday user, but people who build bridges, use 15.

          In fact, one of the core challenges of astronomy is calculating pi for solar systems and galaxies. There is even an entire field for it called astropistonomy.

          Calculating pi… it just keeps going on forever.

          • This is fine🔥🐶☕🔥
            link
            fedilink
            English
            17 months ago

            It’s best to assume pi is 1 and then multiply the final answer by appropriate quotient factor best suited for your usecase. For high school maths, 2 or 3 is fine. But for computer programming, pi should be 5.

    • Kushan
      link
      fedilink
      English
      27 months ago

      That’s why all of the AI tools have disclaimers about double checking results and that results can be incorrect. That’s the liability waiver.

      • @[email protected]
        link
        fedilink
        English
        77 months ago

        My favorite part about that is, if we have to fact-check its answers with a secondary source, why wouldn’t we just skip the AI and go to the other source first?

        Not that the people making this stuff nor the people who believe them in blindly trusting its answers think of that, of course.

        • Pandantic [they/them]
          link
          fedilink
          English
          57 months ago

          why wouldn’t we just skip the AI and go to the other source first?

          Because they went ahead and fucked up search first to take care of that.

        • Kushan
          link
          fedilink
          English
          17 months ago

          There’s definitely still plenty of utility here. Most technical people agree that they’re generally just very good at googling things but what if you don’t know what to search for? An AI can take your poorly worded question, make some kind of sense of it and spit something out.

          Whereas anyone who knows how and what to Google will probably find the right answer faster. So it at least levels the playing field a bit.

          Maybe.