• Diplomjodler
    link
    fedilink
    English
    3476 months ago

    It’s not AI that is the problem, it’s half baked insecure data harvesting products pushed by big corporations that are the problem.

        • NutWrench
          link
          fedilink
          English
          596 months ago

          Also, CoPilot is going to be bundled with Office 365, a subscription service. You’re literally paying them to spy on you.

        • @[email protected]
          link
          fedilink
          English
          146 months ago

          I think the bigger joke is calling LLMs AI

          I have to disagree.

          Frankly, LLMs (which are based on neural networks) seem a Hell of a lot closer to how actual brains work than “classical AI” (which basically boils down to a gigantic pile of if statements) does.

          I guess I could agree that LLMs are undeserving of the term “AI”, but only in the sense that nothing we’ve made so far is deserving of it.

          • @[email protected]
            link
            fedilink
            English
            56 months ago

            seem a Hell of a lot closer

            “seem” is the critical word there. Interacting with an LLM they do seem to be pretty clever.

            • @[email protected]
              link
              fedilink
              English
              8
              edit-2
              6 months ago

              I’m not talking about interacting with it. I’m talking about how it’s implemented, from my perspective as a computer scientist.

              Let me say it more concretely: if even shitty expert systems, which are literally just flowcharts implemented in procedural code, are considered “AI” – and historically speaking, they are – then the bar is really fucking low. LLMs, which at least make an effort to kinda resemble the structure of biological intelligence, are certainly way, way above it.

              • @[email protected]
                link
                fedilink
                English
                26 months ago

                I’m actually sad that the state of AI deserves the hate it gets. Neural networks are so sick, just going through the example of detecting a diagonal on a 2x2 grid was like magic to me. And they made me second guess simulation theory for quite a while lmao

                Tangentially, blockchain was a similar phenomenon for me. Or at least trust networks. One idea was to just throw away Certificate Authorities. Basically federate all the things, and this was before we knew about the fediverse. It gets all the hate because of crypto, but it’s cool tech. The CA thing would probably lead to a bad place too, though.

          • Brickardo
            link
            fedilink
            English
            15 months ago

            Let’s agree to disagree then. An LLM has no notion of semantics, it’s just outputting the most likely word to follow up to what it’s already written and the user’s input.

            On the contrary, expert systems from back in the 90s for, say, predicting the atomic structure of an element, work like a human brain on steroids. It features an arbitrary large search tree that the software knows how to iterarively prune according to a well known set of chemical rules. We do the same when analyzing a set of options.

            Debugging “current” AI models, on the other hand, is impossible because all we’re doing is prescripting a composition of functions and forcing it to minimize a loss function. That’s all we’re doing. How can you currently tell that a certain model is going to work? Unless the mathematical theory ever catches up with the technology, we’ll never know until we execute the code.

      • Aniki 🌱🌿
        link
        fedilink
        English
        18
        edit-2
        6 months ago

        Runs locally, mirrors remotely.

        To ensure a seamless customer experience when their hardware isn’t capable of running the model locally or if there is a problem with the local instance.

        microsoft, probably.

    • @[email protected]
      link
      fedilink
      English
      836 months ago

      That is an accurate description of AI in common usage even if it isn’t an inherent aspect of AI.

      • andrew
        link
        fedilink
        English
        126 months ago

        Right, but AI is not the only way they’re doing the data collection.

    • Pennomi
      link
      fedilink
      English
      216 months ago

      Locally run AI could be great. But sending all your data to an external server for processing is really, really bad.