Bill Gates feels Generative AI has plateaued, says GPT-5 will not be any better::The billionaire philanthropist in an interview with German newspaper Handelsblatt, shared his thoughts on Artificial general intelligence, climate change, and the scope of AI in the future.

    • @[email protected]
      link
      fedilink
      English
      21 year ago

      I’d say the majority of humans know what 2 + 2 is. Chat GPT doesn’t. As it found the answer in some texts it will tell you 4, but all it takes is you telling it that’s wrong and suddenly it’s 5. So even for the most simple math problem it’s extremely easy to throw the whole thing off. Which also means for any prompt you put in it can go in wildly wrong directions at times.

      And this is all with good input data, there’s plenty of trolls online and the data will only get worse (it already did, the original data up to 2021 was okayish, in the last year tons of crap was put out on top, some of it by Chat GPT itself. So the new model might input the crap it produced before, getting worse over time). The problem on top of that is that you don’t know the sources it used. If you ask about a recent event you might receive an insane answer it picked up from a right wing conspiracy site, you simply don’t know. There is no fact checking in place.

      It’s a stunningly good text generator, but that’s all it is and it ever will be, at least until they do much more than just add more compute power to it.

        • @[email protected]
          link
          fedilink
          English
          11 year ago

          GPT can probably give a better answer to any advanced math or science query than the majority of humans

          Only if that answer is already out there and in the model. So pretty much a Google search away.

          GPT isn’t coming up with new math or science facts (at least not real ones).

          It literally is a word predictor, an insanely complex one, it’s the best way to describe it. If you start with layers, parameters and so on most people lose interest. But there’s some really good explanations around.

          Generic AI (real AI) has internal logic, can learn and improve itself and can do self motivated actions. Chat GPT can tell you exactly how to create an account and order something from Amazon, but despite being able to put that text out it will never be able to actually follow them itself.

          • @[email protected]
            link
            fedilink
            English
            21 year ago

            Only if that answer is already out there and in the model.

            That’s not true. I wanted a vba script for Excel. I don’t know vba or excel so I spent hours searching Google for help. There were explanations of functions but no working code. I tried GPT for the fun of it and it spit back working code. Code that was nowhere on the Internet.

            It was able to put together functions into working code based on the definition of functions, not simply cutting and pasting what somebody else had already written.

          • @[email protected]
            link
            fedilink
            English
            11 year ago

            Chat GPT can tell you exactly how to create an account and order something from Amazon, but despite being able to put that text out it will never be able to actually follow them itself.

            This is a really good ELI5 explanation of its limit.

            • @[email protected]
              link
              fedilink
              English
              11 year ago

              That would still give it too much credit in that case. It’s purely an input output system. You put text in (the prompt), you get text out (the result). If there is no input from you, there is no output. It doesn’t have any intrinsic functionality that runs on its own.

              Maybe a bit too much for an ELI5.