• @LostWanderer
    link
    English
    34
    edit-2
    5 months ago

    Good, that shitty OpenAI needs to be drained dry and go into nonexistence ASAP.

        • sunzu
          link
          fedilink
          125 months ago

          Whose labour got exploited for micro-shit to have this much capital to blow?

          I mean I ain’t gonna cry about them potentially losing it but let’s be real we are still a paying for this.

          They get to take risks, if they win, it theirs. If they lose ohh well crack the whips to make up the diffefrence

          • jaxxed
            link
            fedilink
            English
            65 months ago

            They are draining their capital investment. Their employees are not paying the 5b.

            • sunzu
              link
              fedilink
              15 months ago

              I don’t think you understand how things really work lol

          • @[email protected]
            link
            fedilink
            English
            05 months ago

            If we play our cards right, we can all be bourgeoisie. Automate all alienated labor roles, let people do the non-alienated (ie meaningful) tasks themselves if they like.

            For $20/mo, anyone can have a personal assistant that takes dictation, does research, writes reports, etc.

            I know $20/mo isn’t accessible to everyone, but it’s a lot more accessible than it was before to exploit this kind of help.

            Assuming we can keep AI unconscious, nobody has to be an industrial or bureaucratic worker any longer. Unless they enjoy it.

  • @[email protected]
    link
    fedilink
    English
    25
    edit-2
    5 months ago

    Yann and co. just dropped llama 3.1. Now there’s an open source model on par with OAI and Anthropic, so who the hell is going to pay these nutjobs for access to their apis when people can get roughly the same quality for free without the risk of having to give your data to a 3rd party?

    These chuckle fucks are cooked.

    • @[email protected]
      link
      fedilink
      English
      235 months ago

      For “free” except you need thousands of dollars upfront for hardware and a full hardware/software stack you need to maintain.

      This is like saying azure is cooked because you can rack mount your own PC

      • @[email protected]
        link
        fedilink
        English
        17
        edit-2
        5 months ago

        OpenAI is losing money on every user and has no moat other than subsidies from VCs, but that’s ok because they’ll make it up in volume.

      • CubitOom
        link
        fedilink
        English
        115 months ago

        That’s mostly true. But if you have a GPU to play video games on a PC running Linux, you can easily use Ollama and run llama 3 with 7 billion parameters locally without any real overhead.

        • @[email protected]
          link
          fedilink
          English
          135 months ago

          Just an off-the-cuff prediction: I fully anticipate AI bros are gonna put their full focus on local models post-bubble, for two main reasons:

          1. Power efficiency - whilst local models are hardly power-sippers, they don’t require the planet-killing money-burning server farms that the likes of ChatGPT require (and which have helped define AI’s public image, now that I think about it). As such, they won’t need VC billions to keep them going - just some dipshit with cash to spare and a GPU to abuse (and there’s plenty of those out in the wild).

          2. Freedom/Control - Compared to ChatGPT, DALL-E, et al, which are pretty locked down in an attempt to keep users from embarrassing their parent corps or inviting public scrutiny, any local model will answer whatever dumbshit question you ask for make whatever godawful slop you want, no questions asked, no prompt injection/jailbreaking needed. For the kind of weird TESCREAL nerd which AI attracts, the benefits are somewhat obvious.

          • @[email protected]
            link
            fedilink
            English
            95 months ago

            you almost always get better efficiency at scale. If the same work is done by lots of different machines instead of one datacenter, they’d be using more energy overall. You’d be doing the same work, but not on chips specifically designed for the task. If it’s already really inefficient at scale, then you’re just sol.

          • CubitOom
            link
            fedilink
            English
            65 months ago

            I guess it depends how you define what an “ai bro” is. I would define them as the front men of startups with VC funding who like to use big buzz words and will try to milk as much money as they can.

            These types of people don’t care about power efficiency or freedom at all unless they can profit off of it.

            But if you just mean anyone that uses a model at home then yeah you might be right. But I’m not understanding all the harsh wording around someone running a model locally.

        • @[email protected]
          link
          fedilink
          English
          85 months ago

          The whole point of using these things (besides helping summon the Acausal Robot God) is for non-technical people to get immediate results without doing any of the hard stuff, such as, I don’t know, personally maintaining and optimizing an LLM server on their llinux gaming(!) rig. And that’s before you realize how slow inference gets as the context window fills up or how complicated summarizing stuff gets past a threshold of length, and so on and so forth.

        • @[email protected]
          link
          fedilink
          English
          75 months ago

          Azure/AWS/other cloud computing services that host these models are absolutely going to continue to make money hand over fist. But if the bottleneck is the infrastructure, then what’s the point of paying an entire team of engineers 650K a year each to recreate a model that’s qualitatively equivalent to an open-source model?

          • CubitOom
            link
            fedilink
            English
            55 months ago

            For me, the bottleneck is my data. I want to keep my data. And honestly I don’t know why any entity is OK with sharing their data for some small productivity improvements. But I don’t understand a lot.

          • @[email protected]
            link
            fedilink
            English
            4
            edit-2
            5 months ago

            The engineers can generally also do other things, the security will likely be better, and its fully possible API costs will exceed that sum if you need that much expertise inhouse to match your API usage.

            • @[email protected]
              link
              fedilink
              English
              75 months ago

              The engineers can generally also do other things

              What’s the job posting for that going to look like, LLM stack maintainer wanted, must also be accomplished front end developer in case things get slow?

  • luciole (he/him)
    link
    fedilink
    English
    205 months ago

    The big men educated in Finance would probably scoff at my ignorant remark, but anyways here goes: isn’t this the modus operandi of these massively swollen tech “startups”? Spotify for example has been operating at a loss for what, 15 years?

    Something like:

    1. screw people over
    2. get funding
    3. bleed money
    4. ???
    5. get more funding
    • @[email protected]
      link
      fedilink
      English
      145 months ago

      I’d expect the current interest rates make it a less viable strategy.

      Also, money is fake but I’d expect there’s a limit to just how many billions you can burn before you can’t get away with it.

    • @[email protected]
      link
      fedilink
      English
      75 months ago

      They need to be producing value even if it’s not “taxable profit” according to generally accepted accounting principles.

      For Spotify, spending venture capital is fine as long as you secure more users, ready to ne milked later.

      For OpenAI, if they burn through $5b in venture capital, their product needs to be ready to make $5b more than it was previously.

    • @[email protected]
      link
      fedilink
      English
      135 months ago

      Dear CHATGPT how do we get our company more money?

      “That’s a great question. To get more money simply bring up the console and enter ‘rosebud!;! ;! ;! ;!’”