• @[email protected]
    link
    fedilink
    English
    143 months ago

    The LLMs they train on their code will only be accessible internally. They won’t leak their own intellectual property.

    • Jack
      link
      fedilink
      English
      43 months ago

      Will that not be more experiensive than having developers?

      • androogee (they/she)
        link
        fedilink
        English
        43 months ago

        Of course not. It will be more expensive and they’ll still have to pay developers to figure out what’s wrong with their AI code.

      • @Hawk
        link
        English
        33 months ago

        Depends on the use case. Training local llms is a lot cheaper after Galore and there are ways to get useful local models with only a moderate amount of effort, see e.g. augmentoolkit.

        This may or may not be practical in many use cases.

        24 months is pretty generous but no doubt there will be significantly less demand for junior developers in the near future.

      • @[email protected]
        link
        fedilink
        English
        23 months ago

        Possibly. It’s hard to know without seeing the numbers and assessing output quality and volume.

        Also it’s not unheard of that some bigwig wastes millions of company €€ for some project they fancy. (Billions if they happen to be Elon)

    • @[email protected]
      link
      fedilink
      English
      2
      edit-2
      3 months ago

      If only we had an overarching structure that everyone in society has agreed exists for the purposes of enforcing laws and regulating things. Something that governs people living in a region… Maybe then they could be compelled to show exactly what they’re using, and what those models are being trained with.

      Oh well.