• @[email protected]
      link
      fedilink
      English
      81 month ago

      They are copying. These LLM are a product of their input, and solely a product of their input. It’s why they’ll often directly output their training data. Using more data to train reduces this effect, that’s why all these companies are stealing and getting aggressive in stopping others stealing their data.

    • @[email protected]
      link
      fedilink
      English
      41 month ago

      Proof? I am fairly certain I am correct but I will gladly admit fault. This whole LLM thing is indeed new to me also