• @[email protected]OPM
    link
    fedilink
    13 days ago

    That’s still not into the realm where I trust it; the underlying model is a language model. What you’re describing is a recipe for ending up with paltering a significant fraction of the time.

    • @[email protected]
      link
      fedilink
      English
      13 days ago

      Did you even try diffy.chat to test how factually correct it is and how well it cites its sources? How good does it have to be to be useful? How bad does it have to be to be useless?

      • @[email protected]OPM
        link
        fedilink
        13 days ago

        I tried it. It produces reasonably accurate results a meaningful fraction of the time. The problem is that when it’s wrong, it still uses authoritative language, and you can’t tell the difference without underlying knowledge.

        • @[email protected]
          link
          fedilink
          English
          13 days ago

          There does need to be a mechanism to keep the human in the loop to correct the knowledge base by people who have the underlying knowledge. Perhaps notification needs to be sent to people who have previously viewed the incorrect information when a correction is made.