I don’t consider myself very technical. I’ve never taken a computer science course and don’t know python. I’ve learned some things like Linux, the command line, docker and networking/pfSense because I value my privacy. My point is that anyone can do this, even if you aren’t technical.

I tried both LM Studio and Ollama. I prefer Ollama. Then you download models and use them to have your own private, personal GPT. I access it both on my local machine through the command line but I also installed Open WebUI in a docker container so I can access it on any device on my local network (I don’t expose services to the internet).

Having a private ai/gpt is pretty cool. You can download and test new models. And it is private. Yes, there are ethical concerns about how the model got the training. I’m not minimizing those concerns. But if you want your own AI/GPT assistant, give it a try. I set it up in a couple of hours, and as I said… I’m not even that technical.

  • @[email protected]
    link
    fedilink
    English
    34 months ago

    I also recently got into selfhosting LLM. Having an AMD card meant I had to scourge for solutions since everything expects to have CUDA suppport which means having Nvidia cards. Koboldcpp has a fork with ROCM support which works on my machine, so I’m content with that for now.

    • @[email protected]
      link
      fedilink
      English
      34 months ago

      Wasnt there a solution by AMD or someone close to them implementing a translation of CUDA for AMD hardware?

      • @[email protected]
        link
        fedilink
        English
        54 months ago

        AMD asked them to shut it down. So the guy is going to go back to the pre-AMD release and work independently from there.

        • @[email protected]
          link
          fedilink
          English
          24 months ago

          I really hate when companies do that kind of crap. I just imagine a little toddler stomping around going “No! No! Nooo!”

          • @[email protected]
            link
            fedilink
            English
            14 months ago

            NVIDIA didn’t ask to shut it down, but AMD lawyer probably weren’t that hot to what the project had become and AMD asked the creator to shut down the project l, which he did.

            But yeah, lots of work wasted caused by pencil pushers and bean counters.

    • @[email protected]
      link
      fedilink
      English
      24 months ago

      Do you have any links or guides that you found helpful? A friend wanted to try this out but basically gave up when he realized he’d need an Nvidia GPU.

      • Noctis
        link
        fedilink
        English
        34 months ago

        Look up kobold cpp yellow rose fork. It’s pretty easy to set up and run