I’m interested in hosting something like this, and I’d like to know experiences regarding this topic.

The main reason to host this for privacy reasons and also to integrate my own PKM data (markdown files, mainly).

Feel free to recommend me videos, articles, other Lemmy communities, etc.

  • @ReallyActuallyFrankenstein
    link
    English
    01 year ago

    Second, I found it very quickly that the amount of RAM you have is critical. My main server is a core i5 4th gen, and so I put AI software on another one of my servers which is a core i5 7th gen. You would think that the latter would work a lot better, but it had half the ram, and it basically wasn’t even able to get started.

    Is there an amount of RAM that’s currently considered the bare minimum for CPU-only self-hosting?

    • exu
      link
      fedilink
      English
      31 year ago

      If you’re using llama.cpp, have a look at the GGUF models by TheBloke on huggingface. He puts approximate RAM required in the readme based on the quantisation level.

      From personal experience I’d estimate 12G for 7B models based on how full RAM was with 16 gigs. For mixtral at least 32G.

      • @ReallyActuallyFrankenstein
        link
        English
        11 year ago

        Thanks, appreciate it (I’m new to local text CPU models, I know it was a stupid question).