Hi all, i am quite an old fart, so i just recently got excited about self hosting an AI, some LLM…

What i want to do is:

  • chat with it
  • eventually integrate it into other services, where needed

I read about OLLAMA, but it’s all unclear to me.

Where do i start, preferably with containers (but “bare metal”) is also fine?

(i already have a linux server rig with all the good stuff on it, from immich to forjeio to the arrs and more, reverse proxy, Wireguard and the works, i am looking for input on AI/LLM, what to self host and such, not general selfhosting hints)

    • billwashere@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      3 months ago

      100% agree. TechnoTim is quite good. Also take a look at NetworkChuck. But be aware, these two will send you down rabbit holes of self-hosting ideas. Awesome rabbit holes, but rabbit holes nonetheless. I’ve spent weeks playing with stuff they’ve suggested. N8n and MCP is my latest obsession.

  • hendrik@palaver.p3x.de
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    3 months ago

    There’s another community for this: !localllama@sh.itjust.works
    Though we mostly discuss the news and specific questions there, beginner questions are a bit more rare.

    I think you already got a lot of good answers here, LMStudio, OpenWebUI, LocalAI…
    I’d like to add KoboldCpp that’s kind of made for gaming/dialogue, but it can do everything. And from my experience it’s very easy to set up and bundles everything into one program.

    • ProperlyProperTea@lemmy.ml
      link
      fedilink
      English
      arrow-up
      2
      ·
      3 months ago

      Indeed, other than being able to get the model running, having decent hardware is the next most important part.

      3060 12gb is probably cheapest card to get, 3090 or other 24gb card if you can get it

  • vane@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    3 months ago

    You can host ollama and open-webui on container. If you want to wire search you can connect open-webui to playwright (also container) and searxng (also container) and llm will search the web for answers