Hello, I’ve been hearing a lot about this new DeepSeek LLM, and was wondering, would it be possible to get the 600+ billion parameter model running on my GPU? I’ve heard something about people have got it to run on their MacBooks. I have i7 4790K, 32GB DDR3, and 7900 XTX 24GB VRAM. I’m running Arch Linux, this computer is just for AI stuff really, not gaming as much. I did tried running the distilled 14B parameter model, but it didn’t work for me, I was using GPT4All to run it. I’m thinking about getting one of the NVIDIA 5090s in the future. Thanks in advance!

  • Here_for_the_dudes@sh.itjust.works
    link
    fedilink
    arrow-up
    0
    ·
    3 days ago

    I run the 32b Version on my 6700xt with an R9 3700x using ollama. It runs well but it gets a bit slower on complex problems. I once ran an 70b Llama model, but it took a long time to finish.

    • ieatpwns@lemmy.world
      link
      fedilink
      arrow-up
      0
      ·
      2 days ago

      Hey not to side track ops post or your own but I’m new to the home llm space and I was wondering once you have the model set up is there a gui? And how do you input tasks for it to do?

    • Fisch@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 days ago

      I have the same GPU but I always run 7B/8B variants as exl2. Do you use GGUF to use your system RAM?

    • MIXEDUNIVERS@discuss.tchncs.de
      link
      fedilink
      Deutsch
      arrow-up
      0
      ·
      3 days ago

      i also have a 6700xt but i don’t get ollama running on it. it only defaults to the cpu ryzen 5600 I plan to tackle this problem on a free weekend and now i have a new Reason for solving it.