DeepSeek 'shared user data' with TikTok owner ByteDance - eviltoast
    • BakedCatboy@lemmy.ml
      link
      fedilink
      English
      arrow-up
      3
      ·
      5 days ago

      If you have a lot of RAM, you can run small models slowly on the CPU. Your integrated graphics I would guess won’t fit anything useful in it’s vram, so if you really want to run something locally, getting some extra sticks of RAM is probably your cheapest option.

      I have 64G and I run 8-14b models. 32b is pushing it (it’s just really slow)