• Time@sh.itjust.worksOP
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    2
    ·
    edit-2
    5 months ago

    Don’t you need tons of RAM to run LLMs? I thought the newer models needed up to 64GB RAM? Also, what about Stable Diffusion?

    • april@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      1
      ·
      edit-2
      5 months ago

      Ram is important but it has to be vram not system ram.

      Only MacBooks can use the system ram because they have an integrated GPU rather than a dedicated one.

      Stable diffusion is the same situation.

    • Pumpkin Escobar@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      5 months ago

      Taking ollama for instance, either the whole model runs in vram and compute is done on the gpu, or it runs in system ram and compute is done on the cpu. Running models on CPU is horribly slow. You won’t want to do it for large models

      LM studio and others allow you to run part of the model on GPU and part on CPU, splitting memory requirements but still pretty slow.

      Even the smaller 7B parameter models run pretty slow in CPU and the huge models are orders of magnitude slower

      So technically more system ram will let you run some larger models but you will quickly figure out you just don’t want to do it.

    • Findmysec@infosec.pub
      link
      fedilink
      English
      arrow-up
      2
      ·
      5 months ago

      They do, but VRAM. Unfortunately, the cards that do have that much of memory are used by OEMs/corporations and are insanely pricey