James Webb telescope confirms there is something seriously wrong with our understanding of the universe - eviltoast
  • tal@lemmy.today
    link
    fedilink
    English
    arrow-up
    11
    ·
    8 months ago

    12GB of video ram is enough in 2024

    And then Stable Diffusion showed up

    • Deceptichum@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      1
      ·
      edit-2
      8 months ago

      Im getting away with my 8gb for now.

      Its the language/text stuff that really needs like 30gb GPUs.

      • tal@lemmy.today
        link
        fedilink
        English
        arrow-up
        1
        ·
        8 months ago

        Im getting away with my 8gb for now.

        I don’t think that you can do the current XL models with 8GB, even for low-resolution images. Maybe with --lowvram or something.

        I’ve got a 24GB RX 7900 XT and would render higher resolution images if I had the VRAM – yeah, you can sometimes sort of get a similar effect by upscaling in tiles, but it’s not really a replacement. And I am confident that even if they put a consumer card out with 128GB, someone will figure out some new clever extension that does something fascinating and useful…as long as one can devote a little more memory to it…

        • Deceptichum@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          edit-2
          8 months ago

          I do XL all the time, at about 30-45 seconds per image. 8gb is surprisingly enough for SDXL, and I run like 7gb models with 3-6 Lora on top.