Couldn’t make this shit up if I tried.

  • brucethemoose@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    edit-2
    21 hours ago

    Dual 3060s are an option. LLMs can be split across GPUs reasonably well.

    3090s used to be like $700 used, but ironically they’ve gone up in price. I got mine for around $800 awhile ago, and stuffed it into 10L PC.

    Some people buy used P40s. There are rumors of a 24GB Arc B580. Also, AMD Strix Halo APU laptops/mini PCs can host it quite well, with the right software setup… I might buy an ITX board if anyone ever makes one.

    Also, there are 12GB/6GB VRAM distillations too, but 24GB is a huge intelligence step-up.

    • unmagical@lemmy.ml
      link
      fedilink
      English
      arrow-up
      5
      ·
      22 hours ago

      Totally forgot the 3090 had 24GB. It’s definitely still enthusiast territory though.

      • brucethemoose@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        21 hours ago

        For sure.

        The 14B distillation is still quite good, and usable on like 10GB GPUs. Maybe 8 with the right settings.