When I first got into local LLMs nearly 3 years ago, in mid 2023, the frontier closed models were ofcourse impressively capable.

I then tried my hand on running 7b size local models, primarily one called Zephyr-7b (what happened to these models?? Dolphin anyone??), on my gaming PC with 8GB AMD RX580 GPU. Fair to say it was just a curiosity exercise (in terms of model performance).

Fast forward to this month, I revisit local LLM. (Although I no longer have the gaming PC, cost-of-living-crisis anyone 😫 )

And, the 31b size models look very sufficient. #Qwen has taken the helm in this order. Which is still very expensive to setup locally, although within grasp.

I’m rooting for the edge-computing models now - the ~2b size models. Due to their low footprint, they are practical to run in a SBC 24/7 at home for many people.

But these edge models are the ‘curiosity category’ now.

  • ntn888@lemmy.mlOP
    link
    fedilink
    English
    arrow-up
    3
    ·
    2 days ago

    I didn;t try any 7b ones lately, they may be better fit for 16gb I think. I was able to try the 2b ones as I mentioned (on cpu). they are subpar. like mentioned the usable ones were 31b, I think you need atleast 24gb vram for most models though. maybe someone else can suggest better.

    • ☂️-@lemmy.ml
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      2 days ago

      bummer. spilling to old computer sysram is painful for the smarter models too.

      • ntn888@lemmy.mlOP
        link
        fedilink
        English
        arrow-up
        2
        ·
        2 days ago

        you can give “unloading some layers to RAM” a try though… that way you can get your hands on the “usable” 31b models. browse around to find some good 31b ones… GL