Abacus.ai:

We recently released Smaug-72B-v0.1 which has taken first place on the Open LLM Leaderboard by HuggingFace. It is the first open-source model to have an average score more than 80.

  • kakes@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    9
    ·
    11 months ago

    Afaik you can substitute VRAM with RAM at the cost of speed. Not exactly sure how that speed loss correlates to the sheer size of these models, though. I have to imagine it would run insanely slow on a CPU.

    • Infiltrated_ad8271@kbin.social
      link
      fedilink
      arrow-up
      7
      ·
      edit-2
      11 months ago

      I tested it with a 16GB model and barely got 1 token per second. I don’t want to imagine what it would take if I used 16GB of swap instead, let alone 130GB.