Abacus.ai:

We recently released Smaug-72B-v0.1 which has taken first place on the Open LLM Leaderboard by HuggingFace. It is the first open-source model to have an average score more than 80.

  • TheChurn@kbin.social
    link
    fedilink
    arrow-up
    2
    ·
    5 months ago

    Every billion parameters needs about 2 GB of VRAM - if using bfloat16 representation. 16 bits per parameter, 8 bits per byte -> 2 bytes per parameter.

    1 billion parameters ~ 2 Billion bytes ~ 2 GB.

    From the name, this model has 72 Billion parameters, so ~144 GB of VRAM

    • FaceDeer@kbin.social
      link
      fedilink
      arrow-up
      3
      ·
      5 months ago

      It’s been discovered that you can reduce the bits per parameter down to 4 or 5 and still get good results. Just saw a paper this morning describing a technique to get down to 2.5 bits per parameter, even, and apparently it 's fine. We’ll see if that works out in practice I guess