• fizzle@quokk.au
    link
    fedilink
    English
    arrow-up
    12
    ·
    5 hours ago

    Most of the power consumption comes from training and optimising models. You only interact with the finished product, so power per query is very low compared to that required to develop the LLM.

    • lime!@feddit.nu
      link
      fedilink
      arrow-up
      2
      ·
      1 hour ago

      while this is true in isolation, the amount of users means that inference now uses more power than training for the large actors.