• tyrant@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    9 hours ago

    I’d imagine it depends on the size of the llm. My local llm is about 20gb and pegs the GPU for maybe 5-10 seconds (6700xt 12gb) so you could probably extrapolate from that based on that. I’m sure these giant AI gpus would be more efficient though so maybe not.