dysfun@treehouse.systems ("gaytabase") wrote:
if i hear one more person talk about how open weight models are democratising access to LLMs i am going to scream. this 'small' model used 60GB of RAM. i don't have 60GB of RAM and even if i had a job i could not fucking afford it right now.