Yes you read that correctly folks,
Qwen has smashed this one out of the park.
Thanks to the power of Quantization (the art of compressing models like ZIP files), we're able to get the model size down from 93gb to just 22GB while retaining 90% of it's intelligence.. Note it's an 80b model so in this case you'll still have a fantastic time even with a Q1 or Q2 version.
I'm using Q2 K_M and it hasn't failed once.
You too can run this incredible model completely locally, offline and without paying a single dollar.
We'll cover this in depth in tomorrow's live so don't forget to tune in!