Bring the price down!...
TLDR
- Amazingly cheap model and historical low price at 1m token for just a few cents?! That's insane.
- Full open source model that you can host at home if you have like 400GB machine(M3 Ultra 512GB?)...
- Distilled model on ollama is crap and I feel it's even worse than llama3 or phi4 with my own translation tasks.
- The hype seems to be on true open source model (what openai should have done year ago). Hope one day it'll come out with smaller size but as smart as big one.
- The model is unfortunately censored 🚨.