Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

A Mac with 16GB RAM can run qwen 7b, gemma 9b and similar models that are somewhere between GPT3.5 and GPT4.

Quite impressive.



on what metric?

Why would OpenAI bother serving GPT4 if customers would be just as happy with a tiny 9B model?


https://lmarena.ai/

Check out the lmsys leaderboard. It has an overall ranking as well as ranking for specific categories.

OpenAI are also serving gpt4o mini. That said afaiu it’s not known how large/small mini is.

Being more useful than GPT3.5 is not a high bar anymore.


Don't confuse GPT-4 and GPT-4o.

GPT-4o is a much better experience than the smaller local models. You can see that in the lmarena benchmarks or from trying them out yourself.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: