Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

(You can't compare parameter count with a mixture of experts model, which is what the 1.8T rumor says that GPT-4 is.)


You absolutely can since it has a size advantage either way. MoE means the expert model performs better BECAUSE of the overall model size.


Fair enough, although it means we don't know whether a 1.8T MoE GPT-4 will have a "size advantage" over Llama 3 400B.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: