Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Do you think training in multiple languages could act as a regularization? Just as polyglots are smarter in real life?


I haven’t seen the architecture of QwQ but I just assumed it learns languages insofar as to pick up relationships between words. It must mean it picks up logic across languages. Huh


I thought too so. But then o1 thinks in english and Qwen thinks in chinese. Is there advantage in thinking in different languages?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: