Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This comparison may make sense on short-horizon tasks for which there is no possibility of preparation. Given some weeks to prepare, a good human executive will get the context, while today's best AI systems will completely fail to do so.


Today’s AI systems probably won’t excel, but they won’t completely fail either.

Basically give the LLM a computer to do all kinds of stuff against the real world, kick it off with a high level goal like “build a startup”.

The key is to instruct it to manage its own memory in its computer, and when context limit inevitably approaches, programmatically interrupt the LLM loop and instruct it to jot down everything it has for its future self.

It already kinda works today, and I believe AI systems a year from now will excel at this:

https://dwyer.co.za/static/claude-code-is-all-you-need.html

https://www.anthropic.com/research/project-vend-1




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: