Could you try writing even in this simple language a longer program? Just simply increase the input to 20x or something around that. Iām interested in whether it will break and if it does, at what length.
Interesting, it screwed up at step 160. I think it probably ran out of context, if I explicitly told it to output each step in a more compact way it might do better. Or if I had access to the 32k context length it would probably get 4x further.
Actually it might be worth trying to get it to output the original instructions again every 100 steps, so that the instructions are always available in the context. The ChatGPT UI still wouldn't let you output that much at once but the API would.
Do you mean you define a programming language/bytecode and then feed it into the model?
He's an example where GPT-4 did this perfectly for a very sinple language. This was my first attempt, I did not have to do any trial an error.
https://pastebin.com/4YA5wpie