Unfortunately it is not that impressive as presented in the paper https://arcprize.org/blog/hrm-analysis

Reply to this note

Please Login to reply.

Discussion

One can recreate results with a simpler transformer architecture, without multiple levels. The trick is in training setup, and the iterative Q learning loss, not the hierarchy and the recursion via latent space.