Unfortunately it is not that impressive as presented in the paper https://arcprize.org/blog/hrm-analysis
Discussion
One can recreate results with a simpler transformer architecture, without multiple levels. The trick is in training setup, and the iterative Q learning loss, not the hierarchy and the recursion via latent space.