Ah nice. Ok yeah we’re pretty aligned then.
Main diff is this maybe: We can tell GPT-4 about its weaknesses, and APIs that address them, and it then uses those APIs when needed. I think if we can scale the input tokens to about 1-2 million, and pair it with good APIs for its weaknesses (like a what character is in index n, or a physics sim) we might be just a few years from an AI system surpassing us.
How many tokens do you need to represent the non-googleable context you use to do your job?