Re: early response to the medium list in general: This is all 3.5.
4 is significantly smarter. 5 will be too
Yeah, the more it has access to for sure increases it’s capabilities. Like giving it access to modify your terraform to deploy infrastructure. I think it’s really compelling, but I also think there is still a long ways to go. Another ML engineer in a community that I’m in put together a great piece on the current state of AI:
https://medium.com/@khalil.alami/the-boring-state-of-artificial-intelligence-152244075d7f
Re: early response to the medium list in general: This is all 3.5.
4 is significantly smarter. 5 will be too
The same limitations apply… it’s still an LLM limited to a knowledge domain and is acting as an approximation machine. We still need some pathways for generating high quality data for it to be trained off of… if we use it’s “intelligence” as an excuse to stop producing that data, it will stop learning.
I agree with the last half. It’s not clear to me that we aren’t closer to LLMs than many realize, though we’re definitely more like RNNs in some ways. Give it a plug-in to apply a sort of RNN mask to its plug-in memory memory, and that might be enough I think. Definitely not GPT3.5, but give it 5 years and I think they are easily good enough to replace most computer workers.
And totally, we need to figure out how to continue producing high-quality data. Currently we train agents (people) and the number of resources that they can accumulate becomes correlated to their “model” performance. As we see these resources accumulate to them, we become more likely to use them as the definition of good information. The same may be true of future AI agents and additional quality information generation.
Yeah, I think there is a lot of growth that will be happening in the model network architecture space. Using more precise smaller models for tasks that LLMs use to do specialized things, reasoning models (lots of good research going on here), as well as RI for experiments (but interestingly enough those require physics engines that we define, so most or the time they only can solve for that vs actual physics - this in combination with robotics and more sophisticated sensors though has a lot of promise - essentially self driving cars, and there is still a long ways to go there). Collaborative models are super cool though.
Ultimately though, I’m concerned about the data creation. I don’t want there to be a misunderstanding of how important it is that we continue to produce it.
And that certain people stop producing it 🤭
I kid. Mostly. I mean if it really is smart it will hopefully ignore them