I gotcha 🤙
In the AI literature they use the term ‘agent’ to describe something that takes actions in an environment. Autogpt like systems can execute terminal commands, call external APIs, etc, so it’s fair to say they are agents in the AI academic sense.
Have you seen autogpt (https://github.com/Significant-Gravitas/Auto-GPT ) and related systems? If you haven’t, they basically make a call to a GPT API that breaks down an initial task/goal (from the user) and stores each step in a database, then recursively calls gpt apis to attempt to solve each sub task and evaluate if it was solved. If it wasn’t solved, it further creates more tasks, stores in the database, and repeats. I tried some sample tasks with OpenAI gpt3.5 and it was pretty terrible at doing anything. Tried to clone git repos that didn’t exist, etc. I’m waiting on an API Key to try GPT4.
I’d be super interested in how your planning to handle the sequence of actions and the memory of the system. Would it be a hardcoded sequence of steps to do everything you want or would you have a system/agent that can do planning?