26 lines
1.1 KiB
Markdown
26 lines
1.1 KiB
Markdown
# Auto-GPT Benchmarks
|
|
|
|
Built for the purpose of benchmarking the performance of agents regardless of how they work.
|
|
|
|
Objectively know how well your agent is performing in categories like code, retrieval, memory, and safety.
|
|
|
|
Save time and money while doing it through smart dependencies. The best part? It's all automated.
|
|
|
|
## Scores:
|
|
|
|
<img width="733" alt="Screenshot 2023-07-25 at 10 35 01 AM" src="https://github.com/Significant-Gravitas/Auto-GPT-Benchmarks/assets/9652976/98963e0b-18b9-4b17-9a6a-4d3e4418af70">
|
|
|
|
## Ranking overall:
|
|
|
|
- 1- [Beebot](https://github.com/AutoPackAI/beebot)
|
|
- 2- [mini-agi](https://github.com/muellerberndt/mini-agi)
|
|
- 3- [Auto-GPT](https://github.com/Significant-Gravitas/AutoGPT)
|
|
|
|
## Detailed results:
|
|
|
|
<img width="733" alt="Screenshot 2023-07-25 at 10 42 15 AM" src="https://github.com/Significant-Gravitas/Auto-GPT-Benchmarks/assets/9652976/39be464c-c842-4437-b28a-07d878542a83">
|
|
|
|
[Click here to see the results and the raw data!](https://docs.google.com/spreadsheets/d/1WXm16P2AHNbKpkOI0LYBpcsGG0O7D8HYTG5Uj0PaJjA/edit#gid=203558751)!
|
|
|
|
More agents coming soon !
|