The list is done according to our best knowledge, although definitely not comprehensive. Check out also the Awesome List of SDKs for AI Agents.
Discussion and feedback appreciated! ❤️
You have something to add or improve about our list? Do it via pull request. Please keep the agents in an alphabetical order and in correct category. Please only add companies and projects powered by autonomous AI agents.
For adding agents'-related SDKs, frameworks and tools, please visit Awesome SDKs for AI Agents
This list is made by the team behind e2b. E2b is building AWS for AI agents. We help elopers to deploy, test, and monitor AI agents. E2b is agnostic to your tech stack and aims to work with any tooling for building AI agents.
Long-short Term Memory: Language agents in the library are equipped with both long-term memory implemented via VectorDB + Semantic Search and short-term memory (working memory) maintained and updated by an LLM.
Tool Usage: Language agents in the library can use any external tools via function-calling and developers can add customized tools/APIs here.
Web Navigation: Language agents in the library can use search engines to navigate the web and get useful information.
Multi-agent Communication: In addition to single language agents, the library supports building multi-agent systems in which language agents can communicate with other language agents and the environment. Different from most existing frameworks for multi-agent systems that use pre-defined rules to control the order for agents' action, Agents includes a controller function that dynamically decides which agent will perform the next action using an LLM by considering the previous actions, the environment, and the target of the current states. This makes multi-agent communication more flexible.
Human-Agent interaction: In addition to letting language agents communicate with each other in an environment, our framework seamlessly supports human users to play the role of the agent by himself/herself and input his/her own actions, and interact with other language agents in the environment.
Symbolic Control: Different from existing frameworks for language agents that only use a simple task description to control the entire multi-agent system over the whole task completion process, Agents allows users to use an SOP (Standard Operation Process) that defines subgoals/subtasks for the overall task to customize fine-grained workflows for the language agents.
Aider is a command line tool that lets you pair program with GPT-3.5/GPT-4, to edit code stored in your local git repository
You can start a new project or work with an existing repo. And you can fluidly switch back and forth between the aider chat where you ask GPT to edit the code and your own editor to make changes yourself
Aider makes sure edits from you and GPT are committed to git with sensible commit messages. Aider is unique in that it works well with pre-existing, larger codebases
Creates tasks based on the result of previous tasks and a predefined objective.
The script then uses OpenAI's NLP capabilities to create new tasks based on the objective
Leverages OpenAI's GPT-4, pinecone vector search, and LangChainAI framework
Default model is OpenAI GPT3-turbo
The system maintains a task list for managing and prioritizing tasks
It autonomously creates new tasks based on completed results and reprioritizes the task list accordingly, showcasing the adaptability of AI-powered language models
A more advanced version of the original BabyAGI code
Improves upon the original framework, by introducing a more complex task management prompt, allowing for more comprehensive analysis and synthesis of information
Designed to handle multiple functions within one task management prompt
Built on top of the GPT-4 architecture, resulting in slower processing speeds and occasional errors
Provides a framework that can be further built upon and improved, paving the way for more sophisticated AI applications
One of the significant differences between BabyAGI and BabyBeeAGI is the complexity of the task management prompt
An AI agent based on @yoheinakajima's BabyAGI which executes shell commands
Automatic Programming, Successfully created an app automatically just by providing feedback. The procedure can be found here.
Automatic Environment Setup, Successfully installed a Flutter environment on Linux in a container, created the Flutter app, and launched it. The procedure can be found here.
A mod of BabyElfAGI, in a series of mods w the naming of BabyAGI in alphabetical order
Self-improving task lists (FOXY method)
By storing a final reflection at the end, and pulling the most relevant reflection to guide future runs, BabyAGI slowly generates better and better tasks lists
Novel Chat UI w parallel tasks
You can chat w BabyAGI! It has an experimental UI where the chat is separate from the tasks/output panel, allowing you to request multiple tasks in parallel
The Chat UI can use a single skill quickly, or chain multiple skills together using a tasklist
New skills
🎨 DALLE skill with prompt assist
🎶 Music player w Deezer
📊 Airtable search (add your own table/base ID)
🔍 Startup Analyst (example of beefy function call as a skill)
CAMEL is an open-source library designed for the study of autonomous and communicative agents.
1)AI user agent: give instructions to the AI assistant with the goal of completing the task.
AI assistant agent: follow AI user’s instructions and respond with solutions to the task
CAMEL also has an open-source community dedicated to the study of autonomous and communicative agents
ChemCrow is an open source package for the accurate solution of reasoning-intensive chemical tasks
It integrates 13 expert-design tools to augment LLM performance in chemistry and demonstrate effectiveness in automating chemical tasks
Built with Langchain
The LLM is provided with a list of tool names, descriptions of their utility, and details about the expected input/output. It is then instructed to answer a user-given prompt using the tools provided when necessary. The instruction suggests the model to follow the ReAct format - Thought, Action, Action Input, Observation. One interesting observation is that while the LLM-based evaluation concluded that GPT-4 and ChemCrow perform nearly equivalently, human evaluations with experts oriented towards the completion and chemical correctness of the solutions showed that ChemCrow outperforms GPT-4 by a large margin. This indicates a potential problem with using LLM to evaluate its own performance on domains that requires deep expertise. The lack of expertise may cause LLMs not knowing its flaws and thus cannot well judge the correctness of task results. (Source: Weng, Lilian. (Jun 2023). LLM-powered Autonomous Agents". Lil’Log. https://lilianweng.github.io/posts/2023-06-23-agent/.)
The purpose of Clippy is to elop code for or with the user. It can plan, write, debug, and test some projects autonomously. For harder tasks, the best way to use it is to look at its work and provide feedback to it.
An AI assistant designed to let you interactively query your codebase using natural language. By utilizing vector embeddings, chunking, and OpenAI's language models, Cody can help you navigate through your code in an efficient and intuitive manner.
DemoGPT leverages the power of Language Models (LLMs) to provide fast and effective demo creation for applications.
Automates the prototyping process, making it more efficient and saving valuable time.
Understands and processes the given prompts to generate relevant applications.
Integrated with LangChain for generating application code through iterative parsing of LangChain's documentation with a "Tree of Transformations" (ToT) approach.
The roadmap for DemoGPT includes constant updates and improvements based on user feedback and real-world application, working towards refining the technology and solving the hallucination problem.
"We are planning to introduce features that will further enhance the application generation process, making it more user-friendly and efficient."
"Tell your AI team what microservice you want to build, and they will do it for you. Your imagination is the limit!!
Welcome to Dev-GPT, where we bring your ideas to life with the power of advanced artificial intelligence! Our automated development team is designed to create microservices tailored to your specific needs, making your software development process seamless and efficient. Comprised of a virtual Product Manager, Developer, and DevOps, our AI team ensures that every aspect of your project is covered, from concept to deployment.
Welcome to the AI Driven Software Development Automation Solution, abbreviated as DevOpsGPT. We combine LLM (Large Language Model) with DevOps tools to convert natural language requirements into working software. This innovative feature greatly improves development efficiency, shortens development cycles, and reduces communication costs, resulting in higher-quality software delivery.
Improved development efficiency: No need for tedious requirement document writing and explanations. Users can interact directly with DevOpsGPT to quickly convert requirements into functional software.
Shortened development cycles: The automated software development process significantly reduces delivery time, accelerating software deployment and iterations.
Reduced communication costs: By accurately understanding user requirements, DevOpsGPT minimizes the risk of communication errors and misunderstandings, enhancing collaboration efficiency between development and business teams.
High-quality deliverables: DevOpsGPT generates code and performs validation, ensuring the quality and reliability of the delivered software.
[Enterprise Edition] Existing project analysis: Through AI, automatic analysis of existing project information, accurate decomposition and development of required tasks on the basis of existing projects.
[Enterprise Edition] Professional model selection: Support language model services stronger than GPT in the professional field to better complete requirements development tasks, and support private deployment.
[Enterprise Edition] Support more DevOps platforms: can connect with more DevOps platforms to achieve the development and deployment of the whole process.
"We know that all great™ projects start with awesome™ detailed functional specifications. Which is typically written in English, or its many other spoken language alternatives.
So what if, instead of writing code from functional specs, we simply compile it directly to code?
Into a future, where we replace nearly everything, with just written text."
GeniA is able to work along side you on your production enviroment, executing tasks on your behalf in your dev & cloud environments, AWS/k8s/Argo/GitHub etc.
Allows you to enhance the platform by integrating your own tools and APIs.
Specify your project, and the AI agent asks for clarification, and then constructs the entire code base
Features
Made to be easy to adapt, extend, and make your agent learn how you want your code to look. It generates an entire codebase based on a prompt
You can specify the "identity" of the AI agent by editing the files in the identity folder
Editing the identity and evolving the main prompt is currently how you make the agent remember things between projects
Each step in steps.py will have its communication history with GPT4 stored in the logs folder, and can be rerun with scripts/rerun_edited_message_logs.py
Dev tool that writes scalable apps from scratch while the developer oversees the implementation
A research project to see how can GPT-4 be utilized to generate fully working, production-ready, apps
The main idea is that AI can write most of the code for an app (maybe 95%) but for the rest 5%, a developer is and will be needed until we get full AGI
Can produce detailed, factual and unbiased research reports
Offers customization options for focusing on relevant resources, outlines, and lessons
Addresses issues of speed and determinism, offering a more stable performance and increased speed through parallelized agent work, as opposed to synchronous operation
Inspired by AutoGPT and the Plan-and-Solve paper
The main idea is to run "planner" and "execution" agents, whereas the planner generates questions to research, and the execution agents seek the most related information based on each generated research question
Conversation with your files which selected by you, no embedding, no vector database!
It's also a AI Prompt Storybook. You can use it to manage some AI preset with your team. It support any IDE and language developer. We provide cli to run web and VSCode extension, Jetbrains plugin is coming soon.
IX is a LangChain based agent platform that includes all the tools to build and deploy fleets of agents that
collaborate to complete tasks. IX is both an editor and a runtime. The editor is a no-code graph style editor for
the design of agents, chains, tools, retrieval functions, and collaborative workflows.
Intuitive graph style no-code editor.
Horizontally scaling agent worker fleet.
Multi-user, multi-agent chat interface.
Smart input auto-completes @mentions and {file} references.
Supports Chroma and other vector databases for document search.
Supports OpenAI API, Anthropic, PaLM, and LLama based models.
Separation of tasks and human-in-the-loop interactions: Lemon Agent is currently holding a Planner Agent and a Solver Agent to keep the agents focussed and increase accuracy. We are planning on adding additional agents real soon. In addition, Lemon Agent will ask for approval at relevant workflow steps to make sure the intended actions are executed.
Unlimited configuration options: Lemon Agent gives you unlimited configuration options (see example here) when defining your workflow. For instance, you can tell Lemon Agent to ask for permission before executing a workflow step or to drop a 🧔♀️ dad joke every time the model executes a workflow step.
UI flexibility: Build any UI on top or engage with Lemon Agent via the built-in CLI.
[Soon] Model & framework agnostic operations: Lemon Agent is a standalone agent, but can easily be integrated into frameworks like LangChain and be used with any model.
Bonus: Identify weak spots in your agent’s decision-making capabilities and move to a more deterministic behavior by further configuring your Lemon Agent workflows. (.html file that can be run without any additional installation)
Default model: GPT-3.5-turbo (also possible with GPT-4)
Modular Auto-GPT Framework
Plug N Play" API - Extensible and modular "Pythonic" framework, not just a command line tool
Features
"Easy to add new features, integrations and custom agent capabilities, all from python code, no nasty config files!"
"Minimal prompt overhead - Every token counts. We are continuously working on getting the best results with the least possible number of tokens."
"Human in the Loop - Ability to "course correct" agents who go astray via human feedback."
"Full state serialization - can save the complete state of an agent, including memory and the states of its tools to a file or python object. No external databases or vector stores required (but they are still supported)!"
Mentat is the AI tool that assists you with any coding task, right from your command line.
Unlike Copilot, Mentat coordinates edits across multiple locations and files. And unlike ChatGPT, Mentat already has the context of your project - no copy and pasting required!
It provides the entire process of a software company along with carefully orchestrated SOPs. Code = SOP(Team) is the core philosophy
The paper about LLM-based multi-agent work spushes forward the idea of autonomous agents collaborating with each other to do more than one can on its own.
MetaGPT incorporates efficient human workflows as a meta programming approach into LLM-based multi-agent collaboration
The paper illustrates how we may treat different instances of the same language models as a "multiagent society", where individual language model generate and critique the language generations of other instances of the language model
The authors find that the final answer generated after such a procedure is both more factually accurate and solves reasoning questions more accurately
Illustrating the quantitative difference between multiagent debate and single agent generation on different domains in reasoning and factual validity
"Whether you are a technical person or a stakeholder, you can use Pezzo effectively. We don't believe that AI prompts should be designed in a developer's code editor. Aside from the technical issues with this approach, it blocks productivity."
Features
Centralized Prompt Management: Manage all AI prompts in one place for maximum visibility and efficiency.
Streamlined Prompt Design, Publishing & Versioning: Create, edit, test and publish prompts with ease.
Observability: Access detailed prompt execution history, stats and metrics (duration, prompt cost, completion cost, etc.) for better insights.
Troubleshooting: Effortlessly resolve issues with your prompts. Time travel to retroactively fine-tune failed prompts and commit the fix instantly.
Cost Transparency: Gain comprehensive cost transparency across all prompts and AI models.
Simplified Integration: Reduce code overhead by 90% by consuming your AI prompts using the Pezzo Client, regardless of the model provider.
Purpose: Gnerate and compose React components from user stories
Stack
React
TailwindCSS
Typescript
Radix UI
Shandcn UI
OpenAI API
The agent is taking a user story text and generating and composing multiple react components to generate the relevant screens, based on atomic design principles
Features
Generate React Components from user stories
Compose React Components from existing components
Use a local design system to generate React Components
Use React, TailwindCSS, Typescript, Radix UI, Shandcn UI
Built with Atomic Design Principles
It is still experimental but very interesting results, It is completely open-sourced, looking for contributors!
Simplifies the configuration and deployment of LLM Agents to production
"One of the core principals of SuperAgent is to build with any third-party dependencies to proprietary tech"
It provides a range of features and functionalities to make it easier for developers to build, manage and deploy AI agents to production including features such as built in memory and document retrieval via vector dbs, powerful tools, webhooks, cron jobs etc.
There are two main types of agents: action agents and plan-and-execute agents
A paper simulating interactions between tens of agents
Presenting an architecture that extends a language model to store and synthesize the agent's experiences, enabling dynamic behavior planning in an interactive sandbox environment with generative agents
A LLM-powered embodied lifelong learning agent in Minecraft that continuously explores the world, acquires diverse skills, and makes novel discoveries without human intervention
Voyager consists of three key components:
an automatic curriculum that maximizes exploration
an ever-growing skill library of executable code for storing and retrieving complex behaviors
a new iterative prompting mechanism that incorporates environment feedback, execution errors, and self-verification for program improvement
Voyager interacts with GPT-4 via blackbox queries, which bypasses the need for model parameter fine-tuning
WorkGPT is an agent framework in a similar fashion to AutoGPT or LangChain. You give it a directive and an array of APIs and it will converse back and forth with the AI until its directive is complete.
For example, a directive could be to research the web for something, to crawl a website, or to order you an Uber. We support any and all APIs that can be represented with an OpenAPI file.
WorkGPT now has OpenAI's new function invocation feature baked into it
While chaining together APIs was possible before (see AutoGPT), it was slow, expensive, and error prone
"This is a Swift port of BabyAGI, an example of an AI-powered task management system that uses OpenAI and Pinecone APIs to create, prioritize, and execute tasks. The main idea behind this system is that it creates tasks based on the result of previous tasks and a predefined objective."
"Adept is building an entirely new way to get things done. It takes your goals, in plain language, and turns them into actions on the software you use every day."
In early stage
"We’re building a machine learning model that can interact with everything on your computer."
Breaks down a high level research question into a step-by-step plan, and executes it
Diverse tools, including a full web browser
Can access internet information without the need for an API
"We don't generate content using AI, as it can be unreliable. Instead, we extract relevant information from trusted sources, cluster and process it into a user-friendly format."
AI-powered query planner intelligently routes and executes requests, ensuring correctness and diverse source selection
BitBuilder Junior is an intern-level code generation tool that lives in your GitHub repository. It follows instructions to implement unambiguous code changes by changing multiple files, writing tests for generated code, and matching the style, framework, and libraries you're already using.
broadn is a no-code platform that helps non-technical people build AI products in minutes. We're faster and more flexible than traditional no-code tools through an LLM powered conversational interface and an agent architecture that automates complex backend/workflow operations
No coding required thanks to user-friendly interface
Full SEO optimization
Picture Upload: Users can conveniently upload and regenerate their own pictures for unlimited customization of their profiles
Profile Customization: Users have the flexibility to customize their profiles by hiding sections, adding social media links, and sharing contact details, allowing them to showcase their unique personality and brand
Instant Preview: Users can instantly visualize their profile changes through a conveniently placed preview button, ensuring a quick assessment of the desired appearance
30% Faster Speed: The app achieves an impressive 30% increase in website generation speed, providing users with a fast and efficient website building experience.
It’s not supposed to be just another coding copilots like GitHub Copilot or Codeium, but autonomous agents capable of autonomously building software from end to end
With Fine-Tuner, you can build sophisticated, tailored AI agents at scale without any need for technical skills or coding. Just bring your data and ideas, and we'll provide the toolset you need to transform them into powerful AI solutions, capable of handling vast amounts of data and users. Take advantage of our scalable platform to meet your growing needs with ease and efficiency
Connecting Your Chatbot to Your App
FineTuner.ai is a no-code AI platform that enables users to create and deploy custom AI agents and components without any coding.
With an intuitive UI/UX and rapid API deployment, FineTuner.ai simplifies AI development, allowing users to focus on their unique use cases and ideas.
4.1. Access the API tab for an overview of the required tokens and parameters to connect your chatbot to your app using REST API endpoints.
The Fine-Tuner REST API provides API endpoints for Fine-Tuner data types that allow to interact with your AI models remotely by sending and receiving JSON
Authentication to the Fine-Tuner API is performed via HTTP Bearer Authentication
A cloud-based platform-as-a-service that allows developers to build smart agents that couple LLMs with back-end logic to interface to data, systems, and tools
GitWit uses a GPT-based agent to generate code and git to track changes made to files
GitWit ties together large language models and modern developer tools
It can spawn and modify codebases using just a single prompt
GitWit is primarily aimed at full-stack developers, and is particularly loved by those with a learning mindset—such as those learning a new stack or technology
It is in early beta and may require some experimentation with the prompts you enter
Your AI Coach and and AI Copilot for course creators, community builders, and coaches. Built into an all-in-one course creation and community software.
"A suite of powerful AI features meant to augment data people"
Hex can explain and document your code
Hex Magic features know about database schemas, past operations, and the project’s execution graph, so they can make deeper, more insightful recommendations
You can see more – and sign up for the waitlist – over here.
The agent runs and controls the local Google Chrome, which allows it to interact with the world/services/web apps, just like people interact with the world/services/web apps using Google Chrome
The agent itself probably also runs locally and currently, it needs the local Google Chrome to function
Our understanding from the demo video is that they use local code and a custom plugin in ChatGPT to control a web browser (e.g., Google Chrome). This setup enables MultiOn to perform tasks like ordering plane tickets as if a human were interacting with the browser directly
Use cases
A lot of cool real use cases, e.g.,
-Sending an email fully autonomously
-Posting a tweet
-Sending a tweet reply to a specific person with a specific message
-Sending a Facebook message to a friend
-Searching for vacation rentals and check pricing for an upcoming trip
-Searching for a wedding venue and starting the wedding planning process
-Scheduling a car wash
After introducing the GPT function calling, MultiOn can call itself recursively to spawn more sub-agents
Instead of calling multiple functions or APIs you just need one Universal Function that can interact with all services and have it call itself to accomplish more complex tasks in parallel
An end-to-end solution, with which it takes 3 minutes not weeks to get a user-facing agent up and running in your app (currently 3 SDKs including React)
Powerful tools built into the admin dashboard and Admin API including analytics, monitoring, rate-limiting, content moderation, etc.
minimizes or eliminates the need for custom backend infrastructure so you can focus on implementing the business logic
Technology-agnostic solution that supports multiple LLM providers (currently 7 models from OpenAI and Anthropic) allowing you to easily switch between models with 1 click
Ready-to-use, highly customizable and beautiful UI components rendering complex interaction trees with support for advanced features like streaming
v0 is a generative user interface system by Vercel Labs powered by AI. It generates copy-and-paste friendly React code based on Shadcn UI and Tailwind CSS.
This list was made by the team behind E2B. E2b is building a Sandbox Runtime for LLM apps and agents - that is, a set of custom sandboxed cloud environments for AI-powered apps and agentic workflows. Get started here.