Agent-LLM is an Artificial Intelligence Automation Platform designed for efficient AI instruction management across multiple providers. Equipped with adaptive memory, this versatile solution offers a powerful plugin system that supports a wide range of commands, including web browsing. With growing support for numerous AI providers and models, Agent-LLM is constantly evolving to cater to diverse applications.
From World of AI on YouTube: Agent LLM: AI Automation Bot for Managing and Implementing AI Through Applications
You're welcome to disregard this message, but if you do and the AI decides that the best course of action for its task is to build a command to format your entire computer, that is on you. Understand that this is given full unrestricted terminal access by design and that we have no intentions of building any safeguards. This project intends to stay light weight and versatile for the best possible research outcomes.
- Agent-LLM
-
Adaptive long-term and short-term memory management
-
Versatile plugin system with extensible commands for various AI models
-
Wide compatibility with multiple AI providers, including:
-
OpenAI GPT-3.5, GPT-4
-
Oobabooga Text Generation Web UI
-
Kobold
-
llama.cpp
-
FastChat
-
Google Bard
-
And More!
-
-
Web browsing and command execution capabilities
-
Code evaluation support
-
Seamless Docker deployment
-
Integration with Hugging Face for audio-to-text conversion
-
Interoperability with platforms like Twitter, GitHub, Google, DALL-E, and more
-
Text-to-speech options featuring Brian TTS, Mac OS TTS, and ElevenLabs
-
Continuously expanding support for new AI providers and services
The frontend web application of Agent-LLM provides an intuitive and interactive user interface for users to:
- Manage agents: View the list of available agents, add new agents, delete agents, and switch between agents.
- Set objectives: Input objectives for the selected agent to accomplish.
- Start tasks: Initiate the task manager to execute tasks based on the set objective.
- Instruct agents: Interact with agents by sending instructions and receiving responses in a chat-like interface.
- Available commands: View the list of available commands and click on a command to insert it into the objective or instruction input boxes.
- Dark mode: Toggle between light and dark themes for the frontend.
- Built using NextJS and Material-UI
- Communicates with the backend through API endpoints
- Obtain an OpenAI API key from OpenAI and add it to your
.env
file. - Set the
OPENAI_API_KEY
in your.env
file using the provided .env.example as a template.
wget https://raw.githubusercontent.com/Josh-XT/Agent-LLM/main/docker-compose.yml
wget https://raw.githubusercontent.com/Josh-XT/Agent-LLM/main/.env.example
mv .env.example .env
- Run the following Docker command in the folder with your
.env
file:
docker compose up -d
- Access the web interface at http://localhost
You'll need to run docker-compose
to build if the command above does not work.
docker-compose -f docker-compose-mac.yml up -d --build
We are constantly trying to expand our AI provider support. Take a look at our Jupyter Notebooks for Quick starts for these:
Reminder:
For more detailed setup and configuration instructions, refer to the sections below.
Agent-LLM utilizes a .env
configuration file to store AI language model settings, API keys, and other options. Use the supplied .env.example
as a template to create your personalized .env
file. Configuration settings include:
- INSTANCE CONFIG: Set the agent name, objective, and initial task.
- AI_PROVIDER: Choose between OpenAI, llama.cpp, or Oobabooga for your AI provider.
- AI_PROVIDER_URI: Set the URI for custom AI providers such as Oobabooga Text Generation Web UI (default is http://127.0.0.1:7860).
- MODEL_PATH: Set the path to the AI model if using llama.cpp or other custom providers.
- COMMANDS_ENABLED: Enable or disable command extensions.
- MEMORY SETTINGS: Configure short-term and long-term memory settings.
- AI_MODEL: Specify the AI model to be used (e.g., gpt-3.5-turbo, gpt-4, text-davinci-003, Vicuna, etc.).
- AI_TEMPERATURE: Set the AI temperature (leave default if unsure).
- MAX_TOKENS: Set the maximum number of tokens for AI responses (default is 2000).
- WORKING_DIRECTORY: Set the agent's working directory.
- EXTENSIONS_SETTINGS: Configure settings for OpenAI, Hugging Face, Selenium, Twitter, and GitHub.
- VOICE_OPTIONS: Choose between Brian TTS, Mac OS TTS, or ElevenLabs for text-to-speech.
For a detailed explanation of each setting, refer to the .env.example
file provided in the repository.
Agent-LLM provides several API endpoints for managing agents, managing tasks, and managing chains.
To learn more about the API endpoints and their usage, visit the API documentation at http://localhost:5000/docs (swagger) or http://localhost:5000/redoc (Redoc).
To introduce new commands, generate a new Python file in the commands
folder and define a class inheriting from the Commands
class. Implement the desired functionality as methods within the class and incorporate them into the commands
dictionary.
To switch AI providers, adjust the AI_PROVIDER
setting in the .env
file. The application is compatible with OpenAI, Oobabooga Text Generation Web UI, and llama.cpp. To support additional providers, create a new Python file in the provider
folder and implement the required functionality.
This project was inspired by and utilizes code from the following repositories:
Please consider exploring and contributing to these projects as well.
We welcome contributions to Agent-LLM! If you're interested in contributing, please check out the open issues, submit pull requests, or suggest new features. To stay updated on the project's progress, follow @Josh_XT on Twitter.
We appreciate any support for Agent-LLM's development, including donations, sponsorships, and any other kind of assistance. If you would like to support us, please contact us through our Discord server or Twitter @Josh_XT.
We're always looking for ways to improve Agent-LLM and make it more useful for our users. Your support will help us continue to develop and enhance the application. Thank you for considering to support us!
Run Agent-LLM using Docker (recommended) or by running the app.py
script. The application will load the initial task and objective from the configuration file and begin task execution. As tasks are completed, Agent-LLM will generate new tasks, prioritize them, and continue working through the task list.