AgentMake AI: a software developement kit for developing agentic AI applications that support 14 AI backends and work with 7 agentic components, such as tools and agents. (Developer: Eliran Wong)
Supported backends: anthropic, azure, cohere, custom, deepseek, genai, github, googleai, groq, llamacpp, mistral, ollama, openai, vertexai, xai
9-min introduction 24-min introduction
This SDK incorporates the best aspects of our favorite projects, LetMeDoIt AI, Toolmate AI and TeamGen AI, to create a library aimed at further advancing the development of agentic AI applications.
Windows, macOS, Linux, ChromeOS, Android via Termux
anthropic
- Anthropic API
azure
- Azure OpenAI API
cohere
- Cohere API
custom
- any openai-compatible backends that support function calling
deepseek
- DeepSeek API
genai
- Vertex AI or Google AI
github
- Github API
googleai
- Google AI
groq
- Groq Cloud API
llamacpp
- Llama.cpp Server - locat setup required
mistral
- Mistral API
ollama
- Ollama - local setup required
openai
- OpenAI API
vertexai
- Vertex AI
xai
- XAI API
For simplicity, agentmake
uses ollama
as the default backend, if parameter backend
is not specified. Ollama models are automatically downloaded if they have not already been downloaded. Users can change the default backend by modifying environment variable DEFAULT_AI_BACKEND
.
agentmake
is designed to work with seven kinds of components for building agentic applications:
-
system
- System messages are crucial for defining the roles of the AI agents and guiding how AI agents interact with users. Check out our examples.agentmake
supports the use offabric
patterns assystem
components for runningagentmake
function or CLI options READ HERE. -
instruction
- Predefined instructions that are added to users' prompts as prefixes, before they are passed to the AI models. Check out our examples.agentmake
supports the use offabric
patterns asinstruction
components for runningagentmake
function or CLI options READ HERE. -
input_content_plugin
- Input content plugins process or transform user inputs before they are passed to the AI models. Check out our examples. -
output_content_plugin
- Output content plugins process or transform assistant responses after they are generated by AI models. Check out our examples. -
tool
- Tools take simple structured actions in response to users' requests, with the use ofschema
andfunction calling
. Check out our examples. -
agent
- Agents are agentic applications automate multiple-step actions or decisions, to fulfill complicated requests. They can be executed on their own or integrated into an agentic workflow, supported byagentmake
, to work collaboratively with other agents or components. Check out our examples. -
follow_up_prompt
- Predefined prompts that are helpful for automating a series of follow-up responses after the first assistant response is generated. Check out our examples.
agentmake
supports both built-in agentic components, created by our developers or contributors, and cutoms agentic components, created by users to meet their own needs.
Built-in agents components are placed into the following six folders inside the agentmake
folders:
agents
, instructions
, plugins
, prompts
, systems
, tools
To use the built-in components, you only need to specify the component filenames, without parent paths or file extensions, when you run the agentmake
signature function or CLI options.
agentmake
offers two options for users to use their custom components.
Option 1: Specify the full file path of inidividual components
Given the fact that each component can be organised as a single file, to use their own custom components, users only need to specify the file paths of the components they want to use, when they run the agentmake
signature function or CLI options.
Option 2: Place custom components into agentmake
user directory
The default agentmake
user directory is ~/agentmake
, i.e. a folder named agentmake
, created under user's home directory. Uses may define their own path by modifying the environment variable AGENTMAKE_USER_DIR
.
After creating a folder named agentmake
under user directory, create six sub-folders in it, according to the following names and place your custom components in relevant folders, as we do with our built-in components.
If you organize the custom agentic components in this way, you only need to specify the component filenames, without parent paths or file extensions, when you run the agentmake
signature function or CLI options.
In cases where a built-in tool and a custom tool have the same name, the custom tool takes priority over the built-in one. This allows for flexibility, enabling users to copy a built-in tool, modify its content, and retain the same name, thereby effectively overriding the built-in tool.
Basic:
pip install --upgrade agentmake
Basic installation supports all AI backends mentioned above, except for vertexai
.
Extras:
We support Vertex AI via Google GenAI SDK. As this package supports most platforms, except for Android Termux, we separate this package google-genai
as an extra. To support Vertex AI with agentmake
, install with running:
pip install --upgrade agentmake[genai]
This SDK is designed to offer a single signature function agentmake
for interacting with all AI backends, delivering a unified experience for generating AI responses. The main APIs are provided with the function agentmake
located in this file.
Find documentation at https://github.com/eliranwong/agentmake/blob/main/docs/README.md
The following examples assumes Ollama is installed as the default backend.
To import:
from agentmake import agentmake
To run, e.g.:
agentmake("What is AI?")
To work with parameter tool
, e.g.:
agentmake("What is AgentMake AI?", tool="search/google")
agentmake("How many 'r's are there in the word 'strawberry'?", tool="magic")
agentmake("What time is it right now?", tool="magic")
agentmake("Open github.com in a web browser.", tool="magic")
agentmake("Convert file 'music.wav' into mp3 format.", tool="magic")
agentmake("Send an email to Eliran Wong at [email protected] to express my gratitude for his work.", tool="email/gmail")
A cross-platform solution to work with a tool
that is placed in a sub-folder, e.g.:
agentmake("Extract text from image file 'sample.png'.", tool=os.path.join("perplexica", "github"))
To work with parameters input_content_plugin
and output_content_plugin
, e.g.:
agentmake("what AI model best", input_content_plugin="improve_writing", output_content_plugin="translate_into_chinese", stream=True)
To work with plugin
that is placed in a sub-folder, e.g.:
agentmake("你好吗?", output_content_plugin=os.path.join("chinese", "convert_simplified"))
To automate prompt engineering:
agentmake("what best LLM training method", system="auto", input_content_plugin="improve_prompt")
To work with parameter system
, instruction
, follow_up_prompt
, e.g.:
agentmake("Is it better to drink wine in the morning, afternoon, or evening?", instruction="reflect", stream=True)
agentmake("Is it better to drink wine in the morning, afternoon, or evening?", instruction="think", follow_up_prompt=["review", "refine"], stream=True)
agentmake("Provide a detailed introduction to generative AI.", system=["create_agents", "assign_agents"], follow_up_prompt="Who is the best agent to contribute next?", stream=True, model="llama3.3:70b")
To work with parameter agent
, e.g.:
agentmake("Write detailed comments about the works of William Shakespeare, focusing on his literary contributions, dramatic techniques, and the profound impact he has had on the world of literature and theatre.", agent="teamgenai", stream=True, model="llama3.3:70b")
To specify an AI backend:
agentmake("What is Microsoft stock price today?", tool=os.path.join("search", "finance"), backend="azure")
To work collaboratively with different backends, e.g.
messages = agentmake("What is the most effective method for training AI models?", backend="openai")
messages = agentmake(messages, backend="googleai", follow_up_prompt="Can you give me some different options?")
messages = agentmake(messages, backend="xai", follow_up_prompt="What are the limitations or potential biases in this information?")
agentmake(messages, backend="mistral", follow_up_prompt="Please provide a summary of the discussion so far.")
As you may see, the agentmake
function returns the messages
list, which is passed to the next agentmake
function in turns.
Therefore, it is very simple to create a chatbot application, you can do it as few as five lines or less, e.g.:
messages = [{"role": "system", "content": "You are an AI assistant."}]
user_input = "Hello!"
while user_input:
messages = agentmake(messages, follow_up_prompt=user_input, stream=True)
user_input = input("Enter your query:\n(enter a blank entry to exit)\n>>> ")
These are just a few simple and straightforward examples. You may find more examples at:
https://github.com/eliranwong/agentmake/tree/main/agentmake/examples
Command CLI are designed for quick run of AI features.
Check for CLI options, run:
agentmark -h
Two shortcut commands:
ai
== agentmake
aic
== agentmake -c
with chat features enabled
The available CLI options use the same parameter names as the agentmake
function for AI backend configurations, to offer users a unified experience. Below are some CLI examples, that are equivalent to some of the examples mentioned above:
ai What is AI?
ai What is AgentMake AI --tool search/google
ai Convert file music.wav into mp3 format. --tool task
ai Send an email to Eliran Wong at [email protected] to express my gratitude for his work --tool email/gmail
ai Extract text from image file sample.png. --tool=ocr/openai
ai What is Microsoft stock price today? -t search/finance -b azure
ai what AI model best --input_content_plugin improve_writing --output_content_plugin translate_into_chinese
ai what best LLM training method --system auto --input_content_plugin improve_prompt
ai 你好吗? --output_content_plugin=chinese/convert_simplified
ai Is it better to drink wine in the morning, afternoon, or evening? --instruction think --follow_up_prompt review --follow_up_prompt refine
ai Write detailed comments about the works of William Shakespeare, focusing on his literary contributions, dramatic techniques, and the profound impact he has had on the world of literature and theatre --agent teamgenai --model "llama3.3:70b"
CLI options are handy for testing, e.g. simply use a newly developed tool
file with -t
option and run:
ai What is AgentMake AI? -t ~/my_folder/perplexica.py
To use ollama
as the default backend, you need to download and install Ollama. To use backends other than Ollama, you need to use your own API keys. There are a few options you may configure the AI backends to work with agentmake
.
Specify AI backend configurations as parameters when you run the agentmake
signature function agentmake
.
Setting configurations via option 1 overrides the default configurations set by option 2 and option 3, but the overriding is effective only when you run the function, with the specified configurations. Default configurations described below in option 2 and 3 still apply next time when you run the agentmake
function, without specifying the AI backend parameters. This gives you flexibility to specify different settings in addition to the default ones.
You may manually export individual environment variables listed in https://github.com/eliranwong/agentmake/blob/main/agentmake.env
You may edit a copy of agentmake.env
, e.g.
cd agentmake
cp agentmake.env .env
etextedit .env
The changes apply next time when you run agentmake
function or cli.
Alternately, use built-in agentmake
cli option to edit the variables:
agentmake -ec
This command automatically make a copy of agentmake.env
and save it as .env
if it does not exist. Remember to save your changes before exiting the text editor to make the changes effective.
Remarks:
- Please do not edit the file
agentmake.env
directly, as it is restored to its default values upon each upgrade. It is recommended to make a copy of it and edit the copied file. - Multiple API keys are supported for running backends
cohere
,github
,groq
andmistral
. You may configure API keys for these backend in the.env
file by using commas,
as separators, e.g.COHERE_API_KEY=cohere_api_key_1,cohere_api_key_2,cohere_api_key_3
fabric
is a fantastic third-party project that offers a great collection of patterns.
agentmake
supports the use of fabric
patterns as entries for the system
or instruction
parameters when running the agentmake
signature function or CLI options.
To use a fabric pattern in agentmake
:
- Install fabric
- Specify a fabric pattern in
agentmake
parametersystem
orinstruction
, by prefixing the selected pattern withfabric.
agentmake("The United Kingdom is a Christian country.", tool="search/searxng", system="fabric.analyze_claims")
- add examples
- convert availble ToolMate AI tools into tools that runable with this SDK (... in progess ...)
- add documentation about tool creation
- add built-in system messages
- add built-in predefined instructions
- add built-in prompts