π AgentAI Deployment Guide
Pre-trained using your own data for full control and privacy.
This setup enables you to deploy AgentAI with support for the following services:
- π OpenWebUI
- π§ Ollama Server
- π€ N8N Workflow
- π Pipelines Server
- π Example Pipeline Script (
aiagent.py
) - π¦ AIAgent - Tech support chat bot. Rust example application (based on LangChain)
- π Qdrant vector database
With this configuration, you can create custom AI agents using your own infrastructure (on-premises, private cloud, or VPS) with providers like:
οΈπ Hetzner Cloud | οΈπ AWS | οΈπ Google Cloud | οΈπ Azure
All applications are packed into docker containers.
π Accessing Deployed Services
Once the package is deployed, you can access the services at:
Service URL's:
π¬ OpenWebUI
http://<your-server-ip>:3001
http://chat.<yourdomain>.<tld>
π N8N Workflow
http://<your-server-ip>:5678
http://n8n.<yourdomain>.<tld>
π§ Ollama Server
http://<your-server-ip>:11434
ποΈ All Docker files and configurations are located in /root/
on your server.
π§ Custom RAG App Integration with OpenWebUI
To build your own RAG (Retrieval-Augmented Generation) application and connect it to OpenWebUI:
- β The Pipelines server is pre-configured and running in Docker.
- π Add your pipeline script to:
/root/openwebui/pipelines/
oru
se the example file provided:aiagent.py
.
π Connecting Pipelines Server to OpenWebUI
- π Go to OpenWebUI β Account Settings β Admin Settings
β Add the following:
URL: http://pipelines:9099
API Key: Provided in the PDF attachment in your welcome email
- π§Ή Navigate to βPipelinesβ from the left menu.
- β¬οΈ Select your custom script from the dropdown.
- π Save the configuration.
- π¬ Your script will now appear in the model selection dropdown in the chat interface.
π¦ AIAgent - Tech support chat bot
The data source file for the app can be modified in:
/root/aiagent/datasource/csv/example.csv
π Server Requirements
To run AgentAI and related services efficiently, we recommend the following hardware specifications:
- π‘ GPU (Recommended): Greatly improves LLM response speed and model performance.
- π CPU: More cores = better performance. Minimum 8 vCPUs recommended.
- π RAM: At least 16 GB for baseline functionality. More is better, especially when running multiple containers.
We tested the package on an 8 vCPU / 16 GB RAM server without GPU, and it ran slowly. While GPU is not strictly required, it is strongly recommended.
You can install this package on your own server, which may be more cost-effective and customizable than using public cloud services.
β οΈ Disclaimer
β οΈ Experimental Version.
This release is for testing and development purposes only. It is not production-ready.
Companion apps

Workflow automation tool that combines AI capabilities with business process automation
Get DevOps Support Hourly
By selecting this service, you will have priority access to the development team for optimizing and upgrading your stack.
Contact info@try.direct for more information or Hire an expert directly to make post-installation adjustments on an hourly basis. Browse other available services.

Open WebUI is a self-hosted AI platform designed to operate entirely offline, it's extensible, feature-rich, and user-friendly

Ollama makes it easy to get up and running with large language models on your own server