This template allows you to deploy both GPU-accelerated Docker containers like ComfyUI and containers without GPU support, such as n8n. This flexibility enables a wide range of applications, from AI image generation to automated workflows, all within a single, manageable environment. Configure and run your desired Docker containers effortlessly, leveraging the power and convenience of this versatile template.
First, let’s configure Any Docker with your configuration for N8N and persistent data storage so restarts are possible with intact data. This configuration does not include webhooks (for now). Ask support if you need webhooks and we’ll help you configuring them.
Note: you can also use the preconfigured template: n8n
See screenshots of configuration below:
N8N unlocks the potential to run complex workflows directly on your Trooper.AI GPU server. This means you can automate tasks involving image/video processing, data analysis, LLM interactions, and more – leveraging the GPU’s power for accelerated performance.
Specifically, you can run workflows for:
Keep in mind, you’ll need to install AI tools like ComfyUI and Ollama to integrate them into your N8N workflows on the server locally. Also you need enough GPU VRAM to power all models. Do not give that GPUs to the docker running N8N.
You can easily run more docker container and ask for help via: Support Contacts