Machine concepts
Machines in Comfy Deploy are what run your Workflows. Here you’ll define- custom nodes
- particular models
- GPU type
- run timeout
- idle timeout
Create a new machine
Clicking the “New machine +” button on the top right on the machines page.Options
Custom nodes
The search functionality is based on the ComfyUI Manager list, so if it’s in there it will be here. If you have your own nodes that aren’t on the list, hosted in a git repo. You can instead click the “edit” button. to revealModels
Most models you’ll want to use will be in the storage tab. And we’ll be looking to remove this and move fully to the storage tab The models that you can add to a machine are limited to very specific models most noticiably animate-diff This is due to quirks with the custom nodes that use these models.GPU type
OptionsModel | VRAM (GB) | Performance Level | Cost per Second |
---|---|---|---|
T4 | 16 | Standard | $ |
L4 | 16 | Standard | $ |
A10G | 40 | High | $ |
A100 | 80 | Very High | $ |
H100 | 80 | Double A100 | $ |
Run timeout
This is the maximum amount of time your GPUs will be able to run, default is 300 seconds or 5 minutes.Idle timeout
After your first workflow run, you might see a delay, this is due to whats called a “cold start”. Where the GPU must be provisioned and the model loaded into memory before it can be used. However that is a one time cost if the GPU stays up running. What Idle timeout allows you to do is to, keep the GPU “warm” for a period of time (default 60 seconds), where subsequent requests while warm, will avoid the “cold start” problem.Custom Machines
Custom Machines are how you can use Comfy Deploy to route requests to your self-managed GPUs.Setup
You’ll need to have a few things- Endpoint
- Auth token (optional)
Types
- modal-serverless
- comfy-deploy-serverless
- classic
- runpod-serverless
Classic
classic is how you can connect any generic endpoint, we’ll route the request to and you can handle the rest. Use cases that our users are using- AWS Load balancer in front of GPUS
- Pods on runpod
- Personal GPU cluster