Usage
Options
template hash (required, but Note: if you use this field, you can skip search_params, as they are automatically inferred from the template)
template id (optional)
Disable default search param query args (alias:
--no-default)launch args string for create instance ex: “—onstart onstart_wget.sh —env ‘-e ONSTART_PATH=https://s3.amazonaws.com/vast.ai/onstart_OOBA.sh’ —image atinoda/text-generation-webui:default-nightly —disk 64”
deployment endpoint name (allows multiple workergroups to share same deployment endpoint)
deployment endpoint id (allows multiple workergroups to share same deployment endpoint)
number of workers to create to get an performance estimate for while initializing workergroup (default 3)
estimated GPU RAM req (independent of search string)
search param string for search offers ex: “gpu_ram>=23 num_gpus=2 gpu_name=RTX_4090 inet_down>200 direct_port_count>2 disk_space>=64”
[NOTE: this field isn’t currently used at the workergroup level] minimum floor load in perf units/s (token/s for LLms)
[NOTE: this field isn’t currently used at the workergroup level] target capacity utilization (fraction, max 1.0, default 0.9)
[NOTE: this field isn’t currently used at the workergroup level]cold/stopped instance capacity target as multiple of hot capacity target (default 2.0)
min number of workers to keep ‘cold’ for this workergroup
Description
Create a new autoscaling group to manage a pool of worker instances. Example: vastai create workergroup —template_hash HASH —endpoint_name “LLama” —test_workers 5Global Options
The following options are available for all commands:| Option | Description |
|---|---|
--url URL | Server REST API URL |
--retry N | Retry limit |
--raw | Output machine-readable JSON |
--explain | Verbose explanation of API calls |
--api-key KEY | API key (defaults to ~/.config/vastai/vast_api_key) |