Skip to content

Quickstart

Five minutes from zero to your first capacity plan.

1. Install

bash
cargo install --path crates/hwledger-cli --root /tmp/hwl-install
export PATH="/tmp/hwl-install/bin:$PATH"
hwledger --version
Install + verify in one shot
Install + verify in one shot
Keyframes (0 — VLM-friendly)

Loading manifest…

2. Plan your first model

Pass any HF-style config.json:

bash
hwledger plan tests/golden/deepseek-v3.json --seq 32768 --users 2
Colored VRAM breakdown with live classification
Colored VRAM breakdown with live classification
Keyframes (0 — VLM-friendly)

Loading manifest…

3. Probe local GPUs

bash
hwledger probe list
Detect Apple Silicon, NVIDIA, AMD, Intel backends
Detect Apple Silicon, NVIDIA, AMD, Intel backends
Keyframes (0 — VLM-friendly)

Loading manifest…

Need deeper telemetry? Use hwledger probe watch --interval 1s to stream samples.

4. Export a plan

Convert a plan into flags for your inference engine:

bash
hwledger plan tests/golden/deepseek-v3.json --seq 32768 --export vllm
hwledger plan tests/golden/deepseek-v3.json --seq 32768 --export llama-cpp
hwledger plan tests/golden/deepseek-v3.json --seq 32768 --export mlx

Next steps

Released under the Apache 2.0 License.