GPU Configuration
Supported Backends
Backend
Platform
Hardware
Automatic Detection
$ predictor up ./llama.gguf
Detecting hardware...
CUDA: NVIDIA RTX 4090 (24GB VRAM) ✓
Loading model: llama-7b-q4.gguf
Size: 3.8GB
Loading... ████████████████████ 100%
✓ Tunnel establishedForce CPU Mode
GPU Selection (Multi-GPU)
VRAM Requirements
Typical VRAM Usage (7B Models)
Quantization
VRAM Required
Troubleshooting
CUDA Not Detected
Metal Not Detected
Out of Memory Errors
GPU Temperature
Last updated