One-command vLLM installation for NVIDIA DGX Spark with Blackwell GB10 GPUs (sm_121 architecture)
-
Updated
Oct 28, 2025 - Shell
One-command vLLM installation for NVIDIA DGX Spark with Blackwell GB10 GPUs (sm_121 architecture)
Headless remote desktop setup for NVIDIA DGX SPARK using Sunshine streaming
Serve the home! Inference stack for your Nvidia DGX Spark aka the Grace Blackwell AI supercomputer on your desk. Mostly vLLM based for now
vLLM + Qwen3.5-122B-A10B-NVFP4 on NVIDIA DGX Spark (GB10/SM121) — single-GPU NVFP4 W4A4 with MTP speculative decoding, self-contained Docker build
headless remote desktop to your dgx spark in crystal clear 4k
LLM fine-tuning with LoRA + NVFP4/MXFP8 on NVIDIA DGX Spark (Blackwell GB10)
GPU-accelerated WhisperX on NVIDIA Blackwell (SM_121) - DGX Spark compatible
The definitive Strix Halo LLM guide — 65 t/s on a $2,999 mini PC. Live benchmarks, tested optimizations, and everything that doesn't work.
Multi-model LLM serving for NVIDIA DGX Spark with vLLM, web UI, and tool calling
Retrieval Augmented Generation containerized for Nvidia DGX Spark
Docker Compose configs for running LLM inference on DGX Spark (TensorRT-LLM / vLLM / NIM)
NVIDIA DGX Spark ressources
NVML unified memory shim for NVIDIA DGX Spark Grace Blackwell GB10 - enables MAX Engine, PyTorch, and GPU monitoring
Add a description, image, and links to the dgx-spark topic page so that developers can more easily learn about it.
To associate your repository with the dgx-spark topic, visit your repo's landing page and select "manage topics."