Local AI coding assistant using Claude Code with a local model on NVIDIA DGX Spark.
This setup allows Claude Code to run without using the Anthropic cloud API by redirecting requests to a local vLLM inference server running Qwen3.5. oai_citation:0‡vLLM