Skip to Content

Exploring the DGX Spark - Lighthouse Labs

Your very own AI/LLMs in a box!
May 15, 2026 by
Exploring the DGX Spark - Lighthouse Labs
Lighthouse IT Solutions, Matthew Almendinger

In today’s journey into the world of AI, we came across a delightful treat, especially for technology enthusiasts like me. With the rising interest in running AI and Large Language Models (LLMs) locally, we decided to invest in cutting-edge equipment, seizing the opportunity before prices continue to skyrocket. Enter the NVIDIA DGX Spark, a compact powerhouse designed for AI endeavors, right here in our lab.

The Unboxing Experience

Unboxing the NVIDIA DGX Spark felt like opening a gift on Christmas morning. This AI miniature supercomputer boasts impressive specifications, including an NVIDIA Grace Blackwell processor, built on a 20-core ARM architecture, and a graphics processor of NVIDIA’s Blackwell style. It features 128 gigabytes of unified RAM and four terabytes of auto-encrypting storage, with support for clustering through its two 200 gigabit SFP ports. Additional perks include four USB Type-C connectors, Wi-Fi 7, Bluetooth 5.3, and a custom version of Debian dubbed the NVIDIA DGX OS.

Setting Up the DGX Spark

After connecting the DGX Spark, it was time to explore hosting AI locally. We placed our powerhouse device in a networked rack, providing it with a stable environment to perform at its best. Our objective wasn’t to offer a tutorial but an introduction to what the DGX Spark can achieve, highlighting its potential to host LLMs so effortlessly.

Hosting AI Locally with Docker and Ollama

To set up a system capable of running large language models, Docker became our tool of choice due to its ease of use with prepackaged LLMs. Today, we utilized Ollama, a popular engine that offers numerous prebuilt optimization packages, and Open Web UI for communication, similar to tools like ChatGPT.

By downloading the Open Web UI Docker container, we seamlessly connected with Ollama-specific versions. Leveraging the power of GPUs, we ran the Docker container, enabling us to interface with our AI model through a web browser. The system worked smoothly, albeit with initial delays attributed to the model warming up and caching necessary data.

Interacting with AI Locally

Once downloaded and configured, the DGX Spark surprised us with its performance, effectively hosting GPT-OSS, an open-source variant of ChatGPT’s model. Although not as instantaneous as cloud-based AI services, it provided a local, private environment to interact with AI capabilities.

Through Open WebUI, we efficiently engaged with AI models and explored additional features such as augmented support and interfacing with other applications, showcasing the versatility of this setup.

In Conclusion

The NVIDIA DGX Spark proved to be an exceptional investment for anyone looking to explore running AI models locally. With its robust specifications and the versatility offered through Docker and Ollama, hosting AI models privately becomes not just feasible but also enjoyable. As we tread carefully into the future of AI, prioritizing privacy and efficient performance, the DGX Spark stands out as a remarkable ally in AI research and development.

Have a great day exploring the endless potential of AI in your home or workplace!