I've been running some of the biggest open-weight LLMs for free on Nvidia's cloud
The author, a technical editor at XDA, shares his experience running large open-weight language models for free using Nvidia's Build platform, which provides access to powerful models hosted on Nvidia's DGX Cloud without requiring a credit card or incurring direct GPU costs. Despite some models being marked for deprecation, the service offers a broad and regularly updated selection of high-quality models, including notable ones like MiniMax M2.7, Llama 4 Maverick, and Gemma 3. This free tier allows developers and enthusiasts to experiment with models that would otherwise be impractical to run locally due to hardware limitations.
Opening excerpt (first ~120 words) tap to expand
{ "@context": "https://schema.org", "@type": "BreadcrumbList", "itemListElement": [ { "@type": "ListItem", "position": "1", "name": "Home", "item": "https://www.xda-developers.com/" }, { "@type": "ListItem", "position":"2", "name": "AI tools", "item": "https://www.xda-developers.com/ai-tools/" }, { "@type": "ListItem", "position":"3", "name": "I've been running some of the biggest open-weight LLMs for free on Nvidia's cloud", "item": "https://www.xda-developers.com/ive-been-running-some-biggest-open-weight-llms-free-nvidia-cloud/" } ] } I've been running some of the biggest open-weight LLMs for free on Nvidia's cloud By Adam Conway Published Apr 30, 2026, 6:00 AM EDT I’m Adam Conway, an Irish technology fanatic with a BSc in Computer Science and I'm XDA’s Lead Technical Editor.
…
Excerpt limited to ~120 words for fair-use compliance. The full article is at XDA Developers.