back to top
HomeTechAI ModelsThis Free Tool Let Me Run AI Video, Image and Music Models...

This Free Tool Let Me Run AI Video, Image and Music Models Locally Without ComfyUI

- Advertisement -

I’ve used ComfyUI multiple times. It’s powerful, no question. But installing every model in it feels unnecessarily complicated, some require specific dependencies, version conflicts are tricky to fix, and one wrong install can break models that were already working fine.

I wanted something simpler. Portable. Something I could move between drives, use offline anytime That’s where Stability Matrix came in.

In simple terms it’s an open source package manager for AI models. No terminal setup, no Python conflicts. You pick what you want, it installs it, and you use it.

My preferred setup is WAN2GP, it supports image, video, audio and music generation all in one place, which covers pretty much everything I care about. But you can install whatever fits your workflow.

To show you how simple this actually is, let me walk you through one real example. I wanted to generate music locally. Completely offline. For free. Here’s exactly what happened.

Here’s exactly how I Run AI Models

I’ll use HeartMuLA, a free local music generation model, as the example. Same process works for video, image, TTS everything.

Step 1: Install & Launch Stability Matrix

Stability matrix how to use

Download it for your OS and run the installer. On first launch it asks where you want to store everything. I checked the Portable Mode option, this stores all your data and models in the same folder as the app itself. That means you can move the entire thing to a different drive or computer anytime without reinstalling anything. Genuinely useful.

Step 2: Add WAN2GP Package

wan2gp stability matrix

Once inside, hit the Add Package button at the bottom. It shows you a list of available packages. Search for WAN2GP and click install, No Terminal or python needed. It handles everything automatically.

Step 3: Launch WAN2GP

Install AI Models Locally

Go to All Packages, find WAN2GP, and hit the Launch button. Give it 20-30 seconds. You’ll see a local URL appear in the terminal, something like localhost:7860. That’s your app running locally.

Install AI Models Locally in PC

Step 4: Open in browser and pick your model

HeartMuLA music generator install

Open that URL in any browser. Make sure your internet is on for this part, first time only, it needs to download the model. I selected HeartMuLA-3b from the model list. It downloaded automatically, no manual setup needed.

After that first download it’s yours. Completely offline, anytime.

Step 5: Generate

HeartMuLA music generator

I pasted in some lyrics, hit generate, and waited. On my 8GB VRAM GPU it took around 1-2 minutes. Not instant, but for a fully local, completely free music generation that never touched a server. I’ll take it. The output was a proper track.

What else Stability Matrix can do

Music is just one thing. That’s what surprised me most about this tool, the scope of what it actually covers.

Through WAN2GP alone you can run image generation, video generation, text to speech, and music — all locally, all free, all from the same browser interface. Pick a model, download it once, use it offline forever. The process is identical every time.

But WAN2GP is just one package. Stability Matrix supports a long list of others including ComfyUI, Automatic1111, Fooocus, InvokeAI, and more. So if you’ve already been using any of those, you can manage them all from one place instead of juggling separate installs.

The model browser is also worth mentioning. It connects directly to CivitAI and HuggingFace, so you can browse, download, and organize models without leaving the app. No manual folder management, no hunting for the right file location.

And because it runs in portable mode, your entire setup — models, settings, everything — lives in one folder you can move to a new computer or external drive anytime.

Closing Thoughts

There are plenty of ways to run AI models locally. ComfyUI is great — powerful, flexible, and has a massive community behind it. But if you want something more straightforward to get started with, Stability Matrix is worth a look.

LEAVE A REPLY

Please enter your comment!
Please enter your name here

YOU MAY ALSO LIKE
MOSS-TTS-Nano Real-Time Voice AI on CPU

MOSS-TTS-Nano: Real-Time Voice AI on CPU, Part of an Open-Source Stack Rivaling Gemini

0
Most text-to-speech tools fall into two camps. The ones that sound good need serious hardware. The ones that run on anything sound robotic. MOSS-TTS-Nano is trying to be neither. It's a 100 million parameter model that runs on a regular CPU and it actually sounds good. Good enough that the team behind it built an entire family of speech models around the same core technology, one of which has gone head to head with Gemini 2.5 Pro and ElevenLabs and come out ahead on speaker similarity. It just dropped on April 10th and it's the newest addition to the MOSS-TTS family, a collection of five open source speech models from MOSI.AI and the OpenMOSS team. The family doesn't just cover lightweight local deployment. One of its models MOSS-TTSD outperforms Gemini 2.5 Pro and ElevenLabs on speaker similarity in benchmarks. Another generates voices purely from text descriptions with no reference audio needed. And one is built specifically for real-time voice agents with a 180ms first-byte latency. Nano is the entry point. The family is the story.
Gen-Searcher An Open Source AI That Searches the Web Before Generating Images

Gen-Searcher: An Open Source AI That Searches the Web Before Generating Images

0
Your image generator has never seen today. It was trained months ago, maybe longer, and everything it draws comes from that frozen snapshot of the world. Ask it to generate a current news moment, a product that launched last month, or anything that requires knowing what's happening right now and it fills in the gaps with a confident guess. Sometimes that guess is close. Often it isn't. Gen-Searcher does something none of the mainstream tools do. Before it draws a single pixel, it goes and looks things up. It searches the web. It browses sources. It pulls visual references. Then it generates. The result is an image grounded in actual current information. It's open source, the weights are on Hugging Face, and the team released everything including code, training data, benchmark, the lot.
MiniMax M2.7 The Agentic Model That Helped Build Itself

MiniMax M2.7: The Agentic Model That Helped Build Itself

0
MiniMax handed an internal version of M2.7 a programming scaffold and let it run unsupervised. Over 100 rounds it analyzed its own failures, modified its own code, ran evaluations, and decided what to keep and what to revert. The result was a 30% performance improvement with nobody directing each step. That is not a benchmark result. That is a different way of thinking about how AI models get built. M2.7 is now available on HuggingFace with weights you can download and deploy. NVIDIA is offering free API access if you want to try it without the hardware overhead. The license has a commercial limitation worth knowing about, we will get to that.

Don’t miss any Tech Story

Subscribe To Firethering NewsLetter

You Can Unsubscribe Anytime! Read more in our privacy policy