The best local AI model for Home Assistant isn’t always the biggest one


Using a Large Language Model (LLM) with Home Assistant has a number of benefits. It can add natural language understanding, power your voice assistant, and even analyze images. A local LLM can help maintain privacy, but you don’t always need to use the largest model.

Using a local LLM with Home Assistant

Keep your data private

LM Studio setup guide asking to download first local model

There are plenty of ways that an LLM can make Home Assistant more powerful. One popular usage is to hook up a cloud-based model from a company such as OpenAI and use it as a conversation assistant for the Assist voice assistant. This allows you to use natural language commands to control your smart home, rather than having to remember the specific phrases that will turn your lights on and off.

The trouble with using a cloud-based LLM is that data about your smart home has to be sent to the cloud to be processed. It means that information about your smart home ends up on third-party servers. Home Assistant was designed to help maintain your privacy, so sharing information about your smart home with AI companies goes against this core principle.

One solution is to use a local LLM. You can run models on your own hardware that can perform some of the same tasks that cloud-based LLMs can do. How quickly or accurately a local LLM can perform these tasks depends both on the hardware you run it on and the models that you use.

Why the biggest model isn’t always the best

Finding the sweet spot

LLMs often come in different sizes. You might see versions of the same model with values such as 4B, 9B, 70B. These refer to the number of parameters the model has; a 70B model has 70 billion parameters, for example. These larger models often have more capacity for knowledge and reasoning.

The flip side is that the more parameters a model has, the more VRAM is needed to store those parameters. Some 70B models, for example, might need more than 100 GB of VRAM to run. That’s beyond the reach of even high-end consumer GPUs unless you’re running a multi-GPU stack, and if you don’t have enough VRAM, the model won’t run at all or will slow to a crawl.

The challenge is finding a model that’s small enough to run on your hardware, but powerful enough to handle the jobs that you want it to do. There are some useful tools, such as llmfit, that can tell you which models are best suited to run on your hardware.

The good news is that, as the technology has developed, new smaller models have appeared that can outperform the very large models from just a few years ago. You no longer need to have insane amounts of VRAM to get decent performance from a local LLM.

Small local LLMs can run on basic hardware

You don’t need an expensive GPU

A Raspberry Pi in a case lying on top of a Beelink Mini S12 Pro mini PC. Credit: Adam Davidson / How-To Geek

If you don’t have a dedicated GPU, it’s not the end of the world. There are some smaller models that are capable of running on a CPU alone without needing to pass anything off to a GPU. These models can use the system RAM in your PC rather than fitting everything into VRAM. While the performance can’t match larger models running with a GPU, they can still do a job.

I wanted to use a local LLM on my Beelink Mini PC that has 16 GB of RAM and no dedicated GPU. My main purpose was to take a list of events from my calendar and transform it into the text for a spoken morning briefing. I’d seen a lot of people saying they’d found the Qwen 3.5 4B model to be a good sweet spot, so I decided to give it a try.

Using this model, I was able to generate the briefing, although it took around 13 seconds to generate. The text was fine, but it wasn’t particularly inspiring.

In an effort to speed things up, I tried a smaller model, Llama 3.2 3B, which uses fewer parameters. You might expect this model to produce a worse result, but it produced a much more natural-sounding output, and did it in under 6 seconds, less than half the time of the other model.

It seems that size isn’t everything. The largest model you can run isn’t always the best choice; using a smaller model can be faster and may even give you better results.

Beelink Mini S13 Pro PC.

CPU

Celeron FCBGA1264 3.6GHz

Graphics

Integrated Intel Graphics 24EUs 1000MHz

The Beelink Mini S13 Pro desktop PC is a ultra-compact computer powered by the Intel N150 processor. Shipping with 16GB of DDR4 RAM and a 500GB SSD, this micro desktop is perfect for a variety of workloads. From running simple server programs to replacing your old PC, the Beelink S13 Pro is up to the task. 


A small local LLM isn’t suitable for everything

It will struggle as a conversation agent

Home Assistant's Assist voice assistant running on an iPhone. Credit: Adam Davidson/How-To Geek

On a whim, I tried to see if I could use either of these models as conversation agents for Assist. This would let me use natural language voice commands with Assist, instead of having to use specific phrases.

As expected, both models failed miserably at this. On my hardware, there was too much context for these smaller models to process quickly, and it would take more than 20 seconds for my lights to turn on, which just wasn’t usable.

If you’re running your local LLM without powerful hardware, it won’t be suitable for every job, as it will either be too slow or incapable of doing what you want. For some jobs, such as generating my morning briefings, however, a local LLM is a perfect way to get the results I want while maintaining my privacy.


Give a local LLM a try

If you’ve put off trying a local LLM because you didn’t think your hardware could handle it, it’s worth seeing what a small local model can do. Until I win the lottery and can afford a powerful AI rig, these small models will do just fine.



Source link

Leave a Reply

Subscribe to Our Newsletter

Get our latest articles delivered straight to your inbox. No spam, we promise.

Recent Reviews


Do you ever walk past a person on the streets exhibiting mental health issues and wonder what happened to their family? I have a brother—or at least, I used to. I worry about where he is and hope he is safe. He hasn’t taken my call since 2014.

James and his brother as young children playing together before his brother became sick. James is on the right and his brother is on the left.

James and his brother as young children playing together before his brother became sick. James is on the right and his brother is on the left.

When I was 13, I had a very bad day. I was in the back of the car, and what I remember most was the world-crushing sound violently panging off every surface: he was pounding his fists into the steering wheel, and I worried it would break apart. He was screaming at me and my mother, and I remember the web of saliva and tears hanging over his mouth. His eyes were red, and I knew this day would change everything between us. My brother was sick.

Nearly 20 years later, I still have trouble thinking about him. By the time we realized he was mentally ill, he was no longer a minor. The police brought him to a facility for the standard 72-hour hold, where he was diagnosed with paranoid delusional schizophrenia. Concluding he was not a danger to himself or others, they released him.

There was only one problem: at 18, my brother told the facility he was not related to us and that we were imposters. When they let him out, he refused to come home.

My parents sought help and even arranged for medication, but he didn’t take it. Before long, he disappeared.

My brother’s decline and disappearance had nothing to do with the common narratives about drug use or criminal behavior. He was sick. By the time my family discovered his condition, he was already 18 and legally independent from our custody.

The last time he let me visit, I asked about his bed. I remember seeing his dirty mattress on the floor beside broken glass and garbage. I also asked about the laptop my parents had gifted him just a year earlier. He needed the money, he said—and he had maxed out my parents’ credit card.

In secret from my parents, I gave him all the cash I had saved. I just wanted him to be alright.

My parents and I tried texting and calling him; there was no response except the occasional text every few weeks. But weeks turned into months.

Before long, I was graduating from high school. I begged him to come. When I looked in the bleachers, he was nowhere to be seen. I couldn’t help but wonder what I had done wrong.

The last time I heard from him was over the phone in 2014. I tried to tell him about our parents and how much we all missed him. I asked him to be my brother again, but he cut me off, saying he was never my brother. After a pause, he admitted we could be friends. Making the toughest call of my life, I told him he was my brother—and if he ever remembers that, I’ll be there, ready for him to come back.

I’m now 32 years old. I often wonder how different our lives would have been if he had been diagnosed as a minor and received appropriate care. The laws in place do not help families in my situation.

My brother has no social media, and we suspect he traded his phone several years ago. My family has hired private investigators over the years, who have also worked with local police to try to track him down.

One private investigator’s report indicated an artist befriended my brother many years ago. When my mother tried contacting the artist, they said whatever happened between them was best left in the past and declined to respond. My mom had wanted to wish my brother a happy 30th birthday.

My brother grew up in a safe, middle-class home with two parents. He had no history of drug use or criminal record. He loved collecting vintage basketball cards, eating mint chocolate chip ice cream, and listening to Motown music. To my parents, there was no smoking gun indicating he needed help before it was too late.

The next time you think about a person screaming outside on the street, picture their families. We need policies and services that allow families to locate and support their loved ones living with mental illness, and stronger protections to ensure that individuals leaving facilities can transition into stable care. Current laws, including age-based consent rules, the limits of 72-hour holds, and the lack of step-down or supported housing options, leave too many families without resources when a serious diagnosis occurs.

Governments and lawmakers need to do better for people like my brother. As someone who thinks about him every day, I can tell you the burden is too heavy to carry alone.

James Finney-Conlon is a concerned brother and mental health advocate. He can be reached at [email protected].



Source link