The best local AI model for Home Assistant isn’t always the biggest one


Using a Large Language Model (LLM) with Home Assistant has a number of benefits. It can add natural language understanding, power your voice assistant, and even analyze images. A local LLM can help maintain privacy, but you don’t always need to use the largest model.

Using a local LLM with Home Assistant

Keep your data private

LM Studio setup guide asking to download first local model

There are plenty of ways that an LLM can make Home Assistant more powerful. One popular usage is to hook up a cloud-based model from a company such as OpenAI and use it as a conversation assistant for the Assist voice assistant. This allows you to use natural language commands to control your smart home, rather than having to remember the specific phrases that will turn your lights on and off.

The trouble with using a cloud-based LLM is that data about your smart home has to be sent to the cloud to be processed. It means that information about your smart home ends up on third-party servers. Home Assistant was designed to help maintain your privacy, so sharing information about your smart home with AI companies goes against this core principle.

One solution is to use a local LLM. You can run models on your own hardware that can perform some of the same tasks that cloud-based LLMs can do. How quickly or accurately a local LLM can perform these tasks depends both on the hardware you run it on and the models that you use.

Why the biggest model isn’t always the best

Finding the sweet spot

LLMs often come in different sizes. You might see versions of the same model with values such as 4B, 9B, 70B. These refer to the number of parameters the model has; a 70B model has 70 billion parameters, for example. These larger models often have more capacity for knowledge and reasoning.

The flip side is that the more parameters a model has, the more VRAM is needed to store those parameters. Some 70B models, for example, might need more than 100 GB of VRAM to run. That’s beyond the reach of even high-end consumer GPUs unless you’re running a multi-GPU stack, and if you don’t have enough VRAM, the model won’t run at all or will slow to a crawl.

The challenge is finding a model that’s small enough to run on your hardware, but powerful enough to handle the jobs that you want it to do. There are some useful tools, such as llmfit, that can tell you which models are best suited to run on your hardware.

The good news is that, as the technology has developed, new smaller models have appeared that can outperform the very large models from just a few years ago. You no longer need to have insane amounts of VRAM to get decent performance from a local LLM.

Small local LLMs can run on basic hardware

You don’t need an expensive GPU

A Raspberry Pi in a case lying on top of a Beelink Mini S12 Pro mini PC. Credit: Adam Davidson / How-To Geek

If you don’t have a dedicated GPU, it’s not the end of the world. There are some smaller models that are capable of running on a CPU alone without needing to pass anything off to a GPU. These models can use the system RAM in your PC rather than fitting everything into VRAM. While the performance can’t match larger models running with a GPU, they can still do a job.

I wanted to use a local LLM on my Beelink Mini PC that has 16 GB of RAM and no dedicated GPU. My main purpose was to take a list of events from my calendar and transform it into the text for a spoken morning briefing. I’d seen a lot of people saying they’d found the Qwen 3.5 4B model to be a good sweet spot, so I decided to give it a try.

Using this model, I was able to generate the briefing, although it took around 13 seconds to generate. The text was fine, but it wasn’t particularly inspiring.

In an effort to speed things up, I tried a smaller model, Llama 3.2 3B, which uses fewer parameters. You might expect this model to produce a worse result, but it produced a much more natural-sounding output, and did it in under 6 seconds, less than half the time of the other model.

It seems that size isn’t everything. The largest model you can run isn’t always the best choice; using a smaller model can be faster and may even give you better results.

Beelink Mini S13 Pro PC.

CPU

Celeron FCBGA1264 3.6GHz

Graphics

Integrated Intel Graphics 24EUs 1000MHz

The Beelink Mini S13 Pro desktop PC is a ultra-compact computer powered by the Intel N150 processor. Shipping with 16GB of DDR4 RAM and a 500GB SSD, this micro desktop is perfect for a variety of workloads. From running simple server programs to replacing your old PC, the Beelink S13 Pro is up to the task. 


A small local LLM isn’t suitable for everything

It will struggle as a conversation agent

Home Assistant's Assist voice assistant running on an iPhone. Credit: Adam Davidson/How-To Geek

On a whim, I tried to see if I could use either of these models as conversation agents for Assist. This would let me use natural language voice commands with Assist, instead of having to use specific phrases.

As expected, both models failed miserably at this. On my hardware, there was too much context for these smaller models to process quickly, and it would take more than 20 seconds for my lights to turn on, which just wasn’t usable.

If you’re running your local LLM without powerful hardware, it won’t be suitable for every job, as it will either be too slow or incapable of doing what you want. For some jobs, such as generating my morning briefings, however, a local LLM is a perfect way to get the results I want while maintaining my privacy.


Give a local LLM a try

If you’ve put off trying a local LLM because you didn’t think your hardware could handle it, it’s worth seeing what a small local model can do. Until I win the lottery and can afford a powerful AI rig, these small models will do just fine.



Source link

Leave a Reply

Subscribe to Our Newsletter

Get our latest articles delivered straight to your inbox. No spam, we promise.

Recent Reviews


Google Maps has a long list of hidden (and sometimes, just underrated) features that help you navigate seamlessly. But I was not a big fan of using Google Maps for walking: that is, until I started using the right set of features that helped me navigate better.

Add layers to your map

See more information on the screen

Layers are an incredibly useful yet underrated feature that can be utilized for all modes of transport. These help add more details to your map beyond the default view, so you can plan your journey better.

To use layers, open your Google Maps app (Android, iPhone). Tap the layer icon on the upper right side (under your profile picture and nearby attractions options). You can switch your map type from default to satellite or terrain, and overlay your map with details, such as traffic, transit, biking, street view (perfect for walking), and 3D (Android)/raised buildings (iPhone) (for buildings). To turn off map details, go back to Layers and tap again on the details you want to disable.

In particular, adding a street view and 3D/raised buildings layer can help you gauge the terrain and get more information about the landscape, so you can avoid tricky paths and discover shortcuts.

Set up Live View

Just hold up your phone

A feature that can help you set out on walks with good navigation is Google Maps’ Live View. This lets you use augmented reality (AR) technology to see real-time navigation: beyond the directions you see on your map, you are able to see directions in your live view through your camera, overlaying instructions with your real view. This feature is very useful for travel and new areas, since it gives you navigational insights for walking that go beyond a 2D map.

To use Live View, search for a location on Google Maps, then tap “Directions.” Once the route appears, tap “Walk,” then tap “Live View” in the navigation options. You will be prompted to point your camera at things like buildings, stores, and signs around you, so Google Maps can analyze your surroundings and give you accurate directions.

Download maps offline

Google Maps without an internet connection

Whether you’re on a hiking trip in a low-connectivity area or want offline maps for your favorite walking destinations, having specific map routes downloaded can be a great help. Google Maps lets you download maps to your device while you’re connected to Wi-Fi or mobile data, and use them when your device is offline.

For Android, open Google Maps and search for a specific place or location. In the placesheet, swipe right, then tap More > Download offline map > Download. For iPhone, search for a location on Google Maps, then, at the bottom of your screen, tap the name or address of the place. Tap More > Download offline map > Download.

After you download an area, use Google Maps as you normally would. If you go offline, your offline maps will guide you to your destination as long as the entire route is within the offline map.

Enable Detailed Voice Guidance

Get better instructions

Voice guidance is a basic yet powerful navigation tool that can come in handy during walks in unfamiliar locations and can be used to ensure your journey is on the right path. To ensure guidance audio is enabled, go to your Google Maps profile (upper right corner), then tap Settings > Navigation > Sound and Voice. Here, tap “Unmute” on “Guidance Audio.”

Apart from this, you can also use Google Assistant to help you along your journey, asking questions about your destination, nearby sights, detours, additional stops, etc. To use this feature on iPhone, map a walking route to a destination, then tap the mic icon in the upper-right corner. For Android, you can also say “Hey Google” after mapping your destination to activate the assistant.

Voice guidance is handy for both new and old places, like when you’re running errands and need to navigate hands-free.

Add multiple stops

Keep your trip going

If you walk regularly to run errands, Google Maps has a simple yet effective feature that can help you plan your route in a better way. With Maps’ multiple stop feature, you can add several stops between your current and final destination to minimize any wasted time and unnecessary detours.

To add multiple stops on Google Maps, search for a destination, then tap “Directions.” Select the walking option, then click the three dots on top (next to “Your Location”), and tap “Edit Stops.” You can now add a stop by searching for it and tapping “Add Stop,” and swap the stops at your convenience. Repeat this process by tapping “Add Stops” until your route is complete, then tap “Start” to begin your journey.

You can add up to ten stops in a single route on both mobile and desktop, and use the journey for multiple modes (walking, driving, and cycling) except public transport and flights. I find this Google Maps feature to be an essential tool for travel to walkable cities, especially when I’m planning a route I am unfamiliar with.


More to discover

A new feature to keep an eye out for, especially if you use Google Maps for walking and cycling, is Google’s Gemini boost, which will allow you to navigate hands-free and get real-time information about your journey. This feature has been rolling out for both Android and iOS users.



Source link