I let a local LLM take control of my video doorbell—it’s probably the future of smart cameras


Some Ring doorbells can use AI features to interact with visitors when you’re not home. I ditched my Ring doorbell for a Reolink doorbell that runs fully locally, but I wondered if I could recreate a similar feature using a local LLM. I was partially successful.

What I wanted my doorbell to do

An AI-powered concierge

Ring doorbell in use by a woman and a man. Credit: Ring

The idea seemed fairly plausible. When someone rings the doorbell and Home Assistant detects that no one is home, the doorbell should speak to the caller explaining that everyone is out and asking for their name and reason for calling. It should then listen for the response, process what they say, and respond accordingly.

With the use of a cloud-based LLM, this would seem to be a realistic goal. Converting text to speech and speech to text are simple enough to do using cloud-based services. An LLM would sit in the middle, taking what the caller said as the input and generating responses to be spoken by the doorbell.

I knew that doing this with a local LLM would be more challenging. My relatively weak hardware can only run smaller models, and these might not be up to the job. I figured it was worth a try to see whether I could get it all running locally.

Reolink Wi-Fi video doorbell.

Resolution

2K

Power Source

Battery

Reolink’s battery-powered Wi-Fi video doorbell is a great way to know who’s outside. With a 2K resolution and a 150°x150° head-to-toe view, this video doorbell can be powered either over battery or wired, depending on your existing setup.


How I set it up

TTS out, Whisper in, Ollama in the middle

There were three main components that I needed to make this work. I needed a way to transform text to speech (TTS) so that my doorbell could speak aloud to the caller. I needed a way to transform speech to text (STT) so that whatever the caller said could be converted into written text to pass to the LLM. And I needed a way to run a local LLM that would be the brains of the whole operation.

Thankfully, Home Assistant has some great options for each of these components. Piper is a local TTS engine that can turn written text into spoken audio that I can play through my doorbell. It runs entirely locally and is lightweight enough that you can run it on a Raspberry Pi 4.


A snarky notification from Home Assistant describing someone at the door on an iPhone.


How I Use Home Assistant to Describe Who’s At the Door Using AI

Get AI-generated descriptions of anyone your video doorbell detects.

Whisper provides the equivalent local STT component. It can take the audio recorded by my doorbell when the caller is speaking and convert it into text that I can pass to the local LLM. Once again, it runs entirely locally, which was my aim for this project.

The final piece of the puzzle is Ollama. This is a tool that allows you to run local large language models on your own hardware. There’s a Home Assistant integration that you can use to connect Ollama to Home Assistant.

The bottleneck is the capability of the LLM model that you run. Weaker hardware can only run smaller, less capable models, and the larger the model you try to run, the slower the responses are likely to be. I had to use a fairly small model to ensure that it didn’t take too long to generate responses.

Reality didn’t match my hopes

The concept is fine, the execution isn’t

A Reolink video doorbell in the rain. Credit: Reolink

It took me some time to get everything set up. As always with Home Assistant, other people had done most of the hard work; there was a useful GitHub Gist explaining how to play audio and TTS through my Reolink doorbell, which came in very handy.

I had some issues with the audio capture starting while the spoken greeting from the doorbell was still playing, which messed things up, but eventually figured out how to work around it.

The first parts of my idea worked well. When the doorbell was pressed, the LLM would generate a spoken greeting which would play through the doorbell speaker. It would explain that everyone was out and ask the caller for their name and the purpose of their call.

The doorbell would then record their spoken response and STT would turn it into text. So far, so good.

The problem was that trying to have a two-way conversation with the AI-powered doorbell just didn’t work. The small LLM would get confused and start talking nonsense, and the responses would take too long to come through.

It seems likely that the concept would work much better with a powerful enough LLM running the show. Until I win the lottery, however, I’m stuck with what I’ve got.

I built a workable alternative

It’s actually a pretty solid setup

A notification forwarding a message left on a video doorbell.

Since the main sticking point was trying to have a conversation with the caller, I simply cut out that part of the process. Instead, when the caller gives their name and reason for calling, the STT turns this into text, and that text is then sent as a notification to my phone. The doorbell then says that it will pass on the message and ends the conversation.

It means that whenever someone rings the doorbell when we’re out, I get a notification telling me who it was and why they were calling. It works reasonably well most of the time, with the occasional slightly hilarious notification appearing when things go wrong. For the most part, however, it’s a genuinely useful feature.


This is the direction the world is going in

The trend now is for AI in all the things, and it doesn’t look like slowing down any time soon. While Ring’s AI-powered concierge is useful, the company doesn’t have the best reputation for privacy. The good news is that it’s possible to recreate at least parts of these features completely locally with a little effort.



Source link

Leave a Reply

Subscribe to Our Newsletter

Get our latest articles delivered straight to your inbox. No spam, we promise.

Recent Reviews



In short: Accel has raised $5 billion in new capital, comprising a $4 billion Leaders Fund V and a $650 million sidecar, targeting 20-25 late-stage AI investments at an average cheque size of $200 million. The raise follows standout returns from its Anthropic stake (invested at $183B, now valued near $800B) and Cursor (backed at $9.9B, now reportedly around $50B), and lands in a Q1 2026 venture market that deployed a record $297 billion.

Accel, the venture capital firm behind early bets on Facebook, Slack, and more recently Anthropic and Cursor, has raised $5 billion in new capital aimed squarely at AI. The raise, reported by Bloomberg, comprises $4 billion for its fifth Leaders Fund and a $650 million sidecar vehicle, positioning the firm to write average cheques of around $200 million into late-stage AI companies globally.

The fund lands in a venture capital market that has lost any pretence of restraint. Q1 2026 saw $297 billion flow into startups worldwide, 2.5 times the total from Q4 2025 and the most venture funding ever recorded in a three-month period. Andreessen Horowitz has raised $15 billion. Thrive Capital has closed more than $10 billion. Founders Fund is finishing a $6 billion raise. Accel’s $5 billion is substantial but not exceptional in a market where the biggest funds are measured in the tens of billions.

The portfolio that made the pitch

What distinguishes Accel’s fundraise is the portfolio it can point to. The firm invested in Anthropic during its Series G at a $183 billion valuation. Anthropic has since closed a round at $380 billion and is now attracting offers at roughly $800 billion, meaning Accel’s stake has more than quadrupled in value in a matter of months. Anthropic’s annualised revenue has hit $30 billion, a trajectory that no company in history has matched.

The firm’s bet on Cursor has been similarly well-timed. Accel backed the AI code editor in June 2025 at a $9.9 billion valuation. By November, Cursor had raised again at $29.3 billion. By March 2026, the company was reportedly in discussions at a valuation of around $50 billion. For a developer tool that barely existed two years ago, the appreciation is extraordinary.

Accel’s broader AI portfolio extends beyond these two headline positions. The firm has backed Vercel, the frontend deployment platform; n8n, an AI-powered automation tool; Recraft, a professional design platform; and Code Metal, which builds AI development tools for hardware and defence applications. In March 2026, Accel launched an Atoms AI programme in partnership with Google’s AI Futures Fund, selecting five early-stage companies from what it described as a global applicant pool focused on “white space” opportunities in enterprise AI.

The Leaders Fund model

Accel’s Leaders Fund series is designed for later-stage investments, the kind of large cheques that growth-stage AI companies now require. With an average investment size of $200 million and a target of 20 to 25 deals from the new $4 billion fund, the strategy is concentrated: a small number of high-conviction bets on companies that have already demonstrated product-market fit and are scaling revenue.

This is a different game from traditional venture capital. At $200 million per cheque, Accel is competing less with seed and Series A firms and more with the mega-funds, sovereign wealth funds, and corporate investors that have flooded into late-stage AI. The firm’s argument is that its early-stage relationships and technical evaluation capabilities give it an edge in identifying which companies deserve capital at scale, and in securing allocations in rounds that are massively oversubscribed.

Founded in 1983 by Arthur Patterson and Jim Swartz, Accel built its reputation on what the founders called the “prepared mind” approach, a philosophy of deep sector research before investments materialise. The firm’s most famous prepared-mind bet was its 2005 investment of $12.7 million for 10% of Facebook, a stake worth $6.6 billion at the company’s IPO seven years later. The question now is whether Accel’s AI bets will produce returns of comparable magnitude.

What the market is pricing

The sheer volume of capital flowing into AI venture funds reflects a market consensus that artificial intelligence will be the dominant technology platform of the next decade. The numbers are difficult to overstate. OpenAI raised $120 billion in 2026. Anthropic has raised more than $50 billion. xAI closed $20 billion. Waymo secured $16 billion. These are not venture-scale numbers; they are infrastructure-scale capital deployments that would have been unthinkable outside of telecommunications or energy a decade ago.

For limited partners, the investors who commit capital to venture funds, the logic is straightforward: the returns from AI’s winners will be so large that even paying premium valuations will generate exceptional multiples. Accel’s Anthropic position, where a single investment has appreciated several times over in months, is exactly the kind of outcome that makes LPs willing to commit $5 billion to a single firm’s next fund.

The risk is equally visible. Venture capital is a cyclical business, and the current fundraising boom has the characteristics of a cycle peak: record fund sizes, compressed deployment timelines, and a concentration of capital in a single sector. The last time venture capital raised this aggressively, during the 2021 ZIRP era, many of those investments were marked down significantly within two years. AI’s commercial traction is far stronger than the crypto and fintech bets that defined that earlier cycle, but the valuations being paid today leave little margin for error.

The concentration question

Accel’s fund also highlights a structural shift in venture capital. The industry is bifurcating into a small number of mega-firms that can write cheques of $100 million or more and a long tail of smaller funds that compete for earlier-stage deals. The middle ground, the traditional Series B and C investors, is being squeezed by mega-funds moving downstream and by AI companies that skip traditional funding stages entirely, going from seed round to billion-dollar valuations in 18 months.

For a firm like Accel, which operates across offices in Palo Alto, San Francisco, London, and India, the $5 billion raise is a bet that it can maintain its position in the top tier as fund sizes inflate and competition for the best deals intensifies. Its portfolio of 1,199 companies, 107 unicorns, and 46 IPOs provides a track record. But in a market where Anthropic alone could generate returns that justify an entire fund, the temptation to concentrate bets on a handful of AI winners is strong, and the consequences of getting those bets wrong are correspondingly severe.

The broader picture is that AI venture capital has entered a phase where the funds themselves are becoming as large as the companies they once backed. Accel’s $5 billion raise would have made it one of the most valuable startups in Europe just a few years ago. Now it is table stakes for a firm that wants to participate meaningfully in the rounds that matter. Whether this represents rational capital allocation or the peak of a cycle that will eventually correct is the question that every LP writing a cheque today is, implicitly or explicitly, answering in the affirmative.



Source link