They usually need a lot of computer memory (RAM) to work well.

However, a few great models can perform on computers with as little as 4GB of RAM.

Why Use Local LLM Chatbots?

‘Figment’ in a Galaxy Z Fold 6 next to an 8BitDo Pro 2 controller.

Online AI chatbots are powerful tools that can seriously boost your daily efficiency.

You key in in what you want, and these LLMs create text based on your instructions.

Why might you use a local chatbot instead of a popular online option?

An old laptop with a chatbot conversation coming out of the screen.

Lucas Gouveia / How-To Geek |Vodopyanov Vyacheslav/ Shutterstock

First, it’s just fun to have a chatbot only you’re able to talk to.

Ultimately, though, it depends on how much you care about privacy, availability, and cost.

Regardless, there is information I might want to share with a chatbot that should always stay between us.

DistilBERT Homepage-1

It can get quite complicated with student loan payments, interest calculations, and the like.

You also have to consider that big data breaches are frighteningly common.

Others will need occasional internet access for updates.

Hugging Face ALBERT Homepage

Lastly, online chatbots generate their companies hundreds of millions of dollars in subscription fees.

Its close rival, Anthropic, also charges $20 monthly for its most advanced features.

Local chatbots can help mitigate such costs.

GPT-2 124 model repository homepage

Not all of them are free, though.

Some require licensing and or usage fees, like OpenAI’s GPT-3.

However, several open-source local chatbot models are free to download and work with.

GPT-Neo 125 Model Repository

These should be used strategically for easier problems.

So, you upscale to the admittedly more advanced online ones only when you absolutely must.

Lightweight means these models are designed to be highly efficient regarding memory use and processing power.

This does limit their chatbot powers for complex tasks, which other online chatbots could easily handle.

I think DistilBERT is one of, if not the most efficient, models available so far.

I highly recommend starting as a beginner with DistilBERT and ALBERT, even if you have a high-memory PC.

Which are like the Swiss Army knives of the local AI world.

GPT-2 models come in different sizes, some more suited for low-RAM PCs than others.

The124M parameter versionis the lightest.

My favorite lightweight LLM by far is theGPT-Neo 125Mbecause of its adjustable customization options.

It was developed by the respected developers at EleutherAI and is like the open-source cousin of GPT-2.

The Neo 125M is designed to balance performance and resource requirements.

This models performance is on par with GPT-2, but it’s adjusted to use memory more efficiently.

How to Get Started with Local LLM Chatbots

Running your own chatbot is easier than you think.

Using the information provided above, ensure your gear’s system can work with the chatbot you want.

Once you know this, you’ve got the option to download the right software.

Look for LLM software on websites likeHugging FaceandGitHub.

Be sure to read the instructions for each model you use to understand how they work.

Then download and chat away.

Remember to keep an eye out for any possible model software updates as well.

If your setup isn’t very powerful, you really should start with DistilBERT or ALBERT.

You will have questions.

Chances are many have already been answered in online communities or forums.

Don’t let hardware stop you from trying your hand at local LLM chatbots.

There are plenty of options that can run efficiently on low-memory systems.

Give them a try today!