This is the straightforward version: install Ollama, download a model, run it, and connect to it from your browser or another app.
Ollama is one of the easiest ways to run large language models on your own machine. It gives you a local server and a simple command line interface, which is exactly what most people need to get started.
Go to the Ollama website and install it for your system. On Linux, that usually means using their install command. On Windows or Mac, use the normal installer.
Typical Linux install:
curl -fsSL https://ollama.com/install.sh | sh
After that, check that it works:
ollama --version
Once Ollama is installed, pull down a model. Start with something reasonable instead of grabbing the biggest thing you can find.
Example:
ollama run qwen3.5:2b
The first time you run that, Ollama downloads the model and then opens an interactive chat in your terminal.
You can also list what you have later:
ollama list
Ollama works as a local server. On many systems, it starts automatically after install. If not, you can start it yourself.
ollama serve
By default, it usually listens on:
http://127.0.0.1:11434
That means apps on the same machine can talk to it through that address.
To test it from terminal, you can still do something simple like:
ollama run qwen3.5:2b
Or from another tool or script, you point it at the local Ollama API.
A lot of frustration comes from trying to do too much at once. The better path is: install, test, confirm, then expand.
Add a normal link from your homepage to this file, something like:
<a href="ollama.html">Local Ollama Setup</a>
If you rename this file, just update the link to match.
This page is meant to feel personal and useful, not corporate. Clear enough to help, casual enough to still sound human.