Newtuple
Back to Blog
5 minute setupAI + Data

Host LLMs Locally: Ollama Installation & Setup Guide

Step-by-step Ollama install to host large-language models on your machine. Get up and running in minutes with examples.

Shikhar JhaMay 22, 20252 min readUpdated November 28, 2025
Host LLMs Locally: Ollama Installation & Setup Guide

Open source Large Language models are increasingly becoming viable alternatives to the closed source models that are available today. The first step towards understanding whether a model is a fit for your use case is to quickly get it up and running on your laptop or your machine. This article goes into a 5 minute setup guide for Ollama.

Ollama has made it easiest by far by providing a simple lightweight, extensible framework for building and running language models on the local machine. It provides a simple API for creating, running, and managing models, as well as a library of pre-built models that can be easily used in a variety of applications.

With ollama you can run Llama 3.3, DeepSeek-R1, Qwen 3, Mistral, Gemma 3, and other models, locally.

  1. Steps to install ollama in your machine:

Whether you are using mac , linux or windows operating system ollama supports all .

The below steps shown are done in macOS , but similar approach can be used in other operating system.

Visit https://ollama.com/.

ollama website ollama website

  • Click on the download button a zip file ollama-darwin.zip will be download shortly.

  • Go to the downloads folder (or where the zip file is saved) , unzip the file you will see the ollama application extracted.

  • Go ahead and move it to Applications

After successfully moving the ollama application to Application click on ollama. You will be prompted to install ollama.

After download is complete!After download is complete!

  • Make sure you have admin rights to you system as you will be prompted to validation that you are admin .

After the above steps are successful, you can now run any open source model in your machine.

How to run a model !How to run a model !

Lets try it out!!

  • Open a terminal and type.
ollama run <model name>

After successful run (might defer based on model):

Downloading Qwen3!!Downloading Qwen3!!

Congratulations !! you have an AI running in your machine!

Example:

Running Qwen3!Running Qwen3!

Tip !

\# use below command to exit 
/bye

References:

Github : https://github.com/ollama/ollama

Ollama : https://ollama.com/

Stay in the loop

Get new posts, product updates, and research notes once a week.

By subscribing you agree to receive updates from Newtuple. You can unsubscribe anytime.

Comments

Ready to build production AI?

Talk to our team about AI agents, data platforms, and GenAI accelerators.

Get in Touch