Raul CariniFull Stack Developer

Running LLMs Locally

February 6, 2025 (4 days ago)

The field of Artificial Intelligence (AI) is experiencing a revolution, fueled by the rise of Large Language Models (LLMs). These powerful models can generate creative text, translate languages, answer complex questions, and even write different kinds of creative content.

Initially, accessing these capabilities required relying on cloud-based services. However, a significant shift is underway: the move towards running LLMs locally. This is where tools like Ollama and LM Studio become essential.

Both empower users to run LLMs on their own computers, but they cater to slightly different audiences and use cases. This new approach opens many doors for users who want more control and security for their AI interactions.

If you're interested in benchmarking the performance of these local LLMs, check out my project Snappy - LLMs Speed Test, a web application designed to measure and compare the speed of different local language models.

The Dawn of Accessible AI: Why Local Matters

For a long time, the cutting edge of AI was locked behind the doors of massive data centers, requiring substantial computing power and, often, significant financial investment. The emergence of more efficient models and tools like Ollama and LM Studio is democratizing AI. Now, individuals and small businesses can run sophisticated LLMs on standard, relatively inexpensive hardware.

This "new age of AI" is characterized by:

What is Ollama?

Ollama is a free, open-source framework designed for running LLMs locally. It emphasizes simplicity and ease of use, making it accessible even to users without extensive technical expertise. Ollama supports macOS, Linux, and Windows.

Key Features and Uses of Ollama:

What is LM Studio?

LM Studio is a desktop application (available for macOS, Windows, and Linux) that provides a user-friendly interface for discovering, downloading, and running local LLMs. It caters to both beginners and experienced users.

Key Features and Uses of LM Studio:

The Rise of Powerful, Efficient Models

The feasibility of running LLMs locally has been significantly boosted by the development of smaller, more efficient models. Some notable examples include:

Which One Should You Use?

Both Ollama and LM Studio are powerful tools that represent a significant step towards democratizing access to advanced AI. The choice between them depends on your technical expertise, specific project requirements, and the degree of control and customization you desire. The increasing availability of efficient, powerful models like Llama 3, Qwen2, and DeepSeek Coder further solidifies the viability of local LLM deployment, opening up exciting possibilities for developers and users alike.