Running large language models locally

Model Shop

Article from Issue 287/2024
Author(s):

Ollama and Open WebUI let you join the AI revolution without relying on the cloud.

Large language models (LLMs) such as the ones used by OpenAI's [1] ChatGPT [2] are too resource intensive to run locally on your own computer. That's why they're deployed as online services that you pay for. However, since ChatGPT's release, some significant advancements have occurred around smaller LLMs. Many of these smaller LLMs are open source or have a liberal license (see the "Licenses" box). You can run them on your own computer without having to send your input to a cloud server and without having to pay a fee to an online service.

Because these LLMs are computationally intensive and need a lot of RAM, running them on your CPU can be slow. For optimal performance, you need a GPU – GPUs have many parallel compute cores and a lot of dedicated RAM. An NVIDIA or AMD GPU with 8GB RAM or more is recommended.

In addition to the hardware and the models, you also need software that enables you to run the models. One popular package is Ollama [3], named for Meta AI's large language model Llama [4]. Ollama is a command-line application that runs on Linux, macOS, and Windows, and you can also run it as a server that other software connects to.

[...]

Use Express-Checkout link below to read the full article (PDF).

Buy this article as PDF

Express-Checkout as PDF
Price $2.95
(incl. VAT)

Buy Linux Magazine

SINGLE ISSUES
 
SUBSCRIPTIONS
 
TABLET & SMARTPHONE APPS
Get it on Google Play

US / Canada

Get it on Google Play

UK / Australia

Related content

  • Write Code with AI

    Artificial intelligence is increasingly supporting programmers in their daily work. How effective are these tools? What are the dangers? And how can you benefit from AI-assisted development today?

  • Linux Foundation Adopts OMI to Foster Ethical LLMs

    The Open Model Initiative hopes to create community LLMs that rival proprietary models but avoid restrictive licensing that limits usage.

  • RHEL AI

    Jesper Rooth discusses Red Hat's new AI platform solution as well as their future AI plans.

  • Bash PaLM Shell

    The PaLMShell.bash script lets you connect to the Google Pathways API Large Language Model (PaLM) from the command line.

  • News

    In the news: New KDE Slimbook Plasma Available for Preorder; Rhino Linux Announces Latest "Quick Update"; Plasma Desktop Will Soon Ask for Donations: Linux Market Share Hits New High; LibreOffice 24.8 Delivers New Features; Deepin 23 Offers Wayland Support and New AI Tool; and Linux Foundation Adopts OMI to Foster Ethical LLMs.

comments powered by Disqus
Subscribe to our Linux Newsletters
Find Linux and Open Source Jobs
Subscribe to our ADMIN Newsletters

Support Our Work

Linux Magazine content is made possible with support from readers like you. Please consider contributing when you’ve found an article to be beneficial.

Learn More

News