Ollama
Return to LLMs
Get up and running with large language models. Run Llama 3.1, Phi 3, Mistral, Gemma 2, and other LLM models. Customize and create your own LLM models.
Ollama is a tool used to run large language models (LLMs) locally on your machine. It serves as a wrapper around the open-source `llama.cpp` library, which implements LLMs in pure C/C++ to maximize efficiency. Ollama supports various models, including the 8 billion parameter Llama 3 model, which can be run using the command `ollama run llama3`. This model requires approximately 16 GB of RAM. For machines with less RAM, smaller models like the 3.8 billion parameter `phi-3` model can be used [1].
[1] [Build a Large Language Model (From Scratch) (chapter-7) by Sebastian Raschka](https://livebook.manning.com/raschka/chapter-7)
Ollama supports a list of models available on https://ollama.com/library
Here are some example models that can be downloaded:
Model Parameters Size Download
Note
You should have at least 8 GB of RAM available to run the 7B models, 16 GB to run the 13B models, and 32 GB to run the 33B models.