# Introduction [Ollama](https://github.com/ollama/ollama) allows you to run open-source large language models, such as Llama 3, locally. Ollama bundles model weights, configuration, and data into a single package, defined by a Modelfile. Ollama is a lightweight, extensible framework for building and running language models on the local machine. It provides a simple API for creating, running, and managing models, as well as a library of pre-built models that can be easily used in a variety of applications. It's the best choice to deploy large language models on AIPC locally. ## Get Started ### Setup Follow [these instructions](https://github.com/ollama/ollama) to set up and run a local Ollama instance. - Download and install Ollama onto the available supported platforms (including Windows) - Fetch available LLM model via `ollama pull `. View a list of available models via the model library and pull to use locally with the command `ollama pull llama3` - This will download the default tagged version of the model. Typically, the default points to the latest, smallest sized-parameter model. Note: Special settings are necessary to pull models behind the proxy. - Step1: Modify the ollama service configure file. ```bash sudo vim /etc/systemd/system/ollama.service ``` Add your proxy to the above configure file. ```markdown [Service] Environment="http_proxy=${your_proxy}" Environment="https_proxy=${your_proxy}" ``` - Step2: Restart the ollama service. ```bash sudo systemctl daemon-reload sudo systemctl restart ollama ``` ### Usage Here are a few ways to interact with pulled local models: #### In the terminal All of your local models are automatically served on localhost:11434. Run `ollama run ` to start interacting via the command line directly. #### API access Send an application/json request to the API endpoint of Ollama to interact. ```bash curl --noproxy "*" http://localhost:11434/api/generate -d '{ "model": "llama3", "prompt":"Why is the sky blue?" }' ```