Here is the guide for using Ollama models in 16x Prompt.
- Go to Ollama
 - Follow the instructions to install Ollama on your machine
 - Once installed, run the following command to download the models, for example 
deepseek-r1:14b: 
ollama pull deepseek-r1:14bNote: Running
deepseek-r1:14brequires at least 16GB of RAM on a MacBook Pro. You can see the list of available models from Ollama Models.
- Ollama defaults to 2048 context window size, which is too small for most coding tasks. You can adjust the context window size to 8192 by running the following command:
 
ollama run deepseek-r1:14b
/set parameter num_ctx 8192
/save deepseek-r1:14b-8k- This will create a new model called 
deepseek-r1:14b-8kwith the context window size of 8192. - Serve the model by running the following command:
 
ollama serve- Open 16x Prompt and click on the API model settings button in the bottom right corner.
 - Go to "3rd Party API / Ollama Configuration" tab.
 - Enter 
http://127.0.0.1:11434/v1into the "API Endpoint / Base URL" field. - Enter model name you want to use, for example 
deepseek-r1:14b-8kinto the "Model ID" field. - Enter 
ollama(or anything else) into the "API Type" field. - Click on "Close".
 
You should now be able to use models from Ollama in 16x Prompt, and see the logs from Ollama in the console.