Releases: ollama/ollama
v0.1.23
New vision models
The LLaVA model family on Ollama has been updated to version 1.6, and now includes a new 34b
version:
ollama run llava
A new 7B LLaVA model based on mistral.ollama run llava:13b
13B LLaVA modelollama run llava:34b
34B LLaVA model – one of the most powerful open-source vision models available
These new models share new improvements:
- More permissive licenses: LLaVA 1.6 models are distributed via the Apache 2.0 license or the LLaMA 2 Community License.
- Higher image resolution: support for up to 4x more pixels, allowing the model to grasp more details.
- Improved text recognition and reasoning capabilities: these models are trained on additional document, chart and diagram data sets.
keep_alive
parameter: control how long models stay loaded
When making API requests, the new keep_alive
parameter can be used to control how long a model stays loaded in memory:
curl http://localhost:11434/api/generate -d '{
"model": "mistral",
"prompt": "Why is the sky blue?",
"keep_alive": "30s"
}'
- If set to a positive duration (e.g.
20m
,1hr
or30
), the model will stay loaded for the provided duration - If set to a negative duration (e.g.
-1
), the model will stay loaded indefinitely - If set to
0
, the model will be unloaded immediately once finished - If not set, the model will stay loaded for 5 minutes by default
Support for more Nvidia GPUs
- GeForce GTX
TITAN X
980 Ti
980
970
960
950
750 Ti
750
- GeForce GTX
980M
970M
965M
960M
950M
860M
850M
- GeForce
940M
930M
910M
840M
830M
- Quadro
M6000
M5500M
M5000
M2200
M1200
M620
M520
- Tesla
M60
M40
- NVS
810
What's Changed
- New
keep_alive
API parameter to control how long models stay loaded - Image paths can now be provided to
ollama run
when running multimodal models - Fixed issue where downloading models via
ollama pull
would slow down to 99% - Fixed error when running Ollama with Nvidia GPUs and CPUs without AVX instructions
- Support for additional Nvidia GPUs (compute capability 5)
- Fixed issue where system prompt would be repeated in subsequent messages
ollama serve
will now print prompt whenOLLAMA_DEBUG=1
is set- Fixed issue where exceeding context size would cause erroneous responses in
ollama run
and the/api/chat
API ollama run
will now allow sending messages without images to multimodal models
New Contributors
- @jaglinux made their first contribution in #2224
- @textspur made their first contribution in #2252
- @rjmacarthy made their first contribution in #1950
- @hugo53 made their first contribution in #1957
- @RussellCanfield made their first contribution in #2313
Full Changelog: v0.1.22...v0.1.23
v0.1.22
New models
- Stable LM 2: A state-of-the-art 1.6B small language model.
What's Changed
- Fixed issue with Nvidia GPU detection that would cause Ollama to error instead of falling back to CPU
- Fixed issue where AMD integrated GPUs caused an error
Full Changelog: v0.1.21...v0.1.22
v0.1.21
New models
- Qwen: Qwen is a series of large language models by Alibaba Cloud spanning from 1.8B to 72B parameters.
- DuckDB-NSQL: A text-to-sql LLM for DuckDB
- Stable Code: A new code completion model on par with Code Llama 7B and similar models.
- Nous Hermes 2 Mixtral: The Nous Hermes 2 model from Nous Research, now trained over Mixtral.
Saving and loading models and messages
Models can now be saved and loaded with /save <model>
and /load <model>
when using ollama run
. This will save or load conversations and any model changes with /set parameter
, /set system
and more as a new model with the provided name.
MESSAGE
modelfile command
Messages can now be specified in a Modelfile
ahead of time using the MESSAGE
command:
# example Modelfile
FROM llama2
SYSTEM You are a friendly assistant that only answers with 'yes' or 'no'
MESSAGE user Is Toronto in Canada?
MESSAGE assistant yes
MESSAGE user Is Sacramento in Canada?
MESSAGE assistant no
MESSAGE user Is Ontario in Canada?
MESSAGE assistant yes
After creating this model, running it will restore the message history. This is useful for techniques such as Chain-Of-Thought prompting
ollama create -f Modelfile yesno
ollama run yesno
>>> Is Toronto in Canada?
yes
>>> Is Sacramento in Canada?
no
>>> Is Ontario in Canada?
yes
>>> Is Havana in Canada?
no
Python and Javascript libraries
The first versions of the Python and JavaScript libraries for Ollama are now available.
Intel & AMD CPU improvements
Ollama now supports CPUs without AVX. This means Ollama will now run on older CPUs and in environments (such as virtual machines, Rosetta, GitHub actions) that don't provide support for AVX instructions. For newer CPUs that support AVX2, Ollama will receive a small performance boost, running models about 10% faster.
What's Changed
- Support for a much broader set of CPUs, including CPUs without AVX instruction set support
- If a GPU detection error is hit when attempting to run a model, Ollama will fallback to CPU
- Fixed issue where generating responses with the same prompt would hang after around 20 requests
- New
MESSAGE
Modelfile command to set the conversation history when building a model - Ollama will now use AVX2 for faster performance if available
- Improved detection of Nvidia GPUs, especially in WSL
- Fixed issue where models with LoRA layers may not load
- Fixed incorrect error that would occur when retrying network connections in
ollama pull
andollama push
- Fixed issue where
/show parameter
would round decimal numbers - Fixed issue where upon hitting the context window limit, requests would hang
New Contributors
- @fpreiss made their first contribution in #1921
- @eavanvalkenburg made their first contribution in #1931
- @0atman made their first contribution in #1924
- @sachinsachdeva made their first contribution in #2021
- @Arrendy made their first contribution in #2016
- @purificant made their first contribution in #1958
- @lainedfles made their first contribution in #1999
Full Changelog: v0.1.20...v0.1.21
v0.1.20
New models
- MegaDolphin: A new 120B version of the Dolphin model.
- OpenChat: Updated to the latest version
3.5-0106
. - Dolphin Mistral: Updated to the latest DPO Laser version, which achieves higher scores with more robust outputs.
What's Changed
- Fixed additional cases where Ollama would fail with
out of memory
CUDA errors - Multi-GPU machines will now correctly allocate memory across all GPUs
- Fixed issue where Nvidia GPUs would not be detected by Ollama
Full Changelog: v0.1.19...v0.1.20
v0.1.19
This release focuses on performance and fixing a number issues and crashes relating to memory allocation.
New Models
- LLaMa-Pro: An expansion of LLaMa by Tencent to an 8B that specializes in language, programming and mathematics.
What's Changed
- Fixed "out of memory" errors when running models such as
llama2
,mixtral
orllama2:13b
with limited GPU memory - Fixed CUDA errors when running on older GPUs that aren't yet supported
- Increasing context size with
num_ctx
will now work (up to a model's supported context window).
To use a 32K context window with Mistral:
# ollama run
/set parameter num_ctx 32678
# api
curl http://localhost:11434/api/generate -d '{
"model": "mistral",
"prompt": "Why is the sky blue?",
"options": {"num_ctx": 32678}
}'
- Larger models such as
mixtral
can now be run on Macs with less memory - Fixed an issue where pressing up or down arrow keys would cause the wrong prompt to show in
ollama run
- Fixed performance issues on Intel Macs
- Fixed an error that would occur with old Nvidia GPUs
OLLAMA_ORIGINS
now supports browser extension URLs- Ollama will now offload more processing to the GPU where possible
New Contributors
- @sublimator made their first contribution in #1797
- @gbaptista made their first contribution in #1830
Full Changelog: v0.1.18...v0.1.19
v0.1.18
New models
- TinyLlama: a compact 1.1B Llama model on 3 trillion tokens
- OpenHermes 2: A 7B model, fine-tuned on Mistral with strong multi-turn chat skills and system prompt capabilities.
- WizardCoder 33B: a new 33B state of the art code generation model:
ollama run wizardcoder:33b
- Dolphin Phi: a 2.7B uncensored model, based on the Phi language model by Microsoft Research
What's Changed
- Added
/? shortcuts
help command toollama run
to list keyboard shortcuts - Improved performance when sending follow up messages in
ollama run
or via the API. - Fixed issues where certain 7B models would error on GPUs with 4GB of memory or less
- Fixed issue where Llava model prompts couldn't start with a file path
- Fixed issue where model would not be correctly reloaded if options or parameters changed between requests
- Ollama will now automatically pull new models when running older
ggml
format models. If using customggml
format models in aModelfile
, please import GGUF models instead.
New Contributors
- @pgibler made their first contribution in #1595
- @rxlabz made their first contribution in #1564
- @K0IN made their first contribution in #1642
- @Icelain made their first contribution in #1708
- @kghandour made their first contribution in #1747
- @shuhblam made their first contribution in #1766
Full Changelog: v0.1.17...v0.1.18
v0.1.17
Phi-2
This release adds support for the Phi-2 model by Microsoft.
ollama run phi
Phi-2 is a new, powerful 2.7B model with strong reasoning and language understanding capabilities comparable to larger, 13B models. Given its small size, it will run effectively on a wide set of hardware configurations.
Example prompt
By default, phi
includes a prompt template designed for multi-turn conversations:
% ollama run phi
>>> Hello, can you help me find my way to Toronto?
Certainly! What is the exact location in Toronto that you are looking for?
>>> Yonge & Bloor
Sure, Yonge and Bloor is a busy intersection in downtown Toronto. Would you like to take public transportation or drive there?
>>> Public transportation
Great! The easiest way to get there is by taking the TTC subway. You can take Line 1, which runs along Yonge Street and passes through downtown Toronto.
Using Ollama's API:
curl http://localhost:11434/api/chat -d '{
"model": "phi",
"messages": [
{ "role": "user", "content": "why is the sky blue?" }
]
}'
Example prompts (raw mode)
Phi also responds well to a wide variety of prompt formats when using raw mode in Ollama's API, which bypasses all default prompt templating:
Instruct
curl http://localhost:11434/api/generate -d '{
"model": "phi",
"prompt": "Instruct: Write a detailed analogy between mathematics and a lighthouse.\nOutput:",
"options": {
"stop": ["Instruct:", "Output:"]
},
"raw": true,
"stream": false
}'
Code Completion
curl http://localhost:11434/api/generate -d '{
"model": "phi",
"prompt": "def print_prime(n):\n ",
"raw": true,
"stream": false
}'
Text completion
curl http://localhost:11434/api/generate -d '{
"model": "phi",
"prompt": "There once was a mouse named",
"raw": true,
"stream": false
}'
New Models
- Phi-2: A versatile 2.7B model by Microsoft with outstanding reasoning and language understanding capabilities.
- Solar: A compact, yet powerful 10.7B large language model designed for single-turn conversation.
- OpenChat: Updated to OpenChat-3.5-1210, this new version of the 7B model model excels at coding tasks and scores very high on many open-source LLM benchmarks.
- Wizard Math: Updated to WizardMath v1.1, this 7B model excels at Math logic and reasoning and is now based on Mistral
What's Changed
- Fixed issues where
message
objects in/api/chat
would return"images": null
in the response /api/chat
now always returns amessage
object, even ifcontent
is an empty string
New Contributors
- @ianpurton made their first contribution in #1463
- @AugustDev made their first contribution in #1529
Full Changelog: v0.1.16...v0.1.17
v0.1.16
This release adds support Mixtral and other models based on its Mixture of Experts (MoE) architecture. Note: this model requires at least 48GB of memory.
ollama run mixtral
New models
- Mixtral: A high-quality mixture of experts model with open weights.
- Dolphin Mixtral: An uncensored, fine-tuned model based on the Mixtral mixture of experts model that excels at coding tasks.
What's Changed
- Add support Mixtral and other models based on its Mixture of Experts (MoE) architecture
- Fixed issue where
load_duration
was not in the response for/api/generate
Full Changelog: v0.1.15...v0.1.16
v0.1.15
Multimodal model support
Ollama now supports multimodal models that can describe what they see. To use a multimodal model with ollama run
, include the full path of a png
or jpeg
image in the prompt:
% ollama run llava
>>> What does the text in this image say? /Users/mchiang/Downloads/image.png
Added image '/Users/mchiang/Downloads/image.png'
The text in this image says "The Ollamas."
API usage
A new images
parameter has been added to the Generate API, which takes a list of base64-encoded png
or jpeg
images. Images up to 100MB in size are supported.
curl http://localhost:11434/api/generate -d '{
"model": "llava",
"prompt":"What is in this picture?",
"images": ["iVBORw0KGgoAAAANSUhEUgAAAG0AAABmCAYAAADBPx+VAAAACXBIWXMAAAsTAAALEwEAmpwYAAAAAXNSR0IArs4c6QAAAARnQU1BAACxjwv8YQUAAA3VSURBVHgB7Z27r0zdG8fX743i1bi1ikMoFMQloXRpKFFIqI7LH4BEQ+NWIkjQuSWCRIEoULk0gsK1kCBI0IhrQVT7tz/7zZo888yz1r7MnDl7z5xvsjkzs2fP3uu71nNfa7lkAsm7d++Sffv2JbNmzUqcc8m0adOSzZs3Z+/XES4ZckAWJEGWPiCxjsQNLWmQsWjRIpMseaxcuTKpG/7HP27I8P79e7dq1ars/yL4/v27S0ejqwv+cUOGEGGpKHR37tzJCEpHV9tnT58+dXXCJDdECBE2Ojrqjh071hpNECjx4cMHVycM1Uhbv359B2F79+51586daxN/+pyRkRFXKyRDAqxEp4yMlDDzXG1NPnnyJKkThoK0VFd1ELZu3TrzXKxKfW7dMBQ6bcuWLW2v0VlHjx41z717927ba22U9APcw7Nnz1oGEPeL3m3p2mTAYYnFmMOMXybPPXv2bNIPpFZr1NHn4HMw0KRBjg9NuRw95s8PEcz/6DZELQd/09C9QGq5RsmSRybqkwHGjh07OsJSsYYm3ijPpyHzoiacg35MLdDSIS/O1yM778jOTwYUkKNHWUzUWaOsylE00MyI0fcnOwIdjvtNdW/HZwNLGg+sR1kMepSNJXmIwxBZiG8tDTpEZzKg0GItNsosY8USkxDhD0Rinuiko2gfL/RbiD2LZAjU9zKQJj8RDR0vJBR1/Phx9+PHj9Z7REF4nTZkxzX4LCXHrV271qXkBAPGfP/atWvu/PnzHe4C97F48eIsRLZ9+3a3f/9+87dwP1JxaF7/3r17ba+5l4EcaVo0lj3SBq5kGTJSQmLWMjgYNei2GPT1MuMqGTDEFHzeQSP2wi/jGnkmPJ/nhccs44jvDAxpVcxnq0F6eT8h4ni/iIWpR5lPyA6ETkNXoSukvpJAD3AsXLiwpZs49+fPn5ke4j10TqYvegSfn0OnafC+Tv9ooA/JPkgQysqQNBzagXY55nO/oa1F7qvIPWkRL12WRpMWUvpVDYmxAPehxWSe8ZEXL20sadYIozfmNch4QJPAfeJgW3rNsnzphBKNJM2KKODo1rVOMRYik5ETy3ix4qWNI81qAAirizgMIc+yhTytx0JWZuNI03qsrgWlGtwjoS9XwgUhWGyhUaRZZQNNIEwCiXD16tXcAHUs79co0vSD8rrJCIW98pzvxpAWyyo3HYwqS0+H0BjStClcZJT5coMm6D2LOF8TolGJtK9fvyZpyiC5ePFi9nc/oJU4eiEP0jVoAnHa9wyJycITMP78+eMeP37sXrx44d6+fdt6f82aNdkx1pg9e3Zb5W+RSRE+n+VjksQWifvVaTKFhn5O8my63K8Qabdv33b379/PiAP//vuvW7BggZszZ072/+TJk91YgkafPn166zXB1rQHFvouAWHq9z3SEevSUerqCn2/dDCeta2jxYbr69evk4MHDyY7d+7MjhMnTiTPnz9Pfv/+nfQT2ggpO2dMF8cghuoM7Ygj5iWCqRlGFml0QC/ftGmTmzt3rmsaKDsgBSPh0/8yPeLLBihLkOKJc0jp8H8vUzcxIA1k6QJ/c78tWEyj5P3o4u9+jywNPdJi5rAH9x0KHcl4Hg570eQp3+vHXGyrmEeigzQsQsjavXt38ujRo44LQuDDhw+TW7duRS1HGgMxhNXHgflaNTOsHyKvHK5Ijo2jbFjJBQK9YwFd6RVMzfgRBmEfP37suBBm/p49e1qjEP2mwTViNRo0VJWH1deMXcNK08uUjVUu7s/zRaL+oLNxz1bpANco4npUgX4G2eFbpDFyQoQxojBCpEGSytmOH8qrH5Q9vuzD6ofQylkCUmh8DBAr+q8JCyVNtWQIidKQE9wNtLSQnS4jDSsxNHogzFuQBw4cyM61UKVsjfr3ooBkPSqqQHesUPWVtzi9/vQi1T+rJj7WiTz4Pt/l3LxUkr5P2VYZaZ4URpsE+st/dujQoaBBYokbrz/8TJNQYLSonrPS9kUaSkPeZyj1AWSj+d+VBoy1pIWVNed8P0Ll/ee5HdGRhrHhR5GGN0r4LGZBaj8oFDJitBTJzIZgFcmU0Y8ytWMZMzJOaXUSrUs5RxKnrxmbb5YXO9VGUhtpXldhEUogFr3IzIsvlpmdosVcGVGXFWp2oU9kLFL3dEkSz6NHEY1sjSRdIuDFWEhd8KxFqsRi1uM/nz9/zpxnwlESONdg6dKlbsaMGS4EHFHtjFIDHwKOo46l4TxSuxgDzi+rE2jg+BaFruOX4HXa0Nnf1lwAPufZeF8/r6zD97WK2qFnGjBxTw5qNGPxT+5T/r7/7RawFC3j4vTp09koCxkeHjqbHJqArmH5UrFKKksnxrK7FuRIs8STfBZv+luugXZ2pR/pP9Ois4z+TiMzUUkUjD0iEi1fzX8GmXyuxUBRcaUfykV0YZnlJGKQpOiGB76x5GeWkWWJc3mOrK6S7xdND+W5N6XyaRgtWJFe13GkaZnKOsYqGdOVVVbGupsyA/l7emTLHi7vwTdirNEt0qxnzAvBFcnQF16xh/TMpUuXHDowhlA9vQVraQhkudRdzOnK+04ZSP3DUhVSP61YsaLtd/ks7ZgtPcXqPqEafHkdqa84X6aCeL7YWlv6edGFHb+ZFICPlljHhg0bKuk0CSvVznWsotRu433alNdFrqG45ejoaPCaUkWERpLXjzFL2Rpllp7PJU2a/v7Ab8N05/9t27Z16KUqoFGsxnI9EosS2niSYg9SpU6B4JgTrvVW1flt1sT+0ADIJU2maXzcUTraGCRaL1Wp9rUMk16PMom8QhruxzvZIegJjFU7LLCePfS8uaQdPny4jTTL0dbee5mYokQsXTIWNY46kuMbnt8Kmec+LGWtOVIl9cT1rCB0V8WqkjAsRwta93TbwNYoGKsUSChN44lgBNCoHLHzquYKrU6qZ8lolCIN0Rh6cP0Q3U6I6IXILYOQI513hJaSKAorFpuHXJNfVlpRtmYBk1Su1obZr5dnKAO+L10Hrj3WZW+E3qh6IszE37F6EB+68mGpvKm4eb9bFrlzrok7fvr0Kfv727dvWRmdVTJHw0qiiCUSZ6wCK+7XL/AcsgNyL74DQQ730sv78Su7+t/A36MdY0sW5o40ahslXr58aZ5HtZB8GH64m9EmMZ7FpYw4T6QnrZfgenrhFxaSiSGXtPnz57e9TkNZLvTjeqhr734CNtrK41L40sUQckmj1lGKQ0rC37x544r8eNXRpnVE3ZZY7zXo8NomiO0ZUCj2uHz58rbXoZ6gc0uA+F6ZeKS/jhRDUq8MKrTho9fEkihMmhxtBI1DxKFY9XLpVcSkfoi8JGnToZO5sU5aiDQIW716ddt7ZLYtMQlhECdBGXZZMWldY5BHm5xgAroWj4C0hbYkSc/jBmggIrXJWlZM6pSETsEPGqZOndr2uuuR5rF169a2HoHPdurUKZM4CO1WTPqaDaAd+GFGKdIQkxAn9RuEWcTRyN2KSUgiSgF5aWzPTeA/lN5rZubMmR2bE4SIC4nJoltgAV/dVefZm72AtctUCJU2CMJ327hxY9t7EHbkyJFseq+EJSY16RPo3Dkq1kkr7+q0bNmyDuLQcZBEPYmHVdOBiJyIlrRDq41YPWfXOxUysi5fvtyaj+2BpcnsUV/oSoEMOk2CQGlr4ckhBwaetBhjCwH0ZHtJROPJkyc7UjcYLDjmrH7ADTEBXFfOYmB0k9oYBOjJ8b4aOYSe7QkKcYhFlq3QYLQhSidNmtS2RATwy8YOM3EQJsUjKiaWZ+vZToUQgzhkHXudb/PW5YMHD9yZM2faPsMwoc7RciYJXbGuBqJ1UIGKKLv915jsvgtJxCZDubdXr165mzdvtr1Hz5LONA8jrUwKPqsmVesKa49S3Q4WxmRPUEYdTjgiUcfUwLx589ySJUva3oMkP6IYddq6HMS4o55xBJBUeRjzfa4Zdeg56QZ43LhxoyPo7Lf1kNt7oO8wWAbNwaYjIv5lhyS7kRf96dvm5Jah8vfvX3flyhX35cuX6HfzFHOToS1H4BenCaHvO8pr8iDuwoUL7tevX+b5ZdbBair0xkFIlFDlW4ZknEClsp/TzXyAKVOmmHWFVSbDNw1l1+4f90U6IY/q4V27dpnE9bJ+v87QEydjqx/UamVVPRG+mwkNTYN+9tjkwzEx+atCm/X9WvWtDtAb68Wy9LXa1UmvCDDIpPkyOQ5ZwSzJ4jMrvFcr0rSjOUh+GcT4LSg5ugkW1Io0/SCDQBojh0hPlaJdah+tkVYrnTZowP8iq1F1TgMBBauufyB33x1v+NWFYmT5KmppgHC+NkAgbmRkpD3yn9QIseXymoTQFGQmIOKTxiZIWpvAatenVqRVXf2nTrAWMsPnKrMZHz6bJq5jvce6QK8J1cQNgKxlJapMPdZSR64/UivS9NztpkVEdKcrs5alhhWP9NeqlfWopzhZScI6QxseegZRGeg5a8C3Re1Mfl1ScP36ddcUaMuv24iOJtz7sbUjTS4qBvKmstYJoUauiuD3k5qhyr7QdUHMeCgLa1Ear9NquemdXgmum4fvJ6w1lqsuDhNrg1qSpleJK7K3TF0Q2jSd94uSZ60kK1e3qyVpQK6PVWXp2/FC3mp6jBhKKOiY2h3gtUV64TWM6wDETRPLDfSakXmH3w8g9Jlug8ZtTt4kVF0kLUYYmCCtD/DrQ5YhMGbA9L3ucdjh0y8kOHW5gU/VEEmJTcL4Pz/f7mgoAbYkAAAAAElFTkSuQmCC"]
}'
With the new Chat API introduced in version 0.1.14
, images
can also be added to messages from the user
role:
curl http://localhost:11434/api/chat -d '{
"model": "llava",
"messages": [
{
"role": "user",
"content": "What is in this picture?",
"images": ["iVBORw0KGgoAAAANSUhEUgAAAG0AAABmCAYAAADBPx+VAAAACXBIWXMAAAsTAAALEwEAmpwYAAAAAXNSR0IArs4c6QAAAARnQU1BAACxjwv8YQUAAA3VSURBVHgB7Z27r0zdG8fX743i1bi1ikMoFMQloXRpKFFIqI7LH4BEQ+NWIkjQuSWCRIEoULk0gsK1kCBI0IhrQVT7tz/7zZo888yz1r7MnDl7z5xvsjkzs2fP3uu71nNfa7lkAsm7d++Sffv2JbNmzUqcc8m0adOSzZs3Z+/XES4ZckAWJEGWPiCxjsQNLWmQsWjRIpMseaxcuTKpG/7HP27I8P79e7dq1ars/yL4/v27S0ejqwv+cUOGEGGpKHR37tzJCEpHV9tnT58+dXXCJDdECBE2Ojrqjh071hpNECjx4cMHVycM1Uhbv359B2F79+51586daxN/+pyRkRFXKyRDAqxEp4yMlDDzXG1NPnnyJKkThoK0VFd1ELZu3TrzXKxKfW7dMBQ6bcuWLW2v0VlHjx41z717927ba22U9APcw7Nnz1oGEPeL3m3p2mTAYYnFmMOMXybPPXv2bNIPpFZr1NHn4HMw0KRBjg9NuRw95s8PEcz/6DZELQd/09C9QGq5RsmSRybqkwHGjh07OsJSsYYm3ijPpyHzoiacg35MLdDSIS/O1yM778jOTwYUkKNHWUzUWaOsylE00MyI0fcnOwIdjvtNdW/HZwNLGg+sR1kMepSNJXmIwxBZiG8tDTpEZzKg0GItNsosY8USkxDhD0Rinuiko2gfL/RbiD2LZAjU9zKQJj8RDR0vJBR1/Phx9+PHj9Z7REF4nTZkxzX4LCXHrV271qXkBAPGfP/atWvu/PnzHe4C97F48eIsRLZ9+3a3f/9+87dwP1JxaF7/3r17ba+5l4EcaVo0lj3SBq5kGTJSQmLWMjgYNei2GPT1MuMqGTDEFHzeQSP2wi/jGnkmPJ/nhccs44jvDAxpVcxnq0F6eT8h4ni/iIWpR5lPyA6ETkNXoSukvpJAD3AsXLiwpZs49+fPn5ke4j10TqYvegSfn0OnafC+Tv9ooA/JPkgQysqQNBzagXY55nO/oa1F7qvIPWkRL12WRpMWUvpVDYmxAPehxWSe8ZEXL20sadYIozfmNch4QJPAfeJgW3rNsnzphBKNJM2KKODo1rVOMRYik5ETy3ix4qWNI81qAAirizgMIc+yhTytx0JWZuNI03qsrgWlGtwjoS9XwgUhWGyhUaRZZQNNIEwCiXD16tXcAHUs79co0vSD8rrJCIW98pzvxpAWyyo3HYwqS0+H0BjStClcZJT5coMm6D2LOF8TolGJtK9fvyZpyiC5ePFi9nc/oJU4eiEP0jVoAnHa9wyJycITMP78+eMeP37sXrx44d6+fdt6f82aNdkx1pg9e3Zb5W+RSRE+n+VjksQWifvVaTKFhn5O8my63K8Qabdv33b379/PiAP//vuvW7BggZszZ072/+TJk91YgkafPn166zXB1rQHFvouAWHq9z3SEevSUerqCn2/dDCeta2jxYbr69evk4MHDyY7d+7MjhMnTiTPnz9Pfv/+nfQT2ggpO2dMF8cghuoM7Ygj5iWCqRlGFml0QC/ftGmTmzt3rmsaKDsgBSPh0/8yPeLLBihLkOKJc0jp8H8vUzcxIA1k6QJ/c78tWEyj5P3o4u9+jywNPdJi5rAH9x0KHcl4Hg570eQp3+vHXGyrmEeigzQsQsjavXt38ujRo44LQuDDhw+TW7duRS1HGgMxhNXHgflaNTOsHyKvHK5Ijo2jbFjJBQK9YwFd6RVMzfgRBmEfP37suBBm/p49e1qjEP2mwTViNRo0VJWH1deMXcNK08uUjVUu7s/zRaL+oLNxz1bpANco4npUgX4G2eFbpDFyQoQxojBCpEGSytmOH8qrH5Q9vuzD6ofQylkCUmh8DBAr+q8JCyVNtWQIidKQE9wNtLSQnS4jDSsxNHogzFuQBw4cyM61UKVsjfr3ooBkPSqqQHesUPWVtzi9/vQi1T+rJj7WiTz4Pt/l3LxUkr5P2VYZaZ4URpsE+st/dujQoaBBYokbrz/8TJNQYLSonrPS9kUaSkPeZyj1AWSj+d+VBoy1pIWVNed8P0Ll/ee5HdGRhrHhR5GGN0r4LGZBaj8oFDJitBTJzIZgFcmU0Y8ytWMZMzJOaXUSrUs5RxKnrxmbb5YXO9VGUhtpXldhEUogFr3IzIsvlpmdosVcGVGXFWp2oU9kLFL3dEkSz6NHEY1sjSRdIuDFWEhd8KxFqsRi1uM/nz9/zpxnwlESONdg6dKlbsaMGS4EHFHtjFIDHwKOo46l4TxSuxgDzi+rE2jg+BaFruOX4HXa0Nnf1lwAPufZeF8/r6zD97WK2qFnGjBxTw5qNGPxT+5T/r7/7RawFC3j4vTp09koCxkeHjqbHJqArmH5UrFKKksnxrK7FuRIs8STfBZv+luugXZ2pR/pP9Ois4z+TiMzUUkUjD0iEi1fzX8GmXyuxUBRcaUfykV0YZnlJGKQpOiGB76x5GeWkWWJc3mOrK6S7xdND+W5N6XyaRgtWJFe13GkaZnKOsYqGdOVVVbGupsyA/l7emTLHi7vwTdirNEt0qxnzAvBFcnQF16xh/TMpUuXHDowhlA9vQVraQhkudRdzOnK+04ZSP3DUhVSP61YsaLtd/ks7ZgtPcXqPqEafHkdqa84X6aCeL7YWlv6edGFHb+ZFICPlljHhg0bKuk0CSvVznWsotRu433alNdFrqG45ejoaPCaUkWERpLXjzFL2Rpllp7PJU2a/v7Ab8N05/9t27Z16KUqoFGsxnI9EosS2niSYg9SpU6B4JgTrvVW1flt1sT+0ADIJU2maXzcUTraGCRaL1Wp9rUMk16PMom8QhruxzvZIegJjFU7LLCePfS8uaQdPny4jTTL0dbee5mYokQsXTIWNY46kuMbnt8Kmec+LGWtOVIl9cT1rCB0V8WqkjAsRwta93TbwNYoGKsUSChN44lgBNCoHLHzquYKrU6qZ8lolCIN0Rh6cP0Q3U6I6IXILYOQI513hJaSKAorFpuHXJNfVlpRtmYBk1Su1obZr5dnKAO+L10Hrj3WZW+E3qh6IszE37F6EB+68mGpvKm4eb9bFrlzrok7fvr0Kfv727dvWRmdVTJHw0qiiCUSZ6wCK+7XL/AcsgNyL74DQQ730sv78Su7+t/A36MdY0sW5o40ahslXr58aZ5HtZB8GH64m9EmMZ7FpYw4T6QnrZfgenrhFxaSiSGXtPnz57e9TkNZLvTjeqhr734CNtrK41L40sUQckmj1lGKQ0rC37x544r8eNXRpnVE3ZZY7zXo8NomiO0ZUCj2uHz58rbXoZ6gc0uA+F6ZeKS/jhRDUq8MKrTho9fEkihMmhxtBI1DxKFY9XLpVcSkfoi8JGnToZO5sU5aiDQIW716ddt7ZLYtMQlhECdBGXZZMWldY5BHm5xgAroWj4C0hbYkSc/jBmggIrXJWlZM6pSETsEPGqZOndr2uuuR5rF169a2HoHPdurUKZM4CO1WTPqaDaAd+GFGKdIQkxAn9RuEWcTRyN2KSUgiSgF5aWzPTeA/lN5rZubMmR2bE4SIC4nJoltgAV/dVefZm72AtctUCJU2CMJ327hxY9t7EHbkyJFseq+EJSY16RPo3Dkq1kkr7+q0bNmyDuLQcZBEPYmHVdOBiJyIlrRDq41YPWfXOxUysi5fvtyaj+2BpcnsUV/oSoEMOk2CQGlr4ckhBwaetBhjCwH0ZHtJROPJkyc7UjcYLDjmrH7ADTEBXFfOYmB0k9oYBOjJ8b4aOYSe7QkKcYhFlq3QYLQhSidNmtS2RATwy8YOM3EQJsUjKiaWZ+vZToUQgzhkHXudb/PW5YMHD9yZM2faPsMwoc7RciYJXbGuBqJ1UIGKKLv915jsvgtJxCZDubdXr165mzdvtr...
v0.1.14
New Models
- StableLM Zephyr: A lightweight chat model allowing accurate, and responsive output without requiring high-end hardware.
- Magicoder: a family of 7B parameter models trained on 75K synthetic instruction data using OSS-Instruct, a novel approach to enlightening LLMs with open-source code snippets.
What's Changed
- New Chat API for sending a history of messages
curl http://localhost:11434/api/chat -d '{ "model": "mistral", "messages": [ { "role": "system", "content": "You are a helpful assistant that answers concisely." }, { "role": "user", "content": "why is the sky blue?" } ] }'
- Linewrap now works when resizing the terminal with
ollama run
- Fixed an issue where ctrl+z would not suspend
ollama run
as expected - Fixed an issue where requests to
/api/generate
would not work when waiting for another request to finish - Fixed an issue where extra whitespace after a
FROM
command would cause an error - Ollama will now warn you if there's a version mismatch when connecting remotely with
OLLAMA_HOST
- New
/api/version
API for checking Ollama's version
New Contributors
- @ruecat made their first contribution in #1364
- @calderonsamuel made their first contribution in #1399
- @Xe made their first contribution in #1406
Full Changelog: v0.1.13...v0.1.14