v0.1.34
New models
- Llava Llama 3: A new high-performing LLaVA model fine-tuned from Llama 3 Instruct.
- Llava Phi 3: A new small LLaVA model fine-tuned from Phi 3.
- StarCoder2 15B Instruct: A new instruct fine-tune of the StarCoder2 model
- CodeGemma 1.1: A new release of the CodeGemma model.
- StableLM2 12B: A new 12B version of the StableLM 2 model from Stability AI
- Moondream 2: Moondream 2's runtime parameters have been improved for better responses
What's Changed
- Fixed issues with LLaVa models where they would respond incorrectly after the first request
- Fixed out of memory errors when running large models such as Llama 3 70B
- Fixed various issues with Nvidia GPU discovery on Linux and Windows
- Fixed a series of Modelfile errors when running
ollama create
- Fixed
no slots available
error that occurred when cancelling a request and then sending follow up requests - Improved AMD GPU detection on Fedora
- Improved reliability when using the experimental
OLLAMA_NUM_PARALLEL
andOLLAMA_MAX_LOADED
flags ollama serve
will now shut down quickly, even if a model is loading
New Contributors
- @drnic made their first contribution in #4116
- @bernardo-bruning made their first contribution in #4111
- @Drlordbasil made their first contribution in #4174
- @Saif-Shines made their first contribution in #4119
- @HydenLiu made their first contribution in #4194
- @jl-codes made their first contribution in #3621
- @Nurgo made their first contribution in #3473
- @adrienbrault made their first contribution in #3129
- @Darinochka made their first contribution in #3945
Full Changelog: v0.1.33...v0.1.34