Skip to content

Commit 8771222

Browse files
committed
Bump version to 0.3.17
1 parent fbb0ed2 commit 8771222

File tree

2 files changed

+16
-1
lines changed

2 files changed

+16
-1
lines changed

CHANGELOG.md

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,21 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
77

88
## [Unreleased]
99

10+
## [0.3.17]
11+
- feat: Update llama.cpp to [ggml-org/llama.cpp/commit/054a45c3d313387a4becd5eae982285932852b35](https://github.com/ggml-org/llama.cpp/commit/054a45c3d313387a4becd5eae982285932852b35)
12+
- feat: Sync llama.cpp llama/mtmd API Binding 20251121
13+
- feat: **Support clip flash-attn**
14+
- feat: **0day support Qwen3VLChatHandler into llama_chat_format.py**
15+
- Update README.md for Qwen3VL example(Thinking/No Thinking)
16+
- feat: **Better Qwen3VL chat template. (by @alcoftTAO)**
17+
- feat: [Implement LlamaTrieCache into llama_cache.py](https://github.com/JamePeng/llama-cpp-python/commit/2419dc2d9bb0a6be0cd381038ce00fcaea124c76): Optimize LlamaCache lookup from **O(N)** to **O(K)** using a Trie, **improves retrieval speed at least 40x compared to the original linear scan method of finding the longest prefix , thereby enhancing service responsiveness.**
18+
- feat: Update Llava15ChatHandler to accept use_gpu, image_min_tokens, and image_max_tokens.Now can pass the`image_min_tokens`parameter in Qwen3VLChatHandler to support bbox grounding tasks.
19+
- feat: [Add Pillow process code in _load_image for VLM](https://github.com/JamePeng/llama-cpp-python/commit/3b0133365e25840c023aef6b6c8578073cd081e8): that can reliably handle common formats, Supports 20+ image formats (PNG, JPEG, WebP, AVIF, HEIC, SVG, BMP, ICO, TIFF, etc.). Images with alpha channel (PNG, WebP, etc.) → automatically composites on white/black background(white for dark content, black for bright content)
20+
- feat: Optimize CUDA Wheel Build Workflow, now workflow action support python3.10-3.13 cu124-cu126-cu128 Basic(Non AVX)-AVX2 win-linux
21+
22+
23+
More information see : https://github.com/JamePeng/llama-cpp-python/compare/e5392b52036bd2770ece5269352f5600a8db5639...fbb0ed2f089c663a5eb75aadcad08f768041ed72
24+
1025
## [0.3.16]
1126

1227
- feat: Update llama.cpp to [ggml-org/llama.cpp/commit/5e6229a8409ac786e62cb133d09f1679a9aec13e](https://github.com/ggml-org/llama.cpp/commit/5e6229a8409ac786e62cb133d09f1679a9aec13e)

llama_cpp/__init__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
from .llama_cpp import *
22
from .llama import *
33

4-
__version__ = "0.3.16"
4+
__version__ = "0.3.17"

0 commit comments

Comments
 (0)