Skip to content

Commit e1ade17

Browse files
committed
Bump version to 0.3.35
Signed-off-by: JamePeng <jame_peng@sina.com>
1 parent 232092e commit e1ade17

2 files changed

Lines changed: 49 additions & 1 deletion

File tree

CHANGELOG.md

Lines changed: 48 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,54 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
77

88
## [Unreleased]
99

10+
## [0.3.35] Gemma 4 series & LFM 2.5-VL Support, OpenAI OpenAPI Alignment and Logging Architecture Migration
11+
12+
- fix: expand stop sequences for `Gemma4ChatHandler`
13+
- Add `GEMMA4_EOS_TOKEN` and `GEMMA4_STR_TOKEN` to the generation stop criteria.
14+
- Align the stopping logic with the model's `generation_config.json` definitions.
15+
- Prevent potential over-generation by ensuring the model halts correctly at standard EOS or when initiating a tool response.
16+
17+
- feat(types): align with latest OpenAI OpenAPI spec (audio, structured outputs)
18+
- Update llama_types.py OpenAI [OpenAPI Link](https://app.stainless.com/api/spec/documented/openai/openapi.documented.yml)
19+
- Add `developer` role.
20+
- Replace Anyscale-specific JSON schema with official OpenAI `json_schema` response format for Structured Outputs.
21+
- Add `input_audio` and `file` types to request message content parts.
22+
- Add `audio`, `refusal`, and `annotations` (e.g., URL citations) fields to response messages.
23+
- Add `content_filter` to finish reasons and strictly define global `ChatCompletionRole`.
24+
25+
- docs: clarify `enable_thinking` compatibility for **Gemma 4** models
26+
- Update `Gemma4ChatHandler` class docstring and `__init__` args documentation.
27+
- Specify that the `enable_thinking` toggle is exclusively supported by Gemma4 31B and 26BA4B variants.
28+
- Explicitly note that E2B and E4B models do not currently support this feature to prevent configuration errors.
29+
30+
- feat(chat_format): Implemented `Gemma4ChatHandler`, add Gemma 4 chat handler with multimodal and tool support
31+
- Implement `Gemma4ChatHandler` with Gemma 4 specific tokens (`<|turn>`, `<|channel>`, etc.).
32+
- Add complex Jinja2 template for advanced nested tool/function schema formatting.
33+
- Support multimodal content injection for `image_url`, `audio_url`, and `input_audio` (including base64 reconstruction).
34+
- Integrate reasoning/thinking controls via `enable_thinking` toggle and `<|channel>thought` formatting.
35+
- Configure `<turn|>` as the primary stop sequence for generation boundaries.
36+
37+
- feat(chat_format) Implemented `LFM25VLChatHandler` for **LFM2.5-VL** (by **@alcoftTAO**)
38+
39+
- fix Qwen3.5 chat template typos(reported by **@abdullah-cod9**)
40+
41+
- refactor(logger): migrate from llama_log_callback to ggml_log_callback
42+
- Remove the deprecated `llama_log_callback` typedef from `llama_cpp.py`.
43+
- Update `_logger.py` to use `ggml_log_callback` from `_ggml`, aligning with the upstream GGML logging architecture.
44+
- Rename the callback references across the codebase, including the MTMD context initialization in `llama_chat_format.py`.
45+
46+
- feat(ggml): add support for ggml-base library and new function bindings
47+
- Load the new `ggml-base` shared library alongside `ggml`.
48+
- Add `ctypes` bindings for `ggml_log_get`, `ggml_log_set`, and `ggml_set_zero` using the `ggml_base_function` decorator.
49+
50+
- Update README.md
51+
52+
- feat: Update llama.cpp to [ggml-org/llama.cpp/commit/58190cc84d846d8575ba26e8486bc29d9fd8ad55](https://github.com/ggml-org/llama.cpp/commit/58190cc84d846d8575ba26e8486bc29d9fd8ad55)
53+
54+
- feat: Sync llama.cpp llama/mtmd API Binding 20260402
55+
56+
More information see: https://github.com/JamePeng/llama-cpp-python/compare/a184583e908cc138fd15794986b3581521fb9b0c...232092e32b3563159a86aacb168da06c4937192b
57+
1058
## [0.3.34] Dynamic LoRA Routing, Control Vectors, and Assistant Prefill
1159

1260
- **feat(chat_format): added assistant_prefill to seamlessly continue responses**

llama_cpp/__init__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
from .llama_cpp import *
22
from .llama import *
33

4-
__version__ = "0.3.34"
4+
__version__ = "0.3.35"

0 commit comments

Comments
 (0)