Skip to content

Regression 2026.0 - ovms stuck in initializing servable #4076

@jpm-canonical

Description

@jpm-canonical

Describe the bug
Likely a duplicate of #4059

OVMS v2026.0 gets stuck in Initializing Visual Language Model Legacy servable, using 100% on a single thread. SIGTERM (ctrl+c) is ignored, and one needs to send the SIGKILL signal to stop it.

OVMS v2025.4.1 does not have this issue.

To Reproduce
Steps to reproduce the behavior:

  1. Download ovms_ubuntu24_python_on.tar.gz for release 2026.0. Extract.
  2. OVMS launch command:
    #!/bin/bash
    
    # Add staged shared objects
    export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:./lib
    export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:./usr/lib/$ARCH_TRIPLET
    export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:./usr/local/lib
    #export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/local/lib
    
    # Add Python dependencies
    export PYTHONPATH=$PYTHONPATH:./lib/python # part of ovms package
    export PYTHONPATH=$PYTHONPATH:./lib/python3.12/site-packages
    
    exec bin/ovms --rest_port 8328 --rest_bind_address 127.0.0.1 \
    --source_model gemma-3-4b-it-ov-int4-fq \
    --model_repository_path /home/jpmeijers/models \
    --pipeline_type VLM --task text_generation \
    --cache_size 2 --log_level DEBUG --target_device GPU
    
  3. See logs get stuck at
    [modelmanager][info][servable_initializer.cpp:448] Initializing Visual Language Model Legacy servable

Logs
TRACE enabled.

OVMS 2026.0 getting stuck: ovms-logs-2026.0.txt

OVMS 2025.4.1 working as expected: ovms-logs-2024.4.1.txt

Configuration

  1. OVMS version - 2026.0 fails, 2025.4.1 works
  2. OVMS config.json file - none
  3. CPU, accelerator's versions if applicable:
    • Intel(R) Core(TM) i5-3570K CPU @ 3.40GHz
    • Intel Corporation DG2 [Arc A580] (rev 08)
  4. Model repository directory structure
    $ tree models/
    models/
    └── gemma-3-4b-it-ov-int4-fq
        ├── added_tokens.json
        ├── chat_template.jinja
        ├── config.json
        ├── generation_config.json
        ├── graph.pbtxt
        ├── openvino_config.json
        ├── openvino_detokenizer.bin
        ├── openvino_detokenizer.xml
        ├── openvino_language_model.bin
        ├── openvino_language_model.xml
        ├── openvino_text_embeddings_model.bin
        ├── openvino_text_embeddings_model.xml
        ├── openvino_tokenizer.bin
        ├── openvino_tokenizer.xml
        ├── openvino_vision_embeddings_model.bin
        ├── openvino_vision_embeddings_model.xml
        ├── preprocessor_config.json
        ├── processor_config.json
        ├── README.md
        ├── special_tokens_map.json
        ├── tokenizer_config.json
        ├── tokenizer.json
        └── tokenizer.model
    
    2 directories, 23 files
    
  5. Model or publicly available similar model that reproduces the issue:
    git clone --depth 1 https://huggingface.co/llmware/gemma-3-4b-it-ov-int4-fq
  6. OS: Ubuntu Desktop 24.04.4

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions