Skip to content

Conversation

@Jaswanth51
Copy link

@Jaswanth51 Jaswanth51 commented Jan 16, 2026

Added versions deprecation notices and updated documentation

https://jira.devtools.intel.com/browse/CVS-176786


# Use config with session
options = {"device_type": "AUTO", "load_config": "ov_config.json"}
options = {"device_type": "AUTO", "load_config": json.dumps(config)}

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Have you tested this change @Jaswanth51 ?

### `cache_dir`

**DEPRECATED:** This option is deprecated and can be set via `load_config` using the `CACHE_DIR` property.
**DEPRECATED:** This option is deprecated since OpenVINO 2025.3/ORT 1.23 and can be set via `load_config` using the `CACHE_DIR` property. `cache_dir` is configured **per-session** rather than globally.

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

what do you mean by "cache_dir is configured per-session rather than globally". Have you tested it yourself ?

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Does this part need to be mentioned? Aren't all properties set per session ?

```python
import onnxruntime as ort
import json
import openvino

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

what is the need to import openvino ? As far as ik, this is how it is used to wrok previously
if platform.system() == "Windows":
import onnxruntime.tools.add_openvino_win_libs as utils
utils.add_openvino_libs_to_path()

Copy link

@jatinwadhwa921 jatinwadhwa921 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pls make the changes accordingly

@MayureshV1 MayureshV1 changed the title Draft for Documentation update and deprecation notices CVS-176786: Draft for Documentation update and deprecation notices Jan 20, 2026
Copy link

@MayureshV1 MayureshV1 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can we also change line#90 and mention
The following provider options are deprecated since ORT 1.23 and should be migrated to load_config for better compatibility with future releases

### `cache_dir`

**DEPRECATED:** This option is deprecated and can be set via `load_config` using the `CACHE_DIR` property.
**DEPRECATED:** This option is deprecated since OpenVINO 2025.3/ORT 1.23 and can be set via `load_config` using the `CACHE_DIR` property. `cache_dir` is configured **per-session** rather than globally.

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Does this part need to be mentioned? Aren't all properties set per session ?

Enables model caching to significantly reduce subsequent load times. Supports CPU, NPU, and GPU devices with kernel caching on iGPU/dGPU.

**Benefits**
- Saves compiled models and `cl_cache` files for dynamic shapes

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can this be verified? GPU kernel vs model cache is based on a property- OPTIMIZE_SPEED, OPTIMIZE_SIZE.

    • Particularly useful for complex models and frequent application restarts
      This can be updated to: "Particularly useful for optimizing application startup latencies, especially for complex models"

"DEVICE_PROPERTIES": {
"CPU": {
"INFERENCE_PRECISION_HINT": "f32",
"NUM_STREAMS": "3",

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please remove the INFERENCE_NUM_THREADS reference since it is a low level setting that is better ignored by app.

"INFERENCE_NUM_THREADS": "8"
},
"GPU": {
"INFERENCE_PRECISION_HINT": "f16",

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

For GPU you can show an example with using EXECUTION_MODE_HINT: ACCURACY in place of "INFERENCE_PRECISION_HINT".
Note, only 1 should be used since the execution mode internally will set the right inference precision hint.

"AUTO": {
"PERFORMANCE_HINT": "THROUGHPUT",
"PERF_COUNT": "NO",
"DEVICE_PROPERTIES": "{CPU:{INFERENCE_PRECISION_HINT:f32,NUM_STREAMS:3},GPU:{INFERENCE_PRECISION_HINT:f32,NUM_STREAMS:5}}"

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nitpick: Can we be selective in the properties we show as example. We can reuse the ones we have used above.
in general using properties like INFERENCE_PRECISION or CPU would not yield much benefit since CPU only supports FP32.
Also, setting num_streams value to static would yield different result on each platform. it is better to use Performance_Hint which is portable across multiple platforms.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants