Skip to content

ggml-cuda: auto apply iGPU flag for CUDA/HIP if integrated device#23007

Open
fl0rianr wants to merge 1 commit into
ggml-org:masterfrom
fl0rianr:fix/auto-apply_iGPU_flag
Open

ggml-cuda: auto apply iGPU flag for CUDA/HIP if integrated device#23007
fl0rianr wants to merge 1 commit into
ggml-org:masterfrom
fl0rianr:fix/auto-apply_iGPU_flag

Conversation

@fl0rianr
Copy link
Copy Markdown
Contributor

Overview

Report CUDA/HIP devices as GGML_BACKEND_DEVICE_TYPE_IGPU when the runtime reports
cudaDeviceProp::integrated.

This intentionally checks cudaDeviceProp directly instead of
ggml_cuda_info().devices[id].integrated, since the latter one is temporally disabled
by design for CUDA buffer allocation behavior due to #15034. So this change only affects the automatic backend
device classification and does not re-enable the integrated-buffer path.

Requirements

@fl0rianr fl0rianr requested a review from a team as a code owner May 13, 2026 09:02
@github-actions github-actions Bot added Nvidia GPU Issues specific to Nvidia GPUs ggml changes relating to the ggml tensor library for machine learning labels May 13, 2026
@0cc4m
Copy link
Copy Markdown
Contributor

0cc4m commented May 13, 2026

It's good to see the IGPU type expand to other backends, we can use it to adapt downstream behaviour according to the device type to e.g. disable mmap on integrated GPUs.

@fl0rianr fl0rianr changed the title ggml: auto apply iGPU flag for CUDA/HIP if integrated device ggml-cuda: auto apply iGPU flag for CUDA/HIP if integrated device May 13, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ggml changes relating to the ggml tensor library for machine learning Nvidia GPU Issues specific to Nvidia GPUs

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants