GPU-accelerated LLaMA inference wrapper for legacy Vulkan-capable systems a Pythonic way to run AI with knowledge (Ilm) on fire (Vulkan).
machine-learning vulkan python-wrapper fastai amd-gpu intel-gpu llama-cpp gpu-inference llm-inference localllm local-ai open-source-llm llama-cpp-python gguf legacy-gpus
-
Updated
Oct 14, 2025 - Python