GPU-accelerated LLaMA inference wrapper for legacy Vulkan-capable systems a Pythonic way to run AI with knowledge (Ilm) on fire (Vulkan).
-
Updated
Oct 14, 2025 - Python
GPU-accelerated LLaMA inference wrapper for legacy Vulkan-capable systems a Pythonic way to run AI with knowledge (Ilm) on fire (Vulkan).
infinityを少し前の世代のGPUで動作さるためのリポジトリ | Repository to run Infinity on slightly older generation GPUs
Add a description, image, and links to the legacy-gpus topic page so that developers can more easily learn about it.
To associate your repository with the legacy-gpus topic, visit your repo's landing page and select "manage topics."