LXC Ryzen 7 (5800H)
hkfuertes opened this issue · comments
Hello, I'm trying to run ollama faster on an LXC container in my proxmox homelab. I have some questions, and maybe you could guide me through them...
- I'm using my APU, and my understanding is that I could use the iGPU to offload ollama onto ROCm, is that correct? my homelab its a minipc so no option to add a second card for VMs...
- On the same homelab I have Jellyfin using the APU (i believe) with VAAPI to transcode my videos (and as I only have one iGPU... I cannot exclusively use it in a vm, no passthrought, is this also correct?)
- So I installed mesa drivers on proxmox host and bind-mounted
/dev/dri
,/dev/kfd
onto the ollama container. For this container I tried several things (ubuntu, debian, and most recently arch... the outcome is the same for all of them, but I will write down the steps for the last one...)
-
pacman -Syu libva-mesa-driver libva-utils rocm-hip-sdk rocm-opencl-sdk rocminfo
-
#/opt/rocm/bin/rocminfo ROCk module is loaded hsa api call failure at: /usr/src/debug/rocminfo/rocminfo-rocm-6.0.0/rocminfo.cc:1219 Call returned HSA_STATUS_ERROR_OUT_OF_RESOURCES: The runtime failed to allocate the necessary resources. This error may also occur when the core runtime library needs to spawn threads or create internal OS-specific events.
- Will I get any improvements using the iGPU (shared memory)? Because if the improvement is going to be zero compared to the CPU, I will not bother anymore 😄 .
Sorry that I write all down... Thank you in advance for the help that I could get!
Following up:
[root@ollamarch ~]# /opt/rocm/bin/rocm-smi
Exception caught: map::at
======================================== ROCm System Management Interface ========================================
================================================== Concise Info ==================================================
Device [Model : Revision] Temp Power Partitions SCLK MCLK Fan Perf PwrCap VRAM% GPU%
Name (20 chars) (Edge) (Avg) (Mem, Compute)
==================================================================================================================
0 [0x0123 : 0xc1] 34.0°C 11.0W N/A, N/A None 1200Mhz 0% auto Unsupported 1% 0%
Lucienne
==================================================================================================================
============================================== End of ROCm SMI Log ===============================================
Does this means that the GPU is recognized (and usable) by the container?