@wgottwalt Thank you for your insight, I have explored this further if you are interested in taking a look as it sounds like you are far more experienced than myself, though I am aware of the explicit deprecation of the arch. It seems some have had success building from SDK or even in this docker (below) though I do not believe any are on so recent a build as 6.3+ (I think current SDK build is 6.1.2.0). I wonder if any patches from the SDK builder should be rolled in upstream as they seem to address many aspects of compatibility for older GCN. My use case is whisperx + llm+cot+rag . I can accept that I may just need to update my card. I was able to utilise vulkan for llama via kobold (I think https://github.com/LostRuins/koboldcpp - I think someone mentions that in one of the sdk issues) and performance was middling, though for the price of the cards as people upgrade, I feel some reasonable performance could be achieved across several and could be ideal for eg raspi+pcie switch (not sure if that's a thing) or even just single card. arm probably runs it more efficiently lol. i think the 580's performance was about equal to 7700k but with the advantage of not completely bricking your system, maybe i need to try a different scheduler or limit cores but i feel that somewhat defeats the purpose. my lspci -vvv is in one of these:
https://github.com/lamikr/rocm_sdk_builder https://github.com/lamikr/rocm_sdk_builder/issues/220 https://github.com/lamikr/rocm_sdk_builder/issues/173 https://github.com/robertrosenbusch/gfx803_rocm/issues/6#issuecomment-2719117249
Pinned Comments
wgottwalt commented on 2024-11-09 10:46 (UTC) (edited on 2024-11-26 15:23 (UTC) by wgottwalt)
Looks like the ROCm 6.2.2-1 SDK has a malfunctioning compiler. It produces a broken ollama binary (fp16 issues). You may need to stay with ROCm 6.0.2 for now. I don't know if this got fixed in a newer build release. But the initial SDK version "-1" is broken.
ROCm 6.2.4 fixes this issue completely.