Release Discussion: v1.73.1.yr1 | ROCm v6.2.0 #64
Replies: 10 comments 20 replies
-
On my radeon 6900xt works well. I tested different language models and I don't see any problems. Dry works as it should. Performance is slightly better than on the previous version of rocm - example: old 35.77T/s vs new 38.43T/s. In the newer version, slightly faster processing (visible only with large amounts of text). The same vram usage in LLM. Slightly weaker performance in generating images (Stable UI) - for 1024x1024 -> 1.28s/it vs 1.34s/it (the same model and settings). Lower memory usage during image generation (older version vs newer) -> example: 9.3 vs 8.8 and in the final phase 15 vs 14.6 Windows 10 Pro, 22H2. AMD drivers: 24.7.1 |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
On a gfx906 (Radeon VII) I get the following error: "rocBLAS error: Could not initialize Tensile host: No devices found". Had to roll back to the old working 1.72 build. |
Beta Was this translation helpful? Give feedback.
-
I've encountered this error but I don't know what it is except that it comes from rocM Device 0: AMD Radeon RX 6600, compute capability 10.3, VMM: no |
Beta Was this translation helpful? Give feedback.
-
I did some speed comparisons, my testing environment wasn't too well controlled so take these with a grain of salt, but I'll drop these here if anyone's interested. Ryzen 7 5800X3D ROCm 5.7 = KoboldCpp v1.71.1.yr0-ROCm Model 1: Meta-Llama-3.1-8B-Instruct-Q6_K (Plenty of free VRAM)
Model 2: Lumimaid-v0.2-12B-Q4_K_M-imat (Good amount of free VRAM)
Model 3: DaringMaid-20B-V1.1-Q4_K_M (VRAM limited)
Model 4: Lumimaid-v0.2-70B.q4_k_m (Very VRAM limited)
|
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
how do i use the image analyzing option |
Beta Was this translation helpful? Give feedback.
-
Seems like generation speed is still behind Vulkan. Processing speed is much faster though. For RP I think Vulkan is still going to be better with the faster generation speed. Model: daybreak-kunoichi-2dpo-7b-q6_k.gguf Vulkan: v1.74.yr0-ROCm: |
Beta Was this translation helpful? Give feedback.
-
Since ROCM afair supports compiling to a CUDA card (?? does it still, that was advertised at some point), would it be possible to also target CUDA cards at the same time, like for example if you wanted to do multi-gpu across AMD and Nvidia using ROCM. Would that even be feasible? |
Beta Was this translation helpful? Give feedback.
-
https://github.com/YellowRoseCx/koboldcpp-rocm/releases/tag/v1.73.1.yr1-ROCm
Beta Was this translation helpful? Give feedback.
All reactions