another AI question
Martin Cracauer
cracauer at cons.org
Wed Apr 8 12:16:29 EDT 2026
Martin Cracauer wrote on Tue, Apr 07, 2026 at 05:50:21PM -0400:
> The situation with LLMs on FreeBSD is not totally catastrophic.
>
> The NVidia drivers are currently broken on my 5090, so I cannot
> compare Vulkan/FreeBSD to Linux/Cuda.
Made them work, you need loader.conf
hw.nvidia.registry.EnableGpuFirmware=17
Performance on bartowski/Qwen_Qwen3.5-27B-GGUF:Q6_K_L in llama.cpp is:
- FreeBSD Vulkan 49 tokens/second
- Linux CUDA 56 tokens/second
Will get Linux/Vulkan numbers when I have a chance.
But this is encouraging. Windows was also 10% slower than Linux.
Martin
> But they work on my 2080ti with Vulkan and run both ollama and
> llama.cpp, accelerated.
>
> On my laptop with "AMD Ryzen 7 PRO 4750U with Radeon Graphics" also
> runs Vulkan and accelerates ollama (although only by a factor of 3
> compared to CPU). This combo does not run llama.cpp
>
> Now that NVidia drivers are running on at least one of my cards I'll
> give it another go to run CUDA through Linuxulator.
That go failed. No CUDA on Linuxulator still.
Martin
--
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
Martin Cracauer <cracauer at cons.org> http://www.cons.org/cracauer/
More information about the talk
mailing list