Ollama K80

This project has been a roll-coaster and one of the hardest things I have attempted my goal when I set out was simple to run local Ollama models on a cheap £60 k80 GPU. My first attempted end very quickly when I found out that the GPU I got was faulty and it spewed out smoke within a second of first boot.

After I confirmed the rest of my server was unharmed I got a full refund and purchased another K80. This one did not fail on first boot and so I was able to start setting up my GPU. I spent the next month trying on and off failing every time.

I tired so many things I am not going to list them all here instead I will provided Instructions to try and recreate what I finally did:

  1. https://www.youtube.com/watch?v=_hOBAGKLQkI&t=616s this video from techno Tim walks through host setup – you also need to ensure you have above 4GB decoding.
  2. Setup a new VM with no ballooning, host cpu, q35 machine and the pcie device.
  3. Install Ubuntu 20.04
  4. sudo apt install -y nvidia-driver-470
  5. sudo apt install -y software-properties-common lsb-release
  6. wget -O - https://apt.kitware.com/keys/kitware-archive-latest.asc 2>/dev/null | gpg --dearmor - | sudo tee /etc/apt/trusted.gpg.d/kitware.gpg >/dev/null
  7. sudo apt-add-repository "deb https://apt.kitware.com/ubuntu/ $(lsb_release -cs) main"
  8. sudo apt update
  9. sudo apt install cmake
  10. sudo apt install gcc-10
  11. sudo apt install golang
  12. wget https://developer.download.nvidia.com/compute/cuda/11.4.0/local_installers/cuda_11.4.0_470.42.01_linux.run
  13. sudo sh cuda_11.4.0_470.42.01_linux.run –silent –toolkit –samples
  14. git clone https://github.com/idream3000/ollama37.git
  15. cd ollama37
  16. nano CMakeLists.txt
  17. change “native” on line 73 to “37”
  18. save and exit
  19. cmake -B build
  20. cmake –build buildy
  21. Environment=”OLLAMA_HOST=0.0.0.0:11434″
  22. Environment=”CUDA_VISIBLE_DEVICES=0,1″
  23. Environment=”OLLAMA_USE_CUDA=1″
  24. Environment=”PATH=/usr/local/cuda/bin:/usr/bin:/bin”
  25. Environment=”LD_LIBRARY_PATH=/usr/local/cuda/lib64:/usr/lib:/lib”
  26. go run . serve

Leave a Comment

Your email address will not be published. Required fields are marked *