Set Up llama.cpp with GPU (RTX 5080) and Optimize for Mixtral – Test Q5_K_M vs Q4_K_M
I need a skilled developer to remotely configure my Windows PC for optimal use with the Mixtral model via llama.cpp. The goal is to install everything needed, build llama.cpp with full CUDA GPU acceleration... (Budget: $30 - $250 AUD, Jobs: C++ Programming, CUDA, Large Language Models (LLMs), Python, Windows Desktop)
