AMD ROCm on Windows Issues

Troubleshooting common issues with AMD ROCm on Windows

Msty supports AMD ROCm on Windows out of the box and we even have a dedicated installer for it. However, sometimes things can go wrong and your GPU card might not be supported or detected at all. Here are some troubleshooting steps you can take to resolve the issue.

Get the latest version of Msty

We are constantly improving Msty and adding support for new GPUs. Make sure you have the latest version of Msty installed on your system. You can download the latest version from the Msty website.

Try a different installer

Instead of using the dedicated AMD ROCm on Windows installer, you can try using the generic Auto installer and see if that works for you. You can download the Auto installer from the Msty website. Give CPU only installer a try as well if nothing works.

Check if your GPU is supported

Not all AMD GPUs are supported by Msty (and Ollama, the underlying Local AI service ). Check here to see if your GPU is supported: Supported GPUs.

Check if your GPU is detected

cd into %appdata%/Msty from your terminal and run ./msty-local.exe serve to see if your GPU is detected and for other useful information.

Multiple GPUs

If you have multiple GPUs in your system, you can try passing main_gpu parameter to each model to specify which GPU to use for that model. This can be set globally for all models in Settings -> General Settings -> Local AI > Model Configurations > Advanced Configurations or per model in the model's configuration available when chatting with the model. This value can even be set per chat.

You can pass extra params to a model globally

You can pass extra params to a model globally

You can pass extra params (A) per chat (B) or per model (B)

You can pass extra params (A) per chat (B) or per model (B)

Custom Ollama Build

If you have an unsupported GPU, or if you just couldn't get Msty to work with your GPU, you can try using a custom build of Ollama:

  1. From https://github.com/likelovewant/ollama-for-amd/releases, download the latest Ollama AMD zipped file ollama-windows-amd64.7z.
  2. Replace %appdata%\Roaming\Msty\lib\ollama with the downloaded archive (unzip it first).
  3. Download the ROCBLAS packages for your GPU model from https://github.com/likelovewant/ROCmLibs-for-gfx1103-AMD780M-APU/releases.
  4. Rename %appdata%\Roaming\Msty\lib\ollama\rocblas\library to library.bck as a backup.
  5. Extract the library folder in the downloaded ROCBLAS packages to %appdata%\Roaming\Msty\lib\ollama\rocblas\ to replace the old library.
  6. Run ./msty-local.exe serve from %appdata%/Msty to see if your GPU is detected and for other useful information.
  7. Stop the server and restart Msty.