- Home
- Knowledge Base
- Raven Documentation
- Raven Workbench
- Raven Intelligence
- GPU Support for Raven Intelligence
GPU Support for Raven Intelligence
Starting in version 1.1, Raven Intelligence can run model Inference with GPU support for PyTorch and ONNX models. We currently can’t run Tensorflow models with GPU support. Here are steps to get the required dependencies in place for Windows.
Requirements
- A GPU that supports NVIDIA CUDA 12.4 or newer.
- Up-to-date drivers for your GPU.
1. Check Compatibility
Fist, verify your GPU can support the required CUDA versions. 12.8 is recommended. 12.4 may work, but we have done limited testing with it.
In a terminal, run the command nvidia-smi. This utility should be available if you have NVidia drivers installer. The supported CUDA version is in the top right corner. This needs to be 12.4 or greater. If it is not, Raven Intelligence will not be able to support it. Inference will still run fine using your CPU though.

2. Install NVIDIA CUDA
If your GPU supports 12.8 or newer, install NVIDIA CUDA 12.8.
If it supports 12.6 or 12.4, install NVIDIA CUDA 12.4.

3. Install NVIDIA CuDNN
Download and install NVIDIA CUDA Deep Neural Network library cuDNN version 9.20.0.
Make sure you are installing the ‘exe’ version, not ‘tarball’.


4. Install Microsoft Visual C++ Redistributable
Download and install the latest version.

5. RESTART your PC
!! Reboot your computer !!
We know, it’s annoying, but do it.
6. Make sure CUDA has access to ZLib
The CUDA library needs, but does not include, ZLib.
Download ZLib DLL and copy it to your CUDA bin directory, C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.8\bin or C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.4\bin
There are several places to find the DLL, we have placed a copy on our update server for convenience. If you get it from another source, make sure it is the x64 DLL.

Run Raven Intelligence
Raven intelligence will automatically download required support libraries so that it can interface with the NVIDIA CUDA tools. The first time you run a PyTorch model, this download will happen in the background. It may take several minutes depending on your internet speed.
You can verify that the correct libraries were downloaded by finding a hidden folder in your %USERPROFILE% directory, C:\Users\<username>\.djl.ai\pytorch
You should have a folder 2.7.1-cu128-win-x86_64. The cu128 indicates CUDA 12.8.

Not all models are able to run on GPU, but if the model you are working with can, you should see it reflected in the Windows Task Manager while running.
