Drivers and Containerization Platforms

Once you have provisioned a deployment environment with a Linux operating system installed, we need to configure it.

While some cloud providers will automatically install NVIDIA drivers for use with NVIDIA GPUs, many do not, so we will walk through how to install NVIDIA drivers for the GPUs and expose them for our use.

We will also step through installing a containerization platform. We highly recommend Docker, but you may also use Podman if you are using Red Hat Enterprise Linux (RHEL) version 8 or higher, or another similar distribution that does not ship or support Docker.


Other pages in Deepgram's documentation may exclusively list example commands using docker. If you are using a different containerization platform, such as podman, you may need to adjust the commands accordingly.


Make sure you have completed the steps in the AWS guide or Bare-Metal Servers guide.

Note on Different Linux Distributions

Various Linux distributions have a default or preferred package manager for the installation and management of system packages. For example, apt is associated with Ubuntu and dnf is associated with RHEL.

This guide will contain instructions that should be adaptable for many Linux distributions, but are specific to Ubuntu and RHEL. You will see comments above the commands and sections when there is a distribution-specific action. If there are no comments or headers above a set of instructions, it should work cross-platform.

Update System Package Manager

Update your server’s operating system package manager to get information on updated versions of packages and their dependencies, and upgrade these packages as needed.

# Ubuntu
sudo apt update
sudo apt upgrade -y
sudo dnf update -y
sudo dnf upgrade -y

Install GNU Toolchain Components

Install the GNU Compiler Collection (gcc) , GNU Make (make), and GNU Web Get (wget) tool:

# Ubuntu
sudo apt install -y gcc make wget
sudo dnf install -y gcc make wget

Remove Nouveau Drivers

The Nouveau kernel driver is incompatible with NVIDIA drivers, so you will need to disable it before installing any NVIDIA drivers.

  1. In your terminal, create a new configuration file at /etc/modprobe.d/blacklist-nouveau.conf to blacklist the Nouveau drivers.

    sudo sh -c 'printf "blacklist nouveau\noptions nouveau modeset=0\n" > /etc/modprobe.d/blacklist-nouveau.conf'
  2. Regenerate the kernel with the new conf file added:

    # Ubuntu
    sudo update-initramfs -u
    # RHEL
    sudo dracut --force
  3. Reboot the instance:

    sudo reboot
  4. Once the instance has finished rebooting, reconnect via SSH and verify that Nouveau has been removed:

    lsmod | grep nouveau

    If you see no output, Nouveau was successfully removed.

Install NVIDIA Drivers

Install Kernel Development Tools

Many Linux distributions require Linux kernel development tools to be installed to support installing the NVIDIA drivers.

# Ubuntu
sudo apt-get install -y linux-headers-`uname -r`
sudo dnf -y install kernel-devel-`uname -r` kernel-headers-`uname -r`
sudo dnf -y install`rpm -q --queryformat '%{VERSION}' redhat-release | cut -d. -f1`.noarch.rpm
sudo dnf -y install dkms

Download and install the official drivers


If you are using Google Cloud Platform and your VM instance has Secure Boot enabled, see the GCP documentation for details on how to sign the NVIDIA kernel modules.

If you are using Azure and your Ubuntu VM instance has Trusted Launch enabled, which also enables Secure Boot, see the Azure documentation for how to add a Machine Owner Key that will sign a key for the driver installation. Otherwise, during VM creation, you may opt for Standard security instead of Trusted Launch, in order to install the drivers through our standard method as documented on this page.

  1. We are going to identify the latest driver for the GPU you are using and retrieve its download URL by going to the NVIDIA Official Drivers.

  2. Select the product type. For cloud instances, this will often be Data Center/Tesla.

  3. Select the product series and product. You should know the exact GPU you are using if you provisioned it yourself in your own data-center. If you are using a cloud instance, you can lookup the VM instance type on your cloud console, and use your cloud provider's documentation to find the corresponding GPU for that instance type.

    1. The product series will the first letter of the GPU name. For example, the T4 is part of the T-series, and the A10 is part of the A-series.
  4. Select your operating system. For most users, like those on Ubuntu, this will be Linux 64-bit. If you are on RHEL, select the appropriate version instead.

  5. Finally, choose the Download Type (Production Branch), and choose the latest CUDA toolkit.

  6. Select Search and check that the correct driver is displayed, then select Download.

  7. Right-click Agree & Download, then copy the link to save the download URL to your clipboard.

  8. Download the latest driver for your GPU on your deployment environment:



    Be sure to replace the LINK_TO_LATEST_NVIDIA_GPU_DRIVER placeholder value with the URL to the latest driver for the GPU you are using.

  9. Install the drivers, with different instructions for standard Linux distros versus RHEL:

    # Ubuntu
    sudo {DOWNLOADED_FILE_NAME} --silent
    # RHEL
    sudo rpm -i DOWNLOADED_FILE_NAME
    sudo dnf clean all
    sudo dnf -y module install nvidia-driver:latest-dkms


    With the --silent install on Ubuntu and other non-RHEL distros, you will see warnings that are similar to the following (they can be ignored):

    WARNING: Ignoring CC version mismatch:
    The kernel was built with gcc (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0, GNU ld (GNU Binutils for Ubuntu) 2.34, but the current compiler version is cc (Ubuntu 9.4.0-1ubuntu1~20.04) 9.4.0.
    WARNING: nvidia-installer was forced to guess the X library path '/usr/lib64' and X module path '/usr/lib64/xorg/modules'; these paths were not queryable from the system.  If X fails to find the NVIDIA X driver module, please install the `pkg-config` utility and the X.Org SDK/development package for your distribution and reinstall the driver
  10. Test that the NVIDIA drivers are installed. The following command should produce output describing the available GPU:


Install Container Engine

For ease of use, Deepgram provides its products in container images, so you must make sure that you have installed the latest version of Docker Engine (or an alternative such as Podman) on all hosts.

Note that RHEL does not distribute Docker Engine, so you will need to use Podman for your container engine.

  1. Install the container engine.

    1. To install Docker, read Install Using the Repository in Docker’s documentation.
    2. To install Podman, read their installation instructions. RHEL 9 users can simply run sudo dnf install podman.
      1. If you are using Podman, other guides in the On-Prem documentation will contain commands using docker. Change all of these to use podman.
  2. It's possible to grant your user (e.g. ubuntu or ec2-user) sufficient permissions to run container engine commands without elevated privileges (without sudo).

    1. For Docker, see Manage Docker as a Non-Root User in Docker’s optional post-installation documentation.

    2. For Podman, the process to run container engine commands without elevated privileges is somewhat more involved. See this tutorial for basic setup and use of Podman in a rootless environment.


    If you do not follow step 2, you cannot run container engine commands without elevated privileges. You must run any docker, docker-compose, podman, or podman-compose commands with sudo.

Install Container Composition Tools

Container Composition tools allow users to define and manage multi-container applications using simple YAML configuration files that can be checked into source control. It enables the orchestration and coordination of services, automating the deployment, scaling, and management of containerized applications.


Docker Compose V2 is now included with Docker Engine. The plugin for CLI use should be installed with the Install Container Engine steps. If not, you can install it independently:

# Ubuntu
sudo apt install -y docker-compose-plugin

Test the installation:

docker compose version

You should expect the command output to return version 2.X.X.


The open source community maintains a podman-compose tool that seeks to be compatible with Docker Compose. You can install this with their instructions on GitHub, and test your installation:

podman-compose version

Install the NVIDIA Container Runtime

CUDA is NVIDIA's library for interacting with its GPU. CUDA support is made available to containers using the NVIDIA container toolkit.


nvidia-docker exposes the NVIDIA container toolkit for the Docker runtime. Follow the Docker instructions from NVIDIA to setup this runtime.

After you've setup the NVIDIA Docker runtime, you can test it with the following command:

docker run --runtime=nvidia --rm --gpus all nvidia/cuda:12.1.1-base-ubuntu22.04 nvidia-smi


Podman has implemented support for the Container Device Interface (CDI) standard in its container engine, which allows for direct use of the NVIDIA container toolkit. Follow the CDI Support instructions from NVIDIA to install and configure the toolkit.

After you've setup the NVIDIA container toolkit with CDI, you can test it with the following command:

 sudo podman run --rm --device ubuntu nvidia-smi


This guide walked you through installing the NVIDIA drivers to interact with our GPU that will run inference, as well as the containerization platform that we will use to run Deepgram services.

As a reminder, many of our guides assume use of Docker. If you are on Red Hat Enterprise Linux or have another reason to use Podman instead of Docker, keep in mind the commands and configuration may be slightly different.

What’s Next

Now we head to Deepgram Console to generate needed credentials for our deployment.