Getting GPU acceleration working is a common task for those of us running Plex or Jellyfin. There is not much documentation for getting the NVIDIA container stack to work with Podman, even less on Gentoo, plus there have been a lot of changes to NVIDIA’s container toolkit lately.

I have been fighting with Podman for a while now and just recently got it working 1:1 with my Docker setup. Gentoo may not be the most popular or easy to use distro but I documented it in case some poor soul runs across it searching the web.

Feel free to poke holes in it or leave feedback.

  • PriorProject@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    1 year ago

    This is a pretty awesome how-to. I knew nothing about containerizing GPU workloads before this, and it seems quite a lot less scary/involved than I feared.

    FWIW, I think some of your DNS and general networking woes may be due to the macvlan setup rather than using netavark. Netavark seems like the golden path going forward for a batteries-included experience. Not that I have anything against macvlan, in many ways macvlan feels simplest and nicest for homelab setups and I’ve used it with LXC and other container runtimes in the past. But for the most docker-like “it just works” experience, I feel like netavark is getting the upstream love.

    • moist_towelettes@lemm.eeOP
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      Thank you!

      Agree, I really wanted netavark to work - it definitely seems like the way forward. I enabled it and out of the box none of my containers could resolve DNS, even though aardvark was running. It’s so new I wasn’t sure where to poke around, so I went with the legacy method.

      I’ll try again once it stabilizes in Gentoo, somebody else noticed netavark should be the default now and opened a bug with the maintainer.

      • PriorProject@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        I enabled it and out of the box none of my containers could resolve DNS, even though aardvark was running.

        I experienced this on Ubuntu as well, and addressed it by opening up a firewall rule on the network interface for my podman network allowing the ip-range of the podman network to issue DNS requests to the gateway-ip (which is where aardvark-dns sets up shop).

        Also had to add a firewall rule to open whatever ports I exposed from all src-ips to the podman network range before exposing hostPorts would work.

        Again, not critiquing the very capable macvlan setup, just sharing tips I’ve picked up on making netavark work.

  • boonhet@lemm.ee
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    Thank you! Cross-posted this to my tiny Gentoo community at !gentoo@lemm.ee

    Feel free to share your Gentoo knowledge and experience there, it certainly could use more activity.

  • vluz@kbin.social
    link
    fedilink
    arrow-up
    1
    ·
    1 year ago

    Absolutely stellar write up. Thank you!

    I have a couple of questions.
    Imagine I have a powerful consumer gpu card to trow at this solution, 4090ti for the sake of example.
    - How many containers can share one physical card, taking into account total vram memory will not be exceeded?
    - How does one virtual gpu look like in the container? Can I run standard stuff like PyTorch, Tensorflow, and CUDA stuff in general?

    • moist_towelettes@lemm.eeOP
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 year ago

      Thanks!

      As I understand it, it bind-mounts the /dev/nvidia devices and the CUDA toolkit binaries inside the container, giving it direct access just as if it was running on the host. It’s not virtualized, just running under a different namespace so the VRAM is still being managed by the host driver. I would think the same restrictions exist in containers that would apply for running CUDA applications normally on the host. Personally I’ve had up to 4 containers run GPU processes at the same time on 1 card.

      And yes, Nvidia hosts it’s own GPU accelerated container images for PyTorch, Tensorflow and a bunch of others on the NGC. They also have images with the full CUDA SDK on their dockerhub.

      • vluz@kbin.social
        link
        fedilink
        arrow-up
        1
        ·
        1 year ago

        That’s wonderful to know! Thank you again.
        I’ll follow your instructions, this implementation is exactly what I was looking for.