I never understood how to use Docker, what makes it so special? I would really like to use it on my Rapsberry Pi 3 Model B+ to ease the setup process of selfhosting different things.

I’m currently running these things without Docker:

  • Mumble server with a Discord bridge and a music bot
  • Maubot, a plugin-based Matrix bot
  • FTP server
  • Two Discord Music bots

All of these things are running as systemd services in the background. Should I change this? A lot of the things I’m hosting offer Docker images.

It would also be great if someone could give me a quick-start guide for Docker. Thanks in advance!

  • loudwhisper@infosec.pub
    link
    fedilink
    English
    arrow-up
    3
    ·
    9 months ago

    Most of the pro-Docker arguments go around security

    Actually Docker and the success of containers is mostly due to the ease of shipping code that carries its own dependencies and can be run anywhere. Security is a side-effect and definitely not the reason why containers picked-up.

    systemd can provide as much isolation a docker containers and 2) there are other container solutions that are at least as safe as Docker and nobody cares about them.

    Yes, and it’s much harder to achieve the same. In systemd you need to use 30 different options to get what using containers you achieve almost instantly and with much less hussle. I made an example on my blog where I decided to run blocky in Systemd and not in Docker. It’s just less convenient and accessible, harder to debug and also relies on each individual user to do it, while with containers a lot gets packed into the image and therefore harder to mess up.

    Docker isn’t totally proprietary

    There are a many container runtimes (CRI-O, podman, mirantis, containerd, etc.). Docker is just a convenient API, containers are fully implemented just with Linux native features (namespaces, seccomp, capabilities, cgroups) and images follow an open standard (OCI).

    I will avoid comment what looks like a rant, but I want to simply remind you that containers are the successor of VMs (virtualize everything!), platforms that were completely proprietary and in the hands of a handful of vendors, while containers use only native OS features and are therefore a step towards openness.

    • TCB13@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      edit-2
      9 months ago

      Docker and the success of containers is mostly due to the ease of shipping code that carries its own dependencies and can be run anywhere

      I don’t disagree with you, but that also shows that most modern software is poorly written. Usually a bunch of solutions that hardly work and nobody is able to reproduce their setup in a quick, sane and secure way.

      There are a many container runtimes (CRI-O, podman, mirantis, containerd, etc.). Docker is just a convenient API, containers are fully implemented just with Linux native features (namespaces, seccomp, capabilities, cgroups) and images follow an open standard (OCI).

      Yes, that’s exactly point point. There are many options, yet people stick with Docker and DockerHub (that is everything but open).

      In systemd you need to use 30 different options to get what using containers you achieve almost instantly and with much less hussle.

      Yes… maybe we just need some automation/orchestration tool for that. This is like saying that it’s way too hard to download the rootfs of some distro, unpack it and then use unshare to launch a shell on a isolated namespace… Docker as you said provides a convenient API but it doesn’t mean we can’t do the same for systemd.

      but I want to simply remind you that containers are the successor of VMs (virtualize everything!), platforms that were completely proprietary and in the hands of a handful of vendor

      Completely proprietary… like QEMU/libvirt? :P

      • towerful@programming.dev
        link
        fedilink
        English
        arrow-up
        0
        ·
        9 months ago

        I use ghcr, i have no issues pulling images from amazon ECR or wherever.
        Docker got there first with the adoption and marketing.

        Automation tools like ansible and terraform have existed for ages, and are great for running things without containers.
        OCI just makes it a hell of a lot easier and portable

      • loudwhisper@infosec.pub
        link
        fedilink
        English
        arrow-up
        1
        ·
        9 months ago

        but that also shows that most modern software is poorly written

        Does it? I mean, this is especially annoying with old software, maybe dynamically linked or PHP, or stuff like that. Modern tools (go, rust) don’t actually even have this problem. Dependencies are annoying in general, I don’t think it’s a property of modern software.

        Yes, that’s exactly point point. There are many options, yet people stick with Docker and DockerHub (that is everything but open).

        Who are these people? There are tons of registries that people use, github has its own, quay.io, etc. You also can simply publish Dockerfiles and people can build themselves. Ofc Docker has the edge because it was the first mainstream tool, and it’s still a great choice for single machine deployments, but it’s far from the only used. Kubernetes abandoned Docker as default runtime for years, for example… who are you referring to?

        Yes… maybe we just need some automation/orchestration tool for that. This is like saying that it’s way too hard to download the rootfs of some distro, unpack it and then use unshare to launch a shell on a isolated namespace… Docker as you said provides a convenient API but it doesn’t mean we can’t do the same for systemd.

        But Systemd also uses unshare, chroot, etc. They are at the same level of abstraction. Docker (and container runtimes) are simply specialized tools, while systemd is not. Why wouldn’t I use a tool that is meant for this when it’s available. I suppose bubblewrap does something similar too (used by Flatpak), and I am sure there are more.

        Completely proprietary… like QEMU/libvirt? :P

        Right, because organizations generally run QEMU, not VMware, Nutanix and another handful of proprietary platforms… :)

    • towerful@programming.dev
      link
      fedilink
      English
      arrow-up
      0
      ·
      9 months ago

      but I want to simply remind you that containers are the successor of VMs

      Successor implies replacement. I think containers are another tool in the toolkit of servers/hosting, but not a replacement for VMs

      • loudwhisper@infosec.pub
        link
        fedilink
        English
        arrow-up
        1
        ·
        9 months ago

        Well, I did not mean replacement (in fact, most orgs run in clouds which uses VMs) but I meant that a lot of orgs moved from VMs as the way to slice their compute to containers/kubernetes. Often the technologies are combined, so you are right.