Title. Mostly because of two flags: --read-only and --log-driver.

  • aksdb@feddit.de
    link
    fedilink
    arrow-up
    2
    ·
    1 year ago

    The point with an external drive is fine (I did that on my RPi as well), but the point with performance overhead due to containers is incorrect. The processes in the container run directly on the host. You even see the processes in ps. They are simply confined using cgroups to be isolated to different degrees.

    • sir_reginald@lemmy.world
      link
      fedilink
      arrow-up
      1
      arrow-down
      2
      ·
      1 year ago

      docker images have a ton of extra processes from the OS they were built in. Normally a light distro is used to build images, like Alpine Linux. but still, you’re executing a lot more processes than if you were installing things natively.

      Of course the images does not contain the kernel, but still they contain a lot of extra processes that would be unnecessary if executing natively.

      • aksdb@feddit.de
        link
        fedilink
        arrow-up
        2
        ·
        1 year ago

        To execute more than one process, you need to explicitly bring along some supervisor or use a more compicated entrypoint script that orchestrates this. But most container images have a simple entrypoint pointing to a single binary (or at most running a script to do some filesystem/permission setup and then run a single process).

        Containers running multiple processes are possible, but hard to pull off and therefore rarely used.

        What you likely think of are the files included in the images. Sure, some images bring more libs and executables along. But they are not started and/or running in the background (unless you explicitly start them as the entrypoint or using for example docker exec).