Anyone using "docker run" instead of "docker compose"?
For the vast majority of docker images, the documentation only mention a super long and hard to understand "docker run" one liner.
Why nobody is placing an example docker-compose.yml in their documentation? It's so tidy and easy to understand, also much easier to run in the future, just set and forget.
If every image had an yml to just copy, I could get it running in a few seconds, instead I have to decode the line to become an yml
I want to know if it's just me that I'm out of touch and should use "docker run" or it's just that an "one liner" looks much tidier in the docs. Like to say "hey just copy and paste this line to run the container. You don't understand what it does? Who cares"
The worst are the ones that are piping directly from curl to "sudo bash"...
I honestly don’t understand why anyone would make “curl and bash” the officially installation method these days, with docker around. Unless this is the ONLY thing you install on the system, so many things can go wrong.
Out of curiosity, is there much overhead to using docker than installing via curl and bash? I'm guessing there's some redundant layers that docker uses?
Of course, but the amount of overhead completely depends per container. The reason I am willing to accept the -in my experience- very small amount of overhead I typically get is that the repeatability is amazing with docker.
My first server was unRAID (freebsd, not Linux), I setup proxmox (debian with a webui) later. I took my unRAID server down for maintenance but wanted a certain service to stay up. So I copied a backup from unRAID to another server and had the service running in minutes. If it was a package, there is no guarantee that it would have been built for both OSes, both builds were the same version, or they used the same libraries.
My favorite way to extend the above is Docker Compose. I create a folder with a docker-compose.yml file and I can keep EVERYTHING for that service in a single folder. unRAID doesn't use Docker Compose in its webui. So, I try to stick to keeping things in Proxmox for ease of transfer and stuff.
I don't think you're out of touch, just use docker compose. It's not that hard to conver the docker run example command line into a neat docker-compose.yml, if they don't already provide one for you. So much better than just running containers manually.
Also, you should always understand what any command or docker compose file does before you run it! And don't blindly curl | bash either, download the bash script and look at it first.
Plain docker is useful when running some simple containers, or even one-off things. A lot of people thing about containers as long running services, but there's also many containers that are for running essentially a single command to completion and then shuts down.
There's also alternate ways to handle containers, for example Podman is typically used with systemd services as unlike Docker it doesn't work through a persistent daemon, so the configuration goes to a service.
I typically skip the docker-compose for simple containers, and turn to compose for either containers with loads of arguments or multi-container things.
Also switching between Docker and Podman depending on the machine and needs.
I've started replacing my docker compose files with pure ansible that is the equivilent of doing docker run. My ansible playbooks look almost exactly like my compose file but they can also create folders, set config files or cycle services when configs are updated.
It's been a bit of a learning process but it's replaced a lot what was previously documentation with code instead.
Wow, yeah this is exactly the sort of roles/playbooks that I've been building. I'm definitely using this as a source before starting my own from scratch. Thanks for sharing.
I'm actually doing both right now since I had quite a huge compose file that I haven't converted to ansible yet. The biggest frustration I have is that there doesn't seem to be an ansible module that works with compose v2 (the official plugin) which means I'm either stuck on the old version of compose or I have to use shell commands to run stuff like 'docker compose up -d'.
One nice thing I've gained though is for services like Plex. I have an 'update' playbook that I use and it will check to see if Plex is actively streaming before updating the container which isn't something I could do easily with compose.
I do this out of habit because this is how my work does it, but I honestly don't know the benefits of doing it this way. Can you explain (or provide a link?)
I've never tried Podman myself, but managing the containers using systemd would mean that you use exactly the same commands to start a Docker container as you would use to start a regular service. The fact that it's running in a container essentially just becomes an implementation detail, and you don't have to remember what's running in containers vs what's not running in containers.
Docker-compose is a orchestration tool that wraps around the inbuilt docker functions that are exposed like "docker run", when teaching people a tool you generally explain the base functions of the tool and then explain wrappers around that tool in terms of the functions you've already learned.
Similarly when you have a standalone container you generally provide the information to get the container running in terms of base docker, not an orchestration tool... unless the container must be used alongside other containers, then orchestration config is often provided.
Honestly I never really saw the point of it, just seems like another dependency. The compose file and the docket run commands have almost the same info. I'd rather jump to kubectl and skip compose entirely. I'd like to see a tool that can convert between these 3 formats for you. As for piping into bash, no - I'd only do it on a very trusted package.
I prefer to use ansible to define and provision my containers (docker/podman over containerd). For work, of course k8s and helm takes the cake. no reason to run k8s for personal self hosting, though.
I’m curious to hear from the runners. I use compose and I feel the same, it’s more readable and editable and it allows me to backup the command by backing up the docker-compose.yml
When orchestration or provisioning tools are used (Ansible, kurbernetes, etc...), creating networks and containers are equally readable in code. The way docker compose is designed makes it hard to integrate with these tools.
This is the response I was hoping to hear. I'm primarily a home-automation/self-hosted enthusiast, not necessarily a infrastructure enthusiast. As of yet, I haven't felt the need for using more involved orchestration tools/infra.
I always use docker-compose. It is very handy if you ever want to have a good backup or move the whole server to another. Copy over files -> docker compose up -d and you are done
For beginners, they should use docker compose from the start. Easier than docker run
If you ever want to convert those one-liner to a proper .yml then use this converter
First version of my server, I wrote a bunch of custom shell scripts to execute docker run statements to launch all my containers b/c I didn't know docker at all and didn't want to learn compose.
Current version of my server, I use docker compose. But all the containers I use come from linuxserver.io, and they always give examples for both. I use ansible to deploy everything.
I am not using docker-compose personally, and moving away from it at work, because it is only a CLI client and doesn't integrate with other tools except she'll scripts.
Previously my server was just a Debian box where I had a 'docker' directory with a bunch of .sh files containing 'docker run' commands (and a couple of docker-compose files for services that that have closely related containers). That works really well, it's easy to understand and manage. I had nginx running natively to expose stuff as necessary.
Recently I decided to try TrueNAS Scale (I wanted more reliable storage for my media library, which is large enough to be annoying to replace when a simple drive fails), and I'm still trying to figure it out. It's kind of a pain in the ass for running containers since the documentation is garbage. The web interface is kind of nice (other than constantly logging me out), but the learning curve for charts and exposing services has been tough, and it seems that ZFS is just a bad choice for Docker.
I was attracted to the idea of being able to run my services on my NAS server as one appliance, but it's feeling like TrueNAS Scale is way too complicated for home-scale (and way too primitive for commercial, not entirely sure what market they are aiming for) and I'm considering dumping it and setting up two servers, one for NAS and for running my containers and VMs.
I use docker to test individual container images. Anything long running is getting a Kubernetes manifest. I never use docker compose, except when supporting developers.
it turns out GPT converts plain docker commands into docker compose files well enough to me, it's been my go-to when I need to create a compose YAML. Checking a YAML and making one or two small corrections is even faster than entering all info in a form like Docker Compose Generator.
I'm sure someone has written a script to convert docker run commands to compose files.
I am usually customizing variables and tend to use compose for anything I am planning on running in "production". I'll use run if it's a temporary or on-demand use container.
It's not really that much effort to write a compose file with the variables from a run command, but you do have to keep an eye on formatting.