Docker Cheat Sheet
Table of Contents:
- Containers
- Images
- Networks
- Registry and Repository
- Dockerfile
- Layers
- Links
- Volumes
- Exposing Ports
- Best Practices
- Docker-Compose
- Security
- Tips
Containers
Your basic isolated Docker process. Containers are to Virtual Machines as threads are to processes. Or you can think of them as chroots on steroids.
Lifecycle
docker createcreates a container but does not start it.docker renameallows the container to be renamed.docker runcreates and starts a container in one operation.docker rmdeletes a container.docker updateupdates a container’s resource limits.
Normally if you run a container without options it will start and stop immediately, if you want keep it running you can use the command, docker run -td container_id this will use the option -t that will allocate a pseudo-TTY session and -d that will detach automatically the container (run container in background and print container ID).
If you want a transient container, docker run --rm will remove the container after it stops.
If you want to map a directory on the host to a docker container, docker run -v $HOSTDIR:$DOCKERDIR. Also see Volumes.
If you want to remove also the volumes associated with the container, the deletion of the container must include the -v switch like in docker rm -v.
There’s also a logging driver available for individual containers in docker 1.10. To run docker with a custom log driver (i.e., to syslog), use docker run --log-driver=syslog.
Another useful option is docker run --name yourname docker_image because when you specify the --name inside the run command this will allow you to start and stop a container by calling it with the name the you specified when you created it.
Starting and Stopping
docker startstarts a container so it is running.docker stopstops a running container.docker restartstops and starts a container.docker pausepauses a running container, “freezing” it in place.docker unpausewill unpause a running container.docker waitblocks until running container stops.docker killsends a SIGKILL to a running container.docker attachwill connect to a running container.
If you want to detach from a running container, use Ctrl + p, Ctrl + q.
If you want to integrate a container with a host process manager, start the daemon with -r=false then use docker start -a.
If you want to expose container ports through the host, see the exposing ports section.
Restart policies on crashed docker instances are covered here.
CPU Constraints
You can limit CPU, either using a percentage of all CPUs, or by using specific cores.
For example, you can tell the cpu-shares setting. The setting is a bit strange – 1024 means 100% of the CPU, so if you want the container to take 50% of all CPU cores, you should specify 512. See https://goldmann.pl/blog/2014/09/11/resource-management-in-docker/#_cpu for more:
|
|
You can also only use some CPU cores using cpuset-cpus. See https://agileek.github.io/docker/2014/08/06/docker-cpuset/ for details and some nice videos:
|
|
Note that Docker can still see all of the CPUs inside the container – it just isn’t using all of them. See https://github.com/docker/docker/issues/20770 for more details.
Memory Constraints
You can also set memory constraints on Docker:
|
|
Capabilities
Linux capabilities can be set by using cap-add and cap-drop. See https://docs.docker.com/engine/reference/run/#/runtime-privilege-and-linux-capabilities for details. This should be used for greater security.
To mount a FUSE based filesystem, you need to combine both –cap-add and –device:
|
|
Give access to a single device:
|
|
Give access to all devices:
|
|
More info about privileged containers here.
Info
docker psshows running containers.docker logsgets logs from container. (You can use a custom log driver, but logs is only available forjson-fileandjournaldin 1.10).docker inspectlooks at all the info on a container (including IP address).docker eventsgets events from container.docker portshows public facing port of container.docker topshows running processes in container.docker statsshows containers’ resource usage statistics.docker diffshows changed files in the container’s FS.
docker ps -a shows running and stopped containers.
docker stats --all shows a list of all containers, default shows just running.
Import / Export
docker cpcopies files or folders between a container and the local filesystem.docker exportturns container filesystem into tarball archive stream to STDOUT.
Executing Commands
docker execto execute a command in container.
To enter a running container, attach a new shell process to a running container called foo, use: docker exec -it foo /bin/bash.
Images
Images are just templates for docker containers.
Lifecycle
docker imagesshows all images.docker importcreates an image from a tarball.docker buildcreates image from Dockerfile.docker commitcreates image from a container, pausing it temporarily if it is running.docker rmiremoves an image.docker loadloads an image from a tar archive as STDIN, including images and tags (as of 0.7).docker savesaves an image to a tar archive stream to STDOUT with all parent layers, tags & versions (as of 0.7).
Info
docker historyshows history of image.docker tagtags an image to a name (local or registry).
Cleaning up
While you can use the docker rmi command to remove specific images, there’s a tool called docker-gc that will safely clean up images that are no longer used by any containers. As of docker 1.13, docker image prune is also available for removing unused images. See Prune.
Load/Save image
Load an image from file:
|
|
Save an existing image:
|
|
Import/Export container
Import a container as an image from file:
|
|
Export an existing container:
|
|
Difference between loading a saved image and importing an exported container as an image
Loading an image using the load command creates a new image including its history.
Importing a container as an image using the import command creates a new image excluding the history which results in a smaller image size compared to loading an image.
Networks
Docker has a networks feature. Docker automatically creates 3 network interfaces when you install it (bridge, host none). A new container is launched into the bridge network by default. To enable communication between multiple containers, you can create a new network and launch containers in it. This enables containers to communicate to each other while being isolated from containers that are not connected to the network. Furthermore, it allows to map container names to their IP addresses. See working with networks for more details.
Lifecycle
docker network createNAME Create a new network (default type: bridge).docker network rmNAME Remove one or more networks by name or identifier. No containers can be connected to the network when deleting it.
Info
docker network lsList networksdocker network inspectNAME Display detailed information on one or more networks.
Connection
docker network connectNETWORK CONTAINER Connect a container to a networkdocker network disconnectNETWORK CONTAINER Disconnect a container from a network
You can specify a specific IP address for a container:
|
|
Registry & Repository
A repository is a hosted collection of tagged images that together create the file system for a container.
A registry is a host – a server that stores repositories and provides an HTTP API for managing the uploading and downloading of repositories.
Docker.com hosts its own index to a central registry which contains a large number of repositories. Having said that, the central docker registry does not do a good job of verifying images and should be avoided if you’re worried about security.
docker loginto login to a registry.docker logoutto logout from a registry.docker searchsearches registry for image.docker pullpulls an image from registry to local machine.docker pushpushes an image to the registry from local machine.
Run local registry
You can run a local registry by using the docker distribution project and looking at the local deploy instructions.
Also see the mailing list.
Dockerfile
The configuration file. Sets up a Docker container when you run docker build on it. Vastly preferable to docker commit.
Here are some common text editors and their syntax highlighting modules you could use to create Dockerfiles:
- If you use jEdit, I’ve put up a syntax highlighting module for Dockerfile you can use.
- Sublime Text 2
- Atom
- Vim
- Emacs
- TextMate
- VS Code
- Also see Docker meets the IDE
Instructions
- .dockerignore
- FROM Sets the Base Image for subsequent instructions.
- MAINTAINER (deprecated - use LABEL instead) Set the Author field of the generated images.
- RUN execute any commands in a new layer on top of the current image and commit the results.
- CMD provide defaults for an executing container.
- EXPOSE informs Docker that the container listens on the specified network ports at runtime. NOTE: does not actually make ports accessible.
- ENV sets environment variable.
- ADD copies new files, directories or remote file to container. Invalidates caches. Avoid
ADDand useCOPYinstead. - COPY copies new files or directories to container. By default this copies as root regardless of the USER/WORKDIR settings. Use
--chown=<user>:<group>to give ownership to another user/group. (Same forADD.) - ENTRYPOINT configures a container that will run as an executable.
- VOLUME creates a mount point for externally mounted volumes or other containers.
- USER sets the user name for following RUN / CMD / ENTRYPOINT commands.
- WORKDIR sets the working directory.
- ARG defines a build-time variable.
- ONBUILD adds a trigger instruction when the image is used as the base for another build.
- STOPSIGNAL sets the system call signal that will be sent to the container to exit.
- LABEL apply key/value metadata to your images, containers, or daemons.
- SHELL override default shell is used by docker to run commands.
- HEALTHCHECK tells docker how to test a container to check that it is still working.
Tutorial
Examples
- Examples
- Best practices for writing Dockerfiles
- Michael Crosby has some more Dockerfiles best practices / take 2.
- Building Good Docker Images / Building Better Docker Images
- Managing Container Configuration with Metadata
- How to write excellent Dockerfiles
Layers
The versioned filesystem in Docker is based on layers. They’re like git commits or changesets for filesystems.
Links
Links are how Docker containers talk to each other through TCP/IP ports. Atlassian show worked examples. You can also resolve links by hostname.
This has been deprecated to some extent by user-defined networks.
NOTE: If you want containers to ONLY communicate with each other through links, start the docker daemon with -icc=false to disable inter process communication.
If you have a container with the name CONTAINER (specified by docker run --name CONTAINER) and in the Dockerfile, it has an exposed port:
EXPOSE 1337Then if we create another container called LINKED like so:
|
|
Then the exposed ports and aliases of CONTAINER will show up in LINKED with the following environment variables:
|
|
And you can connect to it that way.
To delete links, use docker rm --link.
Generally, linking between docker services is a subset of “service discovery”, a big problem if you’re planning to use Docker at scale in production. Please read The Docker Ecosystem: Service Discovery and Distributed Configuration Stores for more info.
Volumes
Docker volumes are free-floating filesystems. They don’t have to be connected to a particular container. You can use volumes mounted from data-only containers for portability. As of Docker 1.9.0, Docker has named volumes which replace data-only containers. Consider using named volumes to implement it rather than data containers.
Lifecycle
Info
Volumes are useful in situations where you can’t use links (which are TCP/IP only). For instance, if you need to have two docker instances communicate by leaving stuff on the filesystem.
You can mount them in several docker containers at once, using docker run --volumes-from.
Because volumes are isolated filesystems, they are often used to store state from computations between transient containers. That is, you can have a stateless and transient container run from a recipe, blow it away, and then have a second instance of the transient container pick up from where the last one left off.
See advanced volumes for more details. Container42 is also helpful.
You can map MacOS host directories as docker volumes:
|
|
You can use remote NFS volumes if you’re feeling brave.
You may also consider running data-only containers as described here to provide some data portability.
Be aware that you can mount files as volumes.
Exposing ports
Exposing incoming ports through the host container is fiddly but doable.
This is done by mapping the container port to the host port (only using localhost interface) using -p:
|
|
You can tell Docker that the container listens on the specified network ports at runtime by using EXPOSE:
|
|
Note that EXPOSE does not expose the port itself - only -p will do that.
To expose the container’s port on your localhost’s port, run:
|
|
If you’re running Docker in Virtualbox, you then need to forward the port there as well, using forwarded_port. Define a range of ports in your Vagrantfile like this so you can dynamically map them:
Vagrant.configure(VAGRANTFILE_API_VERSION) do |config|
...
(49000..49900).each do |port|
config.vm.network :forwarded_port, :host => port, :guest => port
end
...
endIf you forget what you mapped the port to on the host container, use docker port to show it:
|
|
Best Practices
This is where general Docker best practices and war stories go:
- The Rabbit Hole of Using Docker in Automated Tests
- Bridget Kromhout has a useful blog post on running Docker in production at Dramafever.
- There’s also a best practices blog post from Lyst.
- Building a Development Environment With Docker
- Discourse in a Docker Container
Docker-Compose
Compose is a tool for defining and running multi-container Docker applications. With Compose, you use a YAML file to configure your application’s services. Then, with a single command, you create and start all the services from your configuration. To learn more about all the features of Compose, see the list of features.
By using the following command you can start up your application:
|
|
You can also run docker-compose in detached mode using -d flag, then you can stop it whenever needed by the following command:
|
|
You can bring everything down, removing the containers entirely, with the down command. Pass --volumes to also remove the data volume.
Security
This is where security tips about Docker go. The Docker security page goes into more detail.
First things first: Docker runs as root. If you are in the docker group, you effectively have root access. If you expose the docker unix socket to a container, you are giving the container root access to the host.
Docker should not be your only defense. You should secure and harden it.
For an understanding of what containers leave exposed, you should read Understanding and Hardening Linux Containers by Aaron Grattafiori. This is a complete and comprehensive guide to the issues involved with containers, with a plethora of links and footnotes leading on to yet more useful content. The security tips following are useful if you’ve already hardened containers in the past, but are not a substitute for understanding.
Security Tips
For greatest security, you want to run Docker inside a virtual machine. This is straight from the Docker Security Team Lead – slides / notes. Then, run with AppArmor / seccomp / SELinux / grsec etc to limit the container permissions. See the Docker 1.10 security features for more details.
Docker image ids are sensitive information and should not be exposed to the outside world. Treat them like passwords.
See the Docker Security Cheat Sheet by Thomas Sjögren: some good stuff about container hardening in there.
Check out the docker bench security script, download the white papers.
Snyk’s 10 Docker Image Security Best Practices cheat sheet
You should start off by using a kernel with unstable patches for grsecurity / pax compiled in, such as Alpine Linux. If you are using grsecurity in production, you should spring for commercial support for the stable patches, same as you would do for RedHat. It’s $200 a month, which is nothing to your devops budget.
Since docker 1.11 you can easily limit the number of active processes running inside a container to prevent fork bombs. This requires a linux kernel >= 4.3 with CGROUP_PIDS=y to be in the kernel configuration.
docker run --pids-limit=64Also available since docker 1.11 is the ability to prevent processes from gaining new privileges. This feature have been in the linux kernel since version 3.5. You can read more about it in this blog post.
|
|
From the Docker Security Cheat Sheet (it’s in PDF which makes it hard to use, so copying below) by Container Solutions:
Turn off interprocess communication with:
|
|
Set the container to be read-only:
|
|
Verify images with a hashsum:
|
|
Set volumes to be read only:
|
|
Define and run a user in your Dockerfile so you don’t run as root inside the container:
|
|
User Namespaces
There’s also work on user namespaces – it is in 1.10 but is not enabled by default.
To enable user namespaces (“remap the userns”) in Ubuntu 15.10, follow the blog example.
Security Videos
- Using Docker Safely
- Securing your applications using Docker
- Container security: Do containers actually contain?
- Linux Containers: Future or Fantasy?
Security Roadmap
The Docker roadmap talks about seccomp support. There is an AppArmor policy generator called bane, and they’re working on security profiles.
Tips
Sources:
Prune
The new Data Management Commands have landed as of Docker 1.13:
docker system prunedocker volume prunedocker network prunedocker container prunedocker image prune
df
docker system df presents a summary of the space currently used by different docker objects.
Heredoc Docker Container
|
|
Last IDs
|
|
Commit with command (needs Dockerfile)
|
|
Get IP address
|
|
Or with jq installed:
|
|
Or using a go template:
|
|
Or when building an image from Dockerfile, when you want to pass in a build argument:
|
|
Get port mapping
|
|
Find containers by regular expression
|
|
Get Environment Settings
|
|
Kill running containers
|
|
Delete all containers (force!! running or stopped containers)
|
|
Delete old containers
|
|
Delete stopped containers
|
|
Delete containers after stopping
|
|
Delete dangling images
|
|
Delete all images
|
|
Delete dangling volumes
As of Docker 1.9:
|
|
In 1.9.0, the filter dangling=false does not work - it is ignored and will list all volumes.
Show image dependencies
|
|
Slimming down Docker containers
- Cleaning APT in a
RUNlayer - This should be done in the same layer as otheraptcommands. Otherwise, the previous layers still persist the original information and your images will still be fat.1 2 3RUN {apt commands} \ && apt-get clean \ && rm -rf /var/lib/apt/lists/* /tmp/* /var/tmp/* - Flatten an image
1 2ID=$(docker run -d image-name /bin/bash) docker export $ID | docker import – flat-image-name - For backup
1 2 3ID=$(docker run -d image-name /bin/bash) (docker export $ID | gzip -c > image.tgz) gzip -dc image.tgz | docker import - flat-image-name
Monitor system resource utilization for running containers
To check the CPU, memory, and network I/O usage of a single container, you can use:
|
|
For all containers listed by ID:
|
|
For all containers listed by name:
|
|
For all containers listed by image:
|
|
Remove all untagged images:
|
|
Remove container by a regular expression:
|
|
Remove all exited containers:
|
|
Volumes can be files
Be aware that you can mount files as volumes. For example you can inject a configuration file like this:
|
|