mirror of https://github.com/docker/cli.git
now, with shiney markdown
Docker-DCO-1.1-Signed-off-by: Sven Dowideit <SvenDowideit@fosiki.com> (github: SvenDowideit)
This commit is contained in:
parent
d39c8aea47
commit
88a30886e5
|
@ -0,0 +1,510 @@
|
|||
page_title: Dockerfile Reference
|
||||
page_description: Dockerfiles use a simple DSL which allows you to automate the steps you would normally manually take to create an image.
|
||||
page_keywords: builder, docker, Dockerfile, automation, image creation
|
||||
|
||||
# Dockerfile Reference
|
||||
|
||||
**Docker can act as a builder** and read instructions from a text
|
||||
`Dockerfile` to automate the steps you would
|
||||
otherwise take manually to create an image. Executing
|
||||
`docker build` will run your steps and commit them
|
||||
along the way, giving you a final image.
|
||||
|
||||
## Usage
|
||||
|
||||
To [*build*](../commandline/cli/#cli-build) an image from a source
|
||||
repository, create a description file called `Dockerfile`
|
||||
at the root of your repository. This file will describe the
|
||||
steps to assemble the image.
|
||||
|
||||
Then call `docker build` with the path of your
|
||||
source repository as argument (for example, `.`):
|
||||
|
||||
> `sudo docker build .`
|
||||
|
||||
The path to the source repository defines where to find the *context* of
|
||||
the build. The build is run by the Docker daemon, not by the CLI, so the
|
||||
whole context must be transferred to the daemon. The Docker CLI reports
|
||||
"Uploading context" when the context is sent to the daemon.
|
||||
|
||||
You can specify a repository and tag at which to save the new image if
|
||||
the build succeeds:
|
||||
|
||||
> `sudo docker build -t shykes/myapp .`
|
||||
|
||||
The Docker daemon will run your steps one-by-one, committing the result
|
||||
to a new image if necessary, before finally outputting the ID of your
|
||||
new image. The Docker daemon will automatically clean up the context you
|
||||
sent.
|
||||
|
||||
Note that each instruction is run independently, and causes a new image
|
||||
to be created - so `RUN cd /tmp` will not have any
|
||||
effect on the next instructions.
|
||||
|
||||
Whenever possible, Docker will re-use the intermediate images,
|
||||
accelerating `docker build` significantly (indicated
|
||||
by `Using cache`):
|
||||
|
||||
$ docker build -t SvenDowideit/ambassador .
|
||||
Uploading context 10.24 kB
|
||||
Uploading context
|
||||
Step 1 : FROM docker-ut
|
||||
---> cbba202fe96b
|
||||
Step 2 : MAINTAINER SvenDowideit@home.org.au
|
||||
---> Using cache
|
||||
---> 51182097be13
|
||||
Step 3 : CMD env | grep _TCP= | sed 's/.*_PORT_\([0-9]*\)_TCP=tcp:\/\/\(.*\):\(.*\)/socat TCP4-LISTEN:\1,fork,reuseaddr TCP4:\2:\3 \&/' | sh && top
|
||||
---> Using cache
|
||||
---> 1a5ffc17324d
|
||||
Successfully built 1a5ffc17324d
|
||||
|
||||
When you’re done with your build, you’re ready to look into [*Pushing a
|
||||
repository to its
|
||||
registry*](../../use/workingwithrepository/#image-push).
|
||||
|
||||
## Format
|
||||
|
||||
Here is the format of the Dockerfile:
|
||||
|
||||
# Comment
|
||||
INSTRUCTION arguments
|
||||
|
||||
The Instruction is not case-sensitive, however convention is for them to
|
||||
be UPPERCASE in order to distinguish them from arguments more easily.
|
||||
|
||||
Docker evaluates the instructions in a Dockerfile in order. **The first
|
||||
instruction must be \`FROM\`** in order to specify the [*Base
|
||||
Image*](../../terms/image/#base-image-def) from which you are building.
|
||||
|
||||
Docker will treat lines that *begin* with `#` as a
|
||||
comment. A `#` marker anywhere else in the line will
|
||||
be treated as an argument. This allows statements like:
|
||||
|
||||
# Comment
|
||||
RUN echo 'we are running some # of cool things'
|
||||
|
||||
Here is the set of instructions you can use in a `Dockerfile`
|
||||
for building images.
|
||||
|
||||
## `FROM`
|
||||
|
||||
> `FROM <image>`
|
||||
|
||||
Or
|
||||
|
||||
> `FROM <image>:<tag>`
|
||||
|
||||
The `FROM` instruction sets the [*Base
|
||||
Image*](../../terms/image/#base-image-def) for subsequent instructions.
|
||||
As such, a valid Dockerfile must have `FROM` as its
|
||||
first instruction. The image can be any valid image – it is especially
|
||||
easy to start by **pulling an image** from the [*Public
|
||||
Repositories*](../../use/workingwithrepository/#using-public-repositories).
|
||||
|
||||
`FROM` must be the first non-comment instruction in
|
||||
the `Dockerfile`.
|
||||
|
||||
`FROM` can appear multiple times within a single
|
||||
Dockerfile in order to create multiple images. Simply make a note of the
|
||||
last image id output by the commit before each new `FROM`
|
||||
command.
|
||||
|
||||
If no `tag` is given to the `FROM`
|
||||
instruction, `latest` is assumed. If the
|
||||
used tag does not exist, an error will be returned.
|
||||
|
||||
## `MAINTAINER`
|
||||
|
||||
> `MAINTAINER <name>`
|
||||
|
||||
The `MAINTAINER` instruction allows you to set the
|
||||
*Author* field of the generated images.
|
||||
|
||||
## `RUN`
|
||||
|
||||
RUN has 2 forms:
|
||||
|
||||
- `RUN <command>` (the command is run in a shell -
|
||||
`/bin/sh -c`)
|
||||
- `RUN ["executable", "param1", "param2"]` (*exec*
|
||||
form)
|
||||
|
||||
The `RUN` instruction will execute any commands in a
|
||||
new layer on top of the current image and commit the results. The
|
||||
resulting committed image will be used for the next step in the
|
||||
Dockerfile.
|
||||
|
||||
Layering `RUN` instructions and generating commits
|
||||
conforms to the core concepts of Docker where commits are cheap and
|
||||
containers can be created from any point in an image’s history, much
|
||||
like source control.
|
||||
|
||||
The *exec* form makes it possible to avoid shell string munging, and to
|
||||
`RUN` commands using a base image that does not
|
||||
contain `/bin/sh`.
|
||||
|
||||
### Known Issues (RUN)
|
||||
|
||||
- [Issue 783](https://github.com/dotcloud/docker/issues/783) is about
|
||||
file permissions problems that can occur when using the AUFS file
|
||||
system. You might notice it during an attempt to `rm`
|
||||
a file, for example. The issue describes a workaround.
|
||||
- [Issue 2424](https://github.com/dotcloud/docker/issues/2424) Locale
|
||||
will not be set automatically.
|
||||
|
||||
## `CMD`
|
||||
|
||||
CMD has three forms:
|
||||
|
||||
- `CMD ["executable","param1","param2"]` (like an
|
||||
*exec*, preferred form)
|
||||
- `CMD ["param1","param2"]` (as *default
|
||||
parameters to ENTRYPOINT*)
|
||||
- `CMD command param1 param2` (as a *shell*)
|
||||
|
||||
There can only be one CMD in a Dockerfile. If you list more than one CMD
|
||||
then only the last CMD will take effect.
|
||||
|
||||
**The main purpose of a CMD is to provide defaults for an executing
|
||||
container.** These defaults can include an executable, or they can omit
|
||||
the executable, in which case you must specify an ENTRYPOINT as well.
|
||||
|
||||
When used in the shell or exec formats, the `CMD`
|
||||
instruction sets the command to be executed when running the image.
|
||||
|
||||
If you use the *shell* form of the CMD, then the `<command>`
|
||||
will execute in `/bin/sh -c`:
|
||||
|
||||
FROM ubuntu
|
||||
CMD echo "This is a test." | wc -
|
||||
|
||||
If you want to **run your** `<command>` **without a
|
||||
shell** then you must express the command as a JSON array and give the
|
||||
full path to the executable. **This array form is the preferred format
|
||||
of CMD.** Any additional parameters must be individually expressed as
|
||||
strings in the array:
|
||||
|
||||
FROM ubuntu
|
||||
CMD ["/usr/bin/wc","--help"]
|
||||
|
||||
If you would like your container to run the same executable every time,
|
||||
then you should consider using `ENTRYPOINT` in
|
||||
combination with `CMD`. See
|
||||
[*ENTRYPOINT*](#dockerfile-entrypoint).
|
||||
|
||||
If the user specifies arguments to `docker run` then
|
||||
they will override the default specified in CMD.
|
||||
|
||||
Note
|
||||
|
||||
Don’t confuse `RUN` with `CMD`.
|
||||
`RUN` actually runs a command and commits the
|
||||
result; `CMD` does not execute anything at build
|
||||
time, but specifies the intended command for the image.
|
||||
|
||||
## `EXPOSE`
|
||||
|
||||
> `EXPOSE <port> [<port>...]`
|
||||
|
||||
The `EXPOSE` instructions informs Docker that the
|
||||
container will listen on the specified network ports at runtime. Docker
|
||||
uses this information to interconnect containers using links (see
|
||||
[*links*](../../use/working_with_links_names/#working-with-links-names)),
|
||||
and to setup port redirection on the host system (see [*Redirect
|
||||
Ports*](../../use/port_redirection/#port-redirection)).
|
||||
|
||||
## `ENV`
|
||||
|
||||
> `ENV <key> <value>`
|
||||
|
||||
The `ENV` instruction sets the environment variable
|
||||
`<key>` to the value `<value>`.
|
||||
This value will be passed to all future `RUN`
|
||||
instructions. This is functionally equivalent to prefixing the command
|
||||
with `<key>=<value>`
|
||||
|
||||
The environment variables set using `ENV` will
|
||||
persist when a container is run from the resulting image. You can view
|
||||
the values using `docker inspect`, and change them
|
||||
using `docker run --env <key>=<value>`.
|
||||
|
||||
Note
|
||||
|
||||
One example where this can cause unexpected consequenses, is setting
|
||||
`ENV DEBIAN_FRONTEND noninteractive`. Which will
|
||||
persist when the container is run interactively; for example:
|
||||
`docker run -t -i image bash`
|
||||
|
||||
## `ADD`
|
||||
|
||||
> `ADD <src> <dest>`
|
||||
|
||||
The `ADD` instruction will copy new files from
|
||||
\<src\> and add them to the container’s filesystem at path
|
||||
`<dest>`.
|
||||
|
||||
`<src>` must be the path to a file or directory
|
||||
relative to the source directory being built (also called the *context*
|
||||
of the build) or a remote file URL.
|
||||
|
||||
`<dest>` is the absolute path to which the source
|
||||
will be copied inside the destination container.
|
||||
|
||||
All new files and directories are created with mode 0755, uid and gid 0.
|
||||
|
||||
Note
|
||||
|
||||
if you build using STDIN (`docker build - < somefile`
|
||||
.literal}), there is no build context, so the Dockerfile can only
|
||||
contain an URL based ADD statement.
|
||||
|
||||
Note
|
||||
|
||||
if your URL files are protected using authentication, you will need to
|
||||
use an `RUN wget` , `RUN curl`
|
||||
or other tool from within the container as ADD does not support
|
||||
authentication.
|
||||
|
||||
The copy obeys the following rules:
|
||||
|
||||
- The `<src>` path must be inside the *context* of
|
||||
the build; you cannot `ADD ../something /something`
|
||||
, because the first step of a `docker build`
|
||||
is to send the context directory (and subdirectories) to
|
||||
the docker daemon.
|
||||
|
||||
- If `<src>` is a URL and `<dest>`
|
||||
does not end with a trailing slash, then a file is
|
||||
downloaded from the URL and copied to `<dest>`.
|
||||
|
||||
- If `<src>` is a URL and `<dest>`
|
||||
does end with a trailing slash, then the filename is
|
||||
inferred from the URL and the file is downloaded to
|
||||
`<dest>/<filename>`. For instance,
|
||||
`ADD http://example.com/foobar /` would create
|
||||
the file `/foobar`. The URL must have a
|
||||
nontrivial path so that an appropriate filename can be discovered in
|
||||
this case (`http://example.com` will not work).
|
||||
|
||||
- If `<src>` is a directory, the entire directory
|
||||
is copied, including filesystem metadata.
|
||||
|
||||
- If `<src>` is a *local* tar archive in a
|
||||
recognized compression format (identity, gzip, bzip2 or xz) then it
|
||||
is unpacked as a directory. Resources from *remote* URLs are **not**
|
||||
decompressed.
|
||||
|
||||
When a directory is copied or unpacked, it has the same behavior as
|
||||
`tar -x`: the result is the union of
|
||||
|
||||
1. whatever existed at the destination path and
|
||||
2. the contents of the source tree,
|
||||
|
||||
with conflicts resolved in favor of "2." on a file-by-file basis.
|
||||
|
||||
- If `<src>` is any other kind of file, it is
|
||||
copied individually along with its metadata. In this case, if
|
||||
`<dest>` ends with a trailing slash
|
||||
`/`, it will be considered a directory and the
|
||||
contents of `<src>` will be written at
|
||||
`<dest>/base(<src>)`.
|
||||
|
||||
- If `<dest>` does not end with a trailing slash,
|
||||
it will be considered a regular file and the contents of
|
||||
`<src>` will be written at `<dest>`
|
||||
.
|
||||
|
||||
- If `<dest>` doesn’t exist, it is created along
|
||||
with all missing directories in its path.
|
||||
|
||||
## `ENTRYPOINT`
|
||||
|
||||
ENTRYPOINT has two forms:
|
||||
|
||||
- `ENTRYPOINT ["executable", "param1", "param2"]`
|
||||
(like an *exec*, preferred form)
|
||||
- `ENTRYPOINT command param1 param2` (as a
|
||||
*shell*)
|
||||
|
||||
There can only be one `ENTRYPOINT` in a Dockerfile.
|
||||
If you have more than one `ENTRYPOINT`, then only
|
||||
the last one in the Dockerfile will have an effect.
|
||||
|
||||
An `ENTRYPOINT` helps you to configure a container
|
||||
that you can run as an executable. That is, when you specify an
|
||||
`ENTRYPOINT`, then the whole container runs as if it
|
||||
was just that executable.
|
||||
|
||||
The `ENTRYPOINT` instruction adds an entry command
|
||||
that will **not** be overwritten when arguments are passed to
|
||||
`docker run`, unlike the behavior of `CMD`
|
||||
.literal}. This allows arguments to be passed to the entrypoint. i.e.
|
||||
`docker run <image> -d` will pass the "-d" argument
|
||||
to the ENTRYPOINT.
|
||||
|
||||
You can specify parameters either in the ENTRYPOINT JSON array (as in
|
||||
"like an exec" above), or by using a CMD statement. Parameters in the
|
||||
ENTRYPOINT will not be overridden by the `docker run`
|
||||
arguments, but parameters specified via CMD will be overridden
|
||||
by `docker run` arguments.
|
||||
|
||||
Like a `CMD`, you can specify a plain string for the
|
||||
ENTRYPOINT and it will execute in `/bin/sh -c`:
|
||||
|
||||
FROM ubuntu
|
||||
ENTRYPOINT wc -l -
|
||||
|
||||
For example, that Dockerfile’s image will *always* take stdin as input
|
||||
("-") and print the number of lines ("-l"). If you wanted to make this
|
||||
optional but default, you could use a CMD:
|
||||
|
||||
FROM ubuntu
|
||||
CMD ["-l", "-"]
|
||||
ENTRYPOINT ["/usr/bin/wc"]
|
||||
|
||||
## `VOLUME`
|
||||
|
||||
> `VOLUME ["/data"]`
|
||||
|
||||
The `VOLUME` instruction will create a mount point
|
||||
with the specified name and mark it as holding externally mounted
|
||||
volumes from native host or other containers. For more
|
||||
information/examples and mounting instructions via docker client, refer
|
||||
to [*Share Directories via
|
||||
Volumes*](../../use/working_with_volumes/#volume-def) documentation.
|
||||
|
||||
## `USER`
|
||||
|
||||
> `USER daemon`
|
||||
|
||||
The `USER` instruction sets the username or UID to
|
||||
use when running the image.
|
||||
|
||||
## `WORKDIR`
|
||||
|
||||
> `WORKDIR /path/to/workdir`
|
||||
|
||||
The `WORKDIR` instruction sets the working directory
|
||||
for the `RUN`, `CMD` and
|
||||
`ENTRYPOINT` Dockerfile commands that follow it.
|
||||
|
||||
It can be used multiple times in the one Dockerfile. If a relative path
|
||||
is provided, it will be relative to the path of the previous
|
||||
`WORKDIR` instruction. For example:
|
||||
|
||||
> WORKDIR /a WORKDIR b WORKDIR c RUN pwd
|
||||
|
||||
The output of the final `pwd` command in this
|
||||
Dockerfile would be `/a/b/c`.
|
||||
|
||||
## `ONBUILD`
|
||||
|
||||
> `ONBUILD [INSTRUCTION]`
|
||||
|
||||
The `ONBUILD` instruction adds to the image a
|
||||
"trigger" instruction to be executed at a later time, when the image is
|
||||
used as the base for another build. The trigger will be executed in the
|
||||
context of the downstream build, as if it had been inserted immediately
|
||||
after the *FROM* instruction in the downstream Dockerfile.
|
||||
|
||||
Any build instruction can be registered as a trigger.
|
||||
|
||||
This is useful if you are building an image which will be used as a base
|
||||
to build other images, for example an application build environment or a
|
||||
daemon which may be customized with user-specific configuration.
|
||||
|
||||
For example, if your image is a reusable python application builder, it
|
||||
will require application source code to be added in a particular
|
||||
directory, and it might require a build script to be called *after*
|
||||
that. You can’t just call *ADD* and *RUN* now, because you don’t yet
|
||||
have access to the application source code, and it will be different for
|
||||
each application build. You could simply provide application developers
|
||||
with a boilerplate Dockerfile to copy-paste into their application, but
|
||||
that is inefficient, error-prone and difficult to update because it
|
||||
mixes with application-specific code.
|
||||
|
||||
The solution is to use *ONBUILD* to register in advance instructions to
|
||||
run later, during the next build stage.
|
||||
|
||||
Here’s how it works:
|
||||
|
||||
1. When it encounters an *ONBUILD* instruction, the builder adds a
|
||||
trigger to the metadata of the image being built. The instruction
|
||||
does not otherwise affect the current build.
|
||||
2. At the end of the build, a list of all triggers is stored in the
|
||||
image manifest, under the key *OnBuild*. They can be inspected with
|
||||
*docker inspect*.
|
||||
3. Later the image may be used as a base for a new build, using the
|
||||
*FROM* instruction. As part of processing the *FROM* instruction,
|
||||
the downstream builder looks for *ONBUILD* triggers, and executes
|
||||
them in the same order they were registered. If any of the triggers
|
||||
fail, the *FROM* instruction is aborted which in turn causes the
|
||||
build to fail. If all triggers succeed, the FROM instruction
|
||||
completes and the build continues as usual.
|
||||
4. Triggers are cleared from the final image after being executed. In
|
||||
other words they are not inherited by "grand-children" builds.
|
||||
|
||||
For example you might add something like this:
|
||||
|
||||
[...]
|
||||
ONBUILD ADD . /app/src
|
||||
ONBUILD RUN /usr/local/bin/python-build --dir /app/src
|
||||
[...]
|
||||
|
||||
Warning
|
||||
|
||||
Chaining ONBUILD instructions using ONBUILD ONBUILD isn’t allowed.
|
||||
|
||||
Warning
|
||||
|
||||
ONBUILD may not trigger FROM or MAINTAINER instructions.
|
||||
|
||||
## Dockerfile Examples
|
||||
|
||||
# Nginx
|
||||
#
|
||||
# VERSION 0.0.1
|
||||
|
||||
FROM ubuntu
|
||||
MAINTAINER Guillaume J. Charmes <guillaume@docker.com>
|
||||
|
||||
# make sure the package repository is up to date
|
||||
RUN echo "deb http://archive.ubuntu.com/ubuntu precise main universe" > /etc/apt/sources.list
|
||||
RUN apt-get update
|
||||
|
||||
RUN apt-get install -y inotify-tools nginx apache2 openssh-server
|
||||
|
||||
# Firefox over VNC
|
||||
#
|
||||
# VERSION 0.3
|
||||
|
||||
FROM ubuntu
|
||||
# make sure the package repository is up to date
|
||||
RUN echo "deb http://archive.ubuntu.com/ubuntu precise main universe" > /etc/apt/sources.list
|
||||
RUN apt-get update
|
||||
|
||||
# Install vnc, xvfb in order to create a 'fake' display and firefox
|
||||
RUN apt-get install -y x11vnc xvfb firefox
|
||||
RUN mkdir /.vnc
|
||||
# Setup a password
|
||||
RUN x11vnc -storepasswd 1234 ~/.vnc/passwd
|
||||
# Autostart firefox (might not be the best way, but it does the trick)
|
||||
RUN bash -c 'echo "firefox" >> /.bashrc'
|
||||
|
||||
EXPOSE 5900
|
||||
CMD ["x11vnc", "-forever", "-usepw", "-create"]
|
||||
|
||||
# Multiple images example
|
||||
#
|
||||
# VERSION 0.1
|
||||
|
||||
FROM ubuntu
|
||||
RUN echo foo > bar
|
||||
# Will output something like ===> 907ad6c2736f
|
||||
|
||||
FROM ubuntu
|
||||
RUN echo moo > oink
|
||||
# Will output something like ===> 695d7793cbe4
|
||||
|
||||
# You'll now have two images, 907ad6c2736f with /bar, and 695d7793cbe4 with
|
||||
# /oink.
|
File diff suppressed because it is too large
Load Diff
|
@ -0,0 +1,422 @@
|
|||
page_title: Docker Run Reference
|
||||
page_description: Configure containers at runtime
|
||||
page_keywords: docker, run, configure, runtime
|
||||
|
||||
# [Docker Run Reference](#id2)
|
||||
|
||||
**Docker runs processes in isolated containers**. When an operator
|
||||
executes `docker run`, she starts a process with its
|
||||
own file system, its own networking, and its own isolated process tree.
|
||||
The [*Image*](../../terms/image/#image-def) which starts the process may
|
||||
define defaults related to the binary to run, the networking to expose,
|
||||
and more, but `docker run` gives final control to
|
||||
the operator who starts the container from the image. That’s the main
|
||||
reason [*run*](../commandline/cli/#cli-run) has more options than any
|
||||
other `docker` command.
|
||||
|
||||
Every one of the [*Examples*](../../examples/#example-list) shows
|
||||
running containers, and so here we try to give more in-depth guidance.
|
||||
|
||||
## [General Form](#id3)
|
||||
|
||||
As you’ve seen in the [*Examples*](../../examples/#example-list), the
|
||||
basic run command takes this form:
|
||||
|
||||
docker run [OPTIONS] IMAGE[:TAG] [COMMAND] [ARG...]
|
||||
|
||||
To learn how to interpret the types of `[OPTIONS]`,
|
||||
see [*Option types*](../commandline/cli/#cli-options).
|
||||
|
||||
The list of `[OPTIONS]` breaks down into two groups:
|
||||
|
||||
1. Settings exclusive to operators, including:
|
||||
- Detached or Foreground running,
|
||||
- Container Identification,
|
||||
- Network settings, and
|
||||
- Runtime Constraints on CPU and Memory
|
||||
- Privileges and LXC Configuration
|
||||
|
||||
2. Setting shared between operators and developers, where operators can
|
||||
override defaults developers set in images at build time.
|
||||
|
||||
Together, the `docker run [OPTIONS]` give complete
|
||||
control over runtime behavior to the operator, allowing them to override
|
||||
all defaults set by the developer during `docker build`
|
||||
and nearly all the defaults set by the Docker runtime itself.
|
||||
|
||||
## [Operator Exclusive Options](#id4)
|
||||
|
||||
Only the operator (the person executing `docker run`
|
||||
.literal}) can set the following options.
|
||||
|
||||
- [Detached vs Foreground](#detached-vs-foreground)
|
||||
- [Detached (-d)](#detached-d)
|
||||
- [Foreground](#foreground)
|
||||
- [Container Identification](#container-identification)
|
||||
- [Name (–name)](#name-name)
|
||||
- [PID Equivalent](#pid-equivalent)
|
||||
- [Network Settings](#network-settings)
|
||||
- [Clean Up (–rm)](#clean-up-rm)
|
||||
- [Runtime Constraints on CPU and
|
||||
Memory](#runtime-constraints-on-cpu-and-memory)
|
||||
- [Runtime Privilege and LXC
|
||||
Configuration](#runtime-privilege-and-lxc-configuration)
|
||||
|
||||
### [Detached vs Foreground](#id2)
|
||||
|
||||
When starting a Docker container, you must first decide if you want to
|
||||
run the container in the background in a "detached" mode or in the
|
||||
default foreground mode:
|
||||
|
||||
-d=false: Detached mode: Run container in the background, print new container id
|
||||
|
||||
#### [Detached (-d)](#id3)
|
||||
|
||||
In detached mode (`-d=true` or just `-d`
|
||||
.literal}), all I/O should be done through network connections or shared
|
||||
volumes because the container is no longer listening to the commandline
|
||||
where you executed `docker run`. You can reattach to
|
||||
a detached container with `docker`
|
||||
[*attach*](../commandline/cli/#cli-attach). If you choose to run a
|
||||
container in the detached mode, then you cannot use the `--rm`
|
||||
option.
|
||||
|
||||
#### [Foreground](#id4)
|
||||
|
||||
In foreground mode (the default when `-d` is not
|
||||
specified), `docker run` can start the process in
|
||||
the container and attach the console to the process’s standard input,
|
||||
output, and standard error. It can even pretend to be a TTY (this is
|
||||
what most commandline executables expect) and pass along signals. All of
|
||||
that is configurable:
|
||||
|
||||
-a=[] : Attach to ``stdin``, ``stdout`` and/or ``stderr``
|
||||
-t=false : Allocate a pseudo-tty
|
||||
--sig-proxy=true: Proxify all received signal to the process (even in non-tty mode)
|
||||
-i=false : Keep STDIN open even if not attached
|
||||
|
||||
If you do not specify `-a` then Docker will [attach
|
||||
everything
|
||||
(stdin,stdout,stderr)](https://github.com/dotcloud/docker/blob/75a7f4d90cde0295bcfb7213004abce8d4779b75/commands.go#L1797).
|
||||
You can specify to which of the three standard streams
|
||||
(`stdin`, `stdout`,
|
||||
`stderr`) you’d like to connect instead, as in:
|
||||
|
||||
docker run -a stdin -a stdout -i -t ubuntu /bin/bash
|
||||
|
||||
For interactive processes (like a shell) you will typically want a tty
|
||||
as well as persistent standard input (`stdin`), so
|
||||
you’ll use `-i -t` together in most interactive
|
||||
cases.
|
||||
|
||||
### [Container Identification](#id5)
|
||||
|
||||
#### [Name (–name)](#id6)
|
||||
|
||||
The operator can identify a container in three ways:
|
||||
|
||||
- UUID long identifier
|
||||
("f78375b1c487e03c9438c729345e54db9d20cfa2ac1fc3494b6eb60872e74778")
|
||||
- UUID short identifier ("f78375b1c487")
|
||||
- Name ("evil\_ptolemy")
|
||||
|
||||
The UUID identifiers come from the Docker daemon, and if you do not
|
||||
assign a name to the container with `--name` then
|
||||
the daemon will also generate a random string name too. The name can
|
||||
become a handy way to add meaning to a container since you can use this
|
||||
name when defining
|
||||
[*links*](../../use/working_with_links_names/#working-with-links-names)
|
||||
(or any other place you need to identify a container). This works for
|
||||
both background and foreground Docker containers.
|
||||
|
||||
#### [PID Equivalent](#id7)
|
||||
|
||||
And finally, to help with automation, you can have Docker write the
|
||||
container ID out to a file of your choosing. This is similar to how some
|
||||
programs might write out their process ID to a file (you’ve seen them as
|
||||
PID files):
|
||||
|
||||
--cidfile="": Write the container ID to the file
|
||||
|
||||
### [Network Settings](#id8)
|
||||
|
||||
-n=true : Enable networking for this container
|
||||
--dns=[] : Set custom dns servers for the container
|
||||
|
||||
By default, all containers have networking enabled and they can make any
|
||||
outgoing connections. The operator can completely disable networking
|
||||
with `docker run -n` which disables all incoming and
|
||||
outgoing networking. In cases like this, you would perform I/O through
|
||||
files or STDIN/STDOUT only.
|
||||
|
||||
Your container will use the same DNS servers as the host by default, but
|
||||
you can override this with `--dns`.
|
||||
|
||||
### [Clean Up (–rm)](#id9)
|
||||
|
||||
By default a container’s file system persists even after the container
|
||||
exits. This makes debugging a lot easier (since you can inspect the
|
||||
final state) and you retain all your data by default. But if you are
|
||||
running short-term **foreground** processes, these container file
|
||||
systems can really pile up. If instead you’d like Docker to
|
||||
**automatically clean up the container and remove the file system when
|
||||
the container exits**, you can add the `--rm` flag:
|
||||
|
||||
--rm=false: Automatically remove the container when it exits (incompatible with -d)
|
||||
|
||||
### [Runtime Constraints on CPU and Memory](#id10)
|
||||
|
||||
The operator can also adjust the performance parameters of the
|
||||
container:
|
||||
|
||||
-m="": Memory limit (format: <number><optional unit>, where unit = b, k, m or g)
|
||||
-c=0 : CPU shares (relative weight)
|
||||
|
||||
The operator can constrain the memory available to a container easily
|
||||
with `docker run -m`. If the host supports swap
|
||||
memory, then the `-m` memory setting can be larger
|
||||
than physical RAM.
|
||||
|
||||
Similarly the operator can increase the priority of this container with
|
||||
the `-c` option. By default, all containers run at
|
||||
the same priority and get the same proportion of CPU cycles, but you can
|
||||
tell the kernel to give more shares of CPU time to one or more
|
||||
containers when you start them via Docker.
|
||||
|
||||
### [Runtime Privilege and LXC Configuration](#id11)
|
||||
|
||||
--privileged=false: Give extended privileges to this container
|
||||
--lxc-conf=[]: (lxc exec-driver only) Add custom lxc options --lxc-conf="lxc.cgroup.cpuset.cpus = 0,1"
|
||||
|
||||
By default, Docker containers are "unprivileged" and cannot, for
|
||||
example, run a Docker daemon inside a Docker container. This is because
|
||||
by default a container is not allowed to access any devices, but a
|
||||
"privileged" container is given access to all devices (see
|
||||
[lxc-template.go](https://github.com/dotcloud/docker/blob/master/execdriver/lxc/lxc_template.go)
|
||||
and documentation on [cgroups
|
||||
devices](https://www.kernel.org/doc/Documentation/cgroups/devices.txt)).
|
||||
|
||||
When the operator executes `docker run --privileged`
|
||||
.literal}, Docker will enable to access to all devices on the host as
|
||||
well as set some configuration in AppArmor to allow the container nearly
|
||||
all the same access to the host as processes running outside containers
|
||||
on the host. Additional information about running with
|
||||
`--privileged` is available on the [Docker
|
||||
Blog](http://blog.docker.io/2013/09/docker-can-now-run-within-docker/).
|
||||
|
||||
If the Docker daemon was started using the `lxc`
|
||||
exec-driver (`docker -d --exec-driver=lxc`) then the
|
||||
operator can also specify LXC options using one or more
|
||||
`--lxc-conf` parameters. These can be new parameters
|
||||
or override existing parameters from the
|
||||
[lxc-template.go](https://github.com/dotcloud/docker/blob/master/execdriver/lxc/lxc_template.go).
|
||||
Note that in the future, a given host’s Docker daemon may not use LXC,
|
||||
so this is an implementation-specific configuration meant for operators
|
||||
already familiar with using LXC directly.
|
||||
|
||||
## Overriding `Dockerfile` Image Defaults
|
||||
|
||||
When a developer builds an image from a
|
||||
[*Dockerfile*](../builder/#dockerbuilder) or when she commits it, the
|
||||
developer can set a number of default parameters that take effect when
|
||||
the image starts up as a container.
|
||||
|
||||
Four of the `Dockerfile` commands cannot be
|
||||
overridden at runtime: `FROM, MAINTAINER, RUN`, and
|
||||
`ADD`. Everything else has a corresponding override
|
||||
in `docker run`. We’ll go through what the developer
|
||||
might have set in each `Dockerfile` instruction and
|
||||
how the operator can override that setting.
|
||||
|
||||
- [CMD (Default Command or Options)](#cmd-default-command-or-options)
|
||||
- [ENTRYPOINT (Default Command to Execute at
|
||||
Runtime](#entrypoint-default-command-to-execute-at-runtime)
|
||||
- [EXPOSE (Incoming Ports)](#expose-incoming-ports)
|
||||
- [ENV (Environment Variables)](#env-environment-variables)
|
||||
- [VOLUME (Shared Filesystems)](#volume-shared-filesystems)
|
||||
- [USER](#user)
|
||||
- [WORKDIR](#workdir)
|
||||
|
||||
### [CMD (Default Command or Options)](#id12)
|
||||
|
||||
Recall the optional `COMMAND` in the Docker
|
||||
commandline:
|
||||
|
||||
docker run [OPTIONS] IMAGE[:TAG] [COMMAND] [ARG...]
|
||||
|
||||
This command is optional because the person who created the
|
||||
`IMAGE` may have already provided a default
|
||||
`COMMAND` using the `Dockerfile`
|
||||
`CMD`. As the operator (the person running a
|
||||
container from the image), you can override that `CMD`
|
||||
just by specifying a new `COMMAND`.
|
||||
|
||||
If the image also specifies an `ENTRYPOINT` then the
|
||||
`CMD` or `COMMAND` get appended
|
||||
as arguments to the `ENTRYPOINT`.
|
||||
|
||||
### [ENTRYPOINT (Default Command to Execute at Runtime](#id13)
|
||||
|
||||
--entrypoint="": Overwrite the default entrypoint set by the image
|
||||
|
||||
The ENTRYPOINT of an image is similar to a `COMMAND`
|
||||
because it specifies what executable to run when the container starts,
|
||||
but it is (purposely) more difficult to override. The
|
||||
`ENTRYPOINT` gives a container its default nature or
|
||||
behavior, so that when you set an `ENTRYPOINT` you
|
||||
can run the container *as if it were that binary*, complete with default
|
||||
options, and you can pass in more options via the `COMMAND`
|
||||
.literal}. But, sometimes an operator may want to run something else
|
||||
inside the container, so you can override the default
|
||||
`ENTRYPOINT` at runtime by using a string to specify
|
||||
the new `ENTRYPOINT`. Here is an example of how to
|
||||
run a shell in a container that has been set up to automatically run
|
||||
something else (like `/usr/bin/redis-server`):
|
||||
|
||||
docker run -i -t --entrypoint /bin/bash example/redis
|
||||
|
||||
or two examples of how to pass more parameters to that ENTRYPOINT:
|
||||
|
||||
docker run -i -t --entrypoint /bin/bash example/redis -c ls -l
|
||||
docker run -i -t --entrypoint /usr/bin/redis-cli example/redis --help
|
||||
|
||||
### [EXPOSE (Incoming Ports)](#id14)
|
||||
|
||||
The `Dockerfile` doesn’t give much control over
|
||||
networking, only providing the `EXPOSE` instruction
|
||||
to give a hint to the operator about what incoming ports might provide
|
||||
services. The following options work with or override the
|
||||
`Dockerfile`‘s exposed defaults:
|
||||
|
||||
--expose=[]: Expose a port from the container
|
||||
without publishing it to your host
|
||||
-P=false : Publish all exposed ports to the host interfaces
|
||||
-p=[] : Publish a container's port to the host (format:
|
||||
ip:hostPort:containerPort | ip::containerPort |
|
||||
hostPort:containerPort)
|
||||
(use 'docker port' to see the actual mapping)
|
||||
--link="" : Add link to another container (name:alias)
|
||||
|
||||
As mentioned previously, `EXPOSE` (and
|
||||
`--expose`) make a port available **in** a container
|
||||
for incoming connections. The port number on the inside of the container
|
||||
(where the service listens) does not need to be the same number as the
|
||||
port exposed on the outside of the container (where clients connect), so
|
||||
inside the container you might have an HTTP service listening on port 80
|
||||
(and so you `EXPOSE 80` in the
|
||||
`Dockerfile`), but outside the container the port
|
||||
might be 42800.
|
||||
|
||||
To help a new client container reach the server container’s internal
|
||||
port operator `--expose`‘d by the operator or
|
||||
`EXPOSE`‘d by the developer, the operator has three
|
||||
choices: start the server container with `-P` or
|
||||
`-p,` or start the client container with
|
||||
`--link`.
|
||||
|
||||
If the operator uses `-P` or `-p`
|
||||
then Docker will make the exposed port accessible on the host
|
||||
and the ports will be available to any client that can reach the host.
|
||||
To find the map between the host ports and the exposed ports, use
|
||||
`docker port`)
|
||||
|
||||
If the operator uses `--link` when starting the new
|
||||
client container, then the client container can access the exposed port
|
||||
via a private networking interface. Docker will set some environment
|
||||
variables in the client container to help indicate which interface and
|
||||
port to use.
|
||||
|
||||
### [ENV (Environment Variables)](#id15)
|
||||
|
||||
The operator can **set any environment variable** in the container by
|
||||
using one or more `-e` flags, even overriding those
|
||||
already defined by the developer with a Dockefile `ENV`
|
||||
.literal}:
|
||||
|
||||
$ docker run -e "deep=purple" --rm ubuntu /bin/bash -c export
|
||||
declare -x HOME="/"
|
||||
declare -x HOSTNAME="85bc26a0e200"
|
||||
declare -x OLDPWD
|
||||
declare -x PATH="/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
|
||||
declare -x PWD="/"
|
||||
declare -x SHLVL="1"
|
||||
declare -x container="lxc"
|
||||
declare -x deep="purple"
|
||||
|
||||
Similarly the operator can set the **hostname** with `-h`
|
||||
.literal}.
|
||||
|
||||
`--link name:alias` also sets environment variables,
|
||||
using the *alias* string to define environment variables within the
|
||||
container that give the IP and PORT information for connecting to the
|
||||
service container. Let’s imagine we have a container running Redis:
|
||||
|
||||
# Start the service container, named redis-name
|
||||
$ docker run -d --name redis-name dockerfiles/redis
|
||||
4241164edf6f5aca5b0e9e4c9eccd899b0b8080c64c0cd26efe02166c73208f3
|
||||
|
||||
# The redis-name container exposed port 6379
|
||||
$ docker ps
|
||||
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
|
||||
4241164edf6f dockerfiles/redis:latest /redis-stable/src/re 5 seconds ago Up 4 seconds 6379/tcp redis-name
|
||||
|
||||
# Note that there are no public ports exposed since we didn't use -p or -P
|
||||
$ docker port 4241164edf6f 6379
|
||||
2014/01/25 00:55:38 Error: No public port '6379' published for 4241164edf6f
|
||||
|
||||
Yet we can get information about the Redis container’s exposed ports
|
||||
with `--link`. Choose an alias that will form a
|
||||
valid environment variable!
|
||||
|
||||
$ docker run --rm --link redis-name:redis_alias --entrypoint /bin/bash dockerfiles/redis -c export
|
||||
declare -x HOME="/"
|
||||
declare -x HOSTNAME="acda7f7b1cdc"
|
||||
declare -x OLDPWD
|
||||
declare -x PATH="/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
|
||||
declare -x PWD="/"
|
||||
declare -x REDIS_ALIAS_NAME="/distracted_wright/redis"
|
||||
declare -x REDIS_ALIAS_PORT="tcp://172.17.0.32:6379"
|
||||
declare -x REDIS_ALIAS_PORT_6379_TCP="tcp://172.17.0.32:6379"
|
||||
declare -x REDIS_ALIAS_PORT_6379_TCP_ADDR="172.17.0.32"
|
||||
declare -x REDIS_ALIAS_PORT_6379_TCP_PORT="6379"
|
||||
declare -x REDIS_ALIAS_PORT_6379_TCP_PROTO="tcp"
|
||||
declare -x SHLVL="1"
|
||||
declare -x container="lxc"
|
||||
|
||||
And we can use that information to connect from another container as a
|
||||
client:
|
||||
|
||||
$ docker run -i -t --rm --link redis-name:redis_alias --entrypoint /bin/bash dockerfiles/redis -c '/redis-stable/src/redis-cli -h $REDIS_ALIAS_PORT_6379_TCP_ADDR -p $REDIS_ALIAS_PORT_6379_TCP_PORT'
|
||||
172.17.0.32:6379>
|
||||
|
||||
### [VOLUME (Shared Filesystems)](#id16)
|
||||
|
||||
-v=[]: Create a bind mount with: [host-dir]:[container-dir]:[rw|ro].
|
||||
If "container-dir" is missing, then docker creates a new volume.
|
||||
--volumes-from="": Mount all volumes from the given container(s)
|
||||
|
||||
The volumes commands are complex enough to have their own documentation
|
||||
in section [*Share Directories via
|
||||
Volumes*](../../use/working_with_volumes/#volume-def). A developer can
|
||||
define one or more `VOLUME`s associated with an
|
||||
image, but only the operator can give access from one container to
|
||||
another (or from a container to a volume mounted on the host).
|
||||
|
||||
### [USER](#id17)
|
||||
|
||||
The default user within a container is `root` (id =
|
||||
0), but if the developer created additional users, those are accessible
|
||||
too. The developer can set a default user to run the first process with
|
||||
the `Dockerfile USER` command, but the operator can
|
||||
override it
|
||||
|
||||
-u="": Username or UID
|
||||
|
||||
### [WORKDIR](#id18)
|
||||
|
||||
The default working directory for running binaries within a container is
|
||||
the root directory (`/`), but the developer can set
|
||||
a different default with the `Dockerfile WORKDIR`
|
||||
command. The operator can override this with:
|
||||
|
||||
-w="": Working directory inside the container
|
Loading…
Reference in New Issue