summaryrefslogtreecommitdiff
path: root/runtime/container.go
diff options
context:
space:
mode:
Diffstat (limited to 'runtime/container.go')
-rw-r--r--runtime/container.go1229
1 files changed, 1229 insertions, 0 deletions
diff --git a/runtime/container.go b/runtime/container.go
new file mode 100644
index 0000000000..c8053b146c
--- /dev/null
+++ b/runtime/container.go
@@ -0,0 +1,1229 @@
+package runtime
+
+import (
+ "encoding/json"
+ "errors"
+ "fmt"
+ "github.com/dotcloud/docker/archive"
+ "github.com/dotcloud/docker/engine"
+ "github.com/dotcloud/docker/image"
+ "github.com/dotcloud/docker/links"
+ "github.com/dotcloud/docker/nat"
+ "github.com/dotcloud/docker/runconfig"
+ "github.com/dotcloud/docker/runtime/execdriver"
+ "github.com/dotcloud/docker/runtime/graphdriver"
+ "github.com/dotcloud/docker/utils"
+ "io"
+ "io/ioutil"
+ "log"
+ "os"
+ "path"
+ "strings"
+ "sync"
+ "syscall"
+ "time"
+)
+
+const DefaultPathEnv = "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
+
+var (
+ ErrNotATTY = errors.New("The PTY is not a file")
+ ErrNoTTY = errors.New("No PTY found")
+ ErrContainerStart = errors.New("The container failed to start. Unknown error")
+ ErrContainerStartTimeout = errors.New("The container failed to start due to timed out.")
+)
+
+type Container struct {
+ sync.Mutex
+ root string // Path to the "home" of the container, including metadata.
+ basefs string // Path to the graphdriver mountpoint
+
+ ID string
+
+ Created time.Time
+
+ Path string
+ Args []string
+
+ Config *runconfig.Config
+ State State
+ Image string
+
+ NetworkSettings *NetworkSettings
+
+ ResolvConfPath string
+ HostnamePath string
+ HostsPath string
+ Name string
+ Driver string
+ ExecDriver string
+
+ command *execdriver.Command
+ stdout *utils.WriteBroadcaster
+ stderr *utils.WriteBroadcaster
+ stdin io.ReadCloser
+ stdinPipe io.WriteCloser
+
+ runtime *Runtime
+
+ waitLock chan struct{}
+ Volumes map[string]string
+ // Store rw/ro in a separate structure to preserve reverse-compatibility on-disk.
+ // Easier than migrating older container configs :)
+ VolumesRW map[string]bool
+ hostConfig *runconfig.HostConfig
+
+ activeLinks map[string]*links.Link
+}
+
+// FIXME: move deprecated port stuff to nat to clean up the core.
+type PortMapping map[string]string // Deprecated
+
+type NetworkSettings struct {
+ IPAddress string
+ IPPrefixLen int
+ Gateway string
+ Bridge string
+ PortMapping map[string]PortMapping // Deprecated
+ Ports nat.PortMap
+}
+
+func (settings *NetworkSettings) PortMappingAPI() *engine.Table {
+ var outs = engine.NewTable("", 0)
+ for port, bindings := range settings.Ports {
+ p, _ := nat.ParsePort(port.Port())
+ if len(bindings) == 0 {
+ out := &engine.Env{}
+ out.SetInt("PublicPort", p)
+ out.Set("Type", port.Proto())
+ outs.Add(out)
+ continue
+ }
+ for _, binding := range bindings {
+ out := &engine.Env{}
+ h, _ := nat.ParsePort(binding.HostPort)
+ out.SetInt("PrivatePort", p)
+ out.SetInt("PublicPort", h)
+ out.Set("Type", port.Proto())
+ out.Set("IP", binding.HostIp)
+ outs.Add(out)
+ }
+ }
+ return outs
+}
+
+// Inject the io.Reader at the given path. Note: do not close the reader
+func (container *Container) Inject(file io.Reader, pth string) error {
+ if err := container.Mount(); err != nil {
+ return fmt.Errorf("inject: error mounting container %s: %s", container.ID, err)
+ }
+ defer container.Unmount()
+
+ // Return error if path exists
+ destPath := path.Join(container.basefs, pth)
+ if _, err := os.Stat(destPath); err == nil {
+ // Since err is nil, the path could be stat'd and it exists
+ return fmt.Errorf("%s exists", pth)
+ } else if !os.IsNotExist(err) {
+ // Expect err might be that the file doesn't exist, so
+ // if it's some other error, return that.
+
+ return err
+ }
+
+ // Make sure the directory exists
+ if err := os.MkdirAll(path.Join(container.basefs, path.Dir(pth)), 0755); err != nil {
+ return err
+ }
+
+ dest, err := os.Create(destPath)
+ if err != nil {
+ return err
+ }
+ defer dest.Close()
+
+ if _, err := io.Copy(dest, file); err != nil {
+ return err
+ }
+ return nil
+}
+
+func (container *Container) When() time.Time {
+ return container.Created
+}
+
+func (container *Container) FromDisk() error {
+ data, err := ioutil.ReadFile(container.jsonPath())
+ if err != nil {
+ return err
+ }
+ // Load container settings
+ // udp broke compat of docker.PortMapping, but it's not used when loading a container, we can skip it
+ if err := json.Unmarshal(data, container); err != nil && !strings.Contains(err.Error(), "docker.PortMapping") {
+ return err
+ }
+ return container.readHostConfig()
+}
+
+func (container *Container) ToDisk() (err error) {
+ data, err := json.Marshal(container)
+ if err != nil {
+ return
+ }
+ err = ioutil.WriteFile(container.jsonPath(), data, 0666)
+ if err != nil {
+ return
+ }
+ return container.WriteHostConfig()
+}
+
+func (container *Container) readHostConfig() error {
+ container.hostConfig = &runconfig.HostConfig{}
+ // If the hostconfig file does not exist, do not read it.
+ // (We still have to initialize container.hostConfig,
+ // but that's OK, since we just did that above.)
+ _, err := os.Stat(container.hostConfigPath())
+ if os.IsNotExist(err) {
+ return nil
+ }
+ data, err := ioutil.ReadFile(container.hostConfigPath())
+ if err != nil {
+ return err
+ }
+ return json.Unmarshal(data, container.hostConfig)
+}
+
+func (container *Container) WriteHostConfig() (err error) {
+ data, err := json.Marshal(container.hostConfig)
+ if err != nil {
+ return
+ }
+ return ioutil.WriteFile(container.hostConfigPath(), data, 0666)
+}
+
+func (container *Container) generateEnvConfig(env []string) error {
+ data, err := json.Marshal(env)
+ if err != nil {
+ return err
+ }
+ p, err := container.EnvConfigPath()
+ if err != nil {
+ return err
+ }
+ ioutil.WriteFile(p, data, 0600)
+ return nil
+}
+
+func (container *Container) Attach(stdin io.ReadCloser, stdinCloser io.Closer, stdout io.Writer, stderr io.Writer) chan error {
+ var cStdout, cStderr io.ReadCloser
+
+ var nJobs int
+ errors := make(chan error, 3)
+ if stdin != nil && container.Config.OpenStdin {
+ nJobs += 1
+ if cStdin, err := container.StdinPipe(); err != nil {
+ errors <- err
+ } else {
+ go func() {
+ utils.Debugf("attach: stdin: begin")
+ defer utils.Debugf("attach: stdin: end")
+ // No matter what, when stdin is closed (io.Copy unblock), close stdout and stderr
+ if container.Config.StdinOnce && !container.Config.Tty {
+ defer cStdin.Close()
+ } else {
+ defer func() {
+ if cStdout != nil {
+ cStdout.Close()
+ }
+ if cStderr != nil {
+ cStderr.Close()
+ }
+ }()
+ }
+ if container.Config.Tty {
+ _, err = utils.CopyEscapable(cStdin, stdin)
+ } else {
+ _, err = io.Copy(cStdin, stdin)
+ }
+ if err == io.ErrClosedPipe {
+ err = nil
+ }
+ if err != nil {
+ utils.Errorf("attach: stdin: %s", err)
+ }
+ errors <- err
+ }()
+ }
+ }
+ if stdout != nil {
+ nJobs += 1
+ if p, err := container.StdoutPipe(); err != nil {
+ errors <- err
+ } else {
+ cStdout = p
+ go func() {
+ utils.Debugf("attach: stdout: begin")
+ defer utils.Debugf("attach: stdout: end")
+ // If we are in StdinOnce mode, then close stdin
+ if container.Config.StdinOnce && stdin != nil {
+ defer stdin.Close()
+ }
+ if stdinCloser != nil {
+ defer stdinCloser.Close()
+ }
+ _, err := io.Copy(stdout, cStdout)
+ if err == io.ErrClosedPipe {
+ err = nil
+ }
+ if err != nil {
+ utils.Errorf("attach: stdout: %s", err)
+ }
+ errors <- err
+ }()
+ }
+ } else {
+ go func() {
+ if stdinCloser != nil {
+ defer stdinCloser.Close()
+ }
+ if cStdout, err := container.StdoutPipe(); err != nil {
+ utils.Errorf("attach: stdout pipe: %s", err)
+ } else {
+ io.Copy(&utils.NopWriter{}, cStdout)
+ }
+ }()
+ }
+ if stderr != nil {
+ nJobs += 1
+ if p, err := container.StderrPipe(); err != nil {
+ errors <- err
+ } else {
+ cStderr = p
+ go func() {
+ utils.Debugf("attach: stderr: begin")
+ defer utils.Debugf("attach: stderr: end")
+ // If we are in StdinOnce mode, then close stdin
+ if container.Config.StdinOnce && stdin != nil {
+ defer stdin.Close()
+ }
+ if stdinCloser != nil {
+ defer stdinCloser.Close()
+ }
+ _, err := io.Copy(stderr, cStderr)
+ if err == io.ErrClosedPipe {
+ err = nil
+ }
+ if err != nil {
+ utils.Errorf("attach: stderr: %s", err)
+ }
+ errors <- err
+ }()
+ }
+ } else {
+ go func() {
+ if stdinCloser != nil {
+ defer stdinCloser.Close()
+ }
+
+ if cStderr, err := container.StderrPipe(); err != nil {
+ utils.Errorf("attach: stdout pipe: %s", err)
+ } else {
+ io.Copy(&utils.NopWriter{}, cStderr)
+ }
+ }()
+ }
+
+ return utils.Go(func() error {
+ defer func() {
+ if cStdout != nil {
+ cStdout.Close()
+ }
+ if cStderr != nil {
+ cStderr.Close()
+ }
+ }()
+
+ // FIXME: how to clean up the stdin goroutine without the unwanted side effect
+ // of closing the passed stdin? Add an intermediary io.Pipe?
+ for i := 0; i < nJobs; i += 1 {
+ utils.Debugf("attach: waiting for job %d/%d", i+1, nJobs)
+ if err := <-errors; err != nil {
+ utils.Errorf("attach: job %d returned error %s, aborting all jobs", i+1, err)
+ return err
+ }
+ utils.Debugf("attach: job %d completed successfully", i+1)
+ }
+ utils.Debugf("attach: all jobs completed successfully")
+ return nil
+ })
+}
+
+func populateCommand(c *Container) {
+ var (
+ en *execdriver.Network
+ driverConfig = make(map[string][]string)
+ )
+
+ en = &execdriver.Network{
+ Mtu: c.runtime.config.Mtu,
+ Interface: nil,
+ }
+
+ if !c.Config.NetworkDisabled {
+ network := c.NetworkSettings
+ en.Interface = &execdriver.NetworkInterface{
+ Gateway: network.Gateway,
+ Bridge: network.Bridge,
+ IPAddress: network.IPAddress,
+ IPPrefixLen: network.IPPrefixLen,
+ }
+ }
+
+ // TODO: this can be removed after lxc-conf is fully deprecated
+ mergeLxcConfIntoOptions(c.hostConfig, driverConfig)
+
+ resources := &execdriver.Resources{
+ Memory: c.Config.Memory,
+ MemorySwap: c.Config.MemorySwap,
+ CpuShares: c.Config.CpuShares,
+ }
+ c.command = &execdriver.Command{
+ ID: c.ID,
+ Privileged: c.hostConfig.Privileged,
+ Rootfs: c.RootfsPath(),
+ InitPath: "/.dockerinit",
+ Entrypoint: c.Path,
+ Arguments: c.Args,
+ WorkingDir: c.Config.WorkingDir,
+ Network: en,
+ Tty: c.Config.Tty,
+ User: c.Config.User,
+ Config: driverConfig,
+ Resources: resources,
+ }
+ c.command.SysProcAttr = &syscall.SysProcAttr{Setsid: true}
+}
+
+func (container *Container) ArgsAsString() string {
+ var args []string
+ for _, arg := range container.Args {
+ if strings.Contains(arg, " ") {
+ args = append(args, fmt.Sprintf("'%s'", arg))
+ } else {
+ args = append(args, arg)
+ }
+ }
+ return strings.Join(args, " ")
+}
+
+func (container *Container) Start() (err error) {
+ container.Lock()
+ defer container.Unlock()
+
+ if container.State.IsRunning() {
+ return nil
+ }
+
+ defer func() {
+ if err != nil {
+ container.cleanup()
+ }
+ }()
+
+ if container.ResolvConfPath == "" {
+ if err := container.setupContainerDns(); err != nil {
+ return err
+ }
+ }
+
+ if err := container.Mount(); err != nil {
+ return err
+ }
+
+ if container.runtime.config.DisableNetwork {
+ container.Config.NetworkDisabled = true
+ container.buildHostnameAndHostsFiles("127.0.1.1")
+ } else {
+ if err := container.allocateNetwork(); err != nil {
+ return err
+ }
+ container.buildHostnameAndHostsFiles(container.NetworkSettings.IPAddress)
+ }
+
+ // Make sure the config is compatible with the current kernel
+ if container.Config.Memory > 0 && !container.runtime.sysInfo.MemoryLimit {
+ log.Printf("WARNING: Your kernel does not support memory limit capabilities. Limitation discarded.\n")
+ container.Config.Memory = 0
+ }
+ if container.Config.Memory > 0 && !container.runtime.sysInfo.SwapLimit {
+ log.Printf("WARNING: Your kernel does not support swap limit capabilities. Limitation discarded.\n")
+ container.Config.MemorySwap = -1
+ }
+
+ if container.runtime.sysInfo.IPv4ForwardingDisabled {
+ log.Printf("WARNING: IPv4 forwarding is disabled. Networking will not work")
+ }
+
+ if err := prepareVolumesForContainer(container); err != nil {
+ return err
+ }
+
+ // Setup environment
+ env := []string{
+ "HOME=/",
+ "PATH=" + DefaultPathEnv,
+ "HOSTNAME=" + container.Config.Hostname,
+ }
+
+ if container.Config.Tty {
+ env = append(env, "TERM=xterm")
+ }
+
+ // Init any links between the parent and children
+ runtime := container.runtime
+
+ children, err := runtime.Children(container.Name)
+ if err != nil {
+ return err
+ }
+
+ if len(children) > 0 {
+ container.activeLinks = make(map[string]*links.Link, len(children))
+
+ // If we encounter an error make sure that we rollback any network
+ // config and ip table changes
+ rollback := func() {
+ for _, link := range container.activeLinks {
+ link.Disable()
+ }
+ container.activeLinks = nil
+ }
+
+ for linkAlias, child := range children {
+ if !child.State.IsRunning() {
+ return fmt.Errorf("Cannot link to a non running container: %s AS %s", child.Name, linkAlias)
+ }
+
+ link, err := links.NewLink(
+ container.NetworkSettings.IPAddress,
+ child.NetworkSettings.IPAddress,
+ linkAlias,
+ child.Config.Env,
+ child.Config.ExposedPorts,
+ runtime.eng)
+
+ if err != nil {
+ rollback()
+ return err
+ }
+
+ container.activeLinks[link.Alias()] = link
+ if err := link.Enable(); err != nil {
+ rollback()
+ return err
+ }
+
+ for _, envVar := range link.ToEnv() {
+ env = append(env, envVar)
+ }
+ }
+ }
+
+ // because the env on the container can override certain default values
+ // we need to replace the 'env' keys where they match and append anything
+ // else.
+ env = utils.ReplaceOrAppendEnvValues(env, container.Config.Env)
+ if err := container.generateEnvConfig(env); err != nil {
+ return err
+ }
+
+ if container.Config.WorkingDir != "" {
+ container.Config.WorkingDir = path.Clean(container.Config.WorkingDir)
+
+ pthInfo, err := os.Stat(path.Join(container.basefs, container.Config.WorkingDir))
+ if err != nil {
+ if !os.IsNotExist(err) {
+ return err
+ }
+ if err := os.MkdirAll(path.Join(container.basefs, container.Config.WorkingDir), 0755); err != nil {
+ return err
+ }
+ }
+ if pthInfo != nil && !pthInfo.IsDir() {
+ return fmt.Errorf("Cannot mkdir: %s is not a directory", container.Config.WorkingDir)
+ }
+ }
+
+ envPath, err := container.EnvConfigPath()
+ if err != nil {
+ return err
+ }
+
+ populateCommand(container)
+ container.command.Env = env
+
+ if err := setupMountsForContainer(container, envPath); err != nil {
+ return err
+ }
+
+ // Setup logging of stdout and stderr to disk
+ if err := container.runtime.LogToDisk(container.stdout, container.logPath("json"), "stdout"); err != nil {
+ return err
+ }
+ if err := container.runtime.LogToDisk(container.stderr, container.logPath("json"), "stderr"); err != nil {
+ return err
+ }
+ container.waitLock = make(chan struct{})
+
+ callbackLock := make(chan struct{})
+ callback := func(command *execdriver.Command) {
+ container.State.SetRunning(command.Pid())
+ if command.Tty {
+ // The callback is called after the process Start()
+ // so we are in the parent process. In TTY mode, stdin/out/err is the PtySlace
+ // which we close here.
+ if c, ok := command.Stdout.(io.Closer); ok {
+ c.Close()
+ }
+ }
+ if err := container.ToDisk(); err != nil {
+ utils.Debugf("%s", err)
+ }
+ close(callbackLock)
+ }
+
+ // We use a callback here instead of a goroutine and an chan for
+ // syncronization purposes
+ cErr := utils.Go(func() error { return container.monitor(callback) })
+
+ // Start should not return until the process is actually running
+ select {
+ case <-callbackLock:
+ case err := <-cErr:
+ return err
+ }
+ return nil
+}
+
+func (container *Container) Run() error {
+ if err := container.Start(); err != nil {
+ return err
+ }
+ container.Wait()
+ return nil
+}
+
+func (container *Container) Output() (output []byte, err error) {
+ pipe, err := container.StdoutPipe()
+ if err != nil {
+ return nil, err
+ }
+ defer pipe.Close()
+ if err := container.Start(); err != nil {
+ return nil, err
+ }
+ output, err = ioutil.ReadAll(pipe)
+ container.Wait()
+ return output, err
+}
+
+// Container.StdinPipe returns a WriteCloser which can be used to feed data
+// to the standard input of the container's active process.
+// Container.StdoutPipe and Container.StderrPipe each return a ReadCloser
+// which can be used to retrieve the standard output (and error) generated
+// by the container's active process. The output (and error) are actually
+// copied and delivered to all StdoutPipe and StderrPipe consumers, using
+// a kind of "broadcaster".
+
+func (container *Container) StdinPipe() (io.WriteCloser, error) {
+ return container.stdinPipe, nil
+}
+
+func (container *Container) StdoutPipe() (io.ReadCloser, error) {
+ reader, writer := io.Pipe()
+ container.stdout.AddWriter(writer, "")
+ return utils.NewBufReader(reader), nil
+}
+
+func (container *Container) StderrPipe() (io.ReadCloser, error) {
+ reader, writer := io.Pipe()
+ container.stderr.AddWriter(writer, "")
+ return utils.NewBufReader(reader), nil
+}
+
+func (container *Container) buildHostnameAndHostsFiles(IP string) {
+ container.HostnamePath = path.Join(container.root, "hostname")
+ ioutil.WriteFile(container.HostnamePath, []byte(container.Config.Hostname+"\n"), 0644)
+
+ hostsContent := []byte(`
+127.0.0.1 localhost
+::1 localhost ip6-localhost ip6-loopback
+fe00::0 ip6-localnet
+ff00::0 ip6-mcastprefix
+ff02::1 ip6-allnodes
+ff02::2 ip6-allrouters
+`)
+
+ container.HostsPath = path.Join(container.root, "hosts")
+
+ if container.Config.Domainname != "" {
+ hostsContent = append([]byte(fmt.Sprintf("%s\t%s.%s %s\n", IP, container.Config.Hostname, container.Config.Domainname, container.Config.Hostname)), hostsContent...)
+ } else if !container.Config.NetworkDisabled {
+ hostsContent = append([]byte(fmt.Sprintf("%s\t%s\n", IP, container.Config.Hostname)), hostsContent...)
+ }
+
+ ioutil.WriteFile(container.HostsPath, hostsContent, 0644)
+}
+
+func (container *Container) allocateNetwork() error {
+ if container.Config.NetworkDisabled {
+ return nil
+ }
+
+ var (
+ env *engine.Env
+ err error
+ eng = container.runtime.eng
+ )
+
+ if container.State.IsGhost() {
+ if container.runtime.config.DisableNetwork {
+ env = &engine.Env{}
+ } else {
+ currentIP := container.NetworkSettings.IPAddress
+
+ job := eng.Job("allocate_interface", container.ID)
+ if currentIP != "" {
+ job.Setenv("RequestIP", currentIP)
+ }
+
+ env, err = job.Stdout.AddEnv()
+ if err != nil {
+ return err
+ }
+
+ if err := job.Run(); err != nil {
+ return err
+ }
+ }
+ } else {
+ job := eng.Job("allocate_interface", container.ID)
+ env, err = job.Stdout.AddEnv()
+ if err != nil {
+ return err
+ }
+ if err := job.Run(); err != nil {
+ return err
+ }
+ }
+
+ if container.Config.PortSpecs != nil {
+ utils.Debugf("Migrating port mappings for container: %s", strings.Join(container.Config.PortSpecs, ", "))
+ if err := migratePortMappings(container.Config, container.hostConfig); err != nil {
+ return err
+ }
+ container.Config.PortSpecs = nil
+ if err := container.WriteHostConfig(); err != nil {
+ return err
+ }
+ }
+
+ var (
+ portSpecs = make(nat.PortSet)
+ bindings = make(nat.PortMap)
+ )
+
+ if !container.State.IsGhost() {
+ if container.Config.ExposedPorts != nil {
+ portSpecs = container.Config.ExposedPorts
+ }
+ if container.hostConfig.PortBindings != nil {
+ bindings = container.hostConfig.PortBindings
+ }
+ } else {
+ if container.NetworkSettings.Ports != nil {
+ for port, binding := range container.NetworkSettings.Ports {
+ portSpecs[port] = struct{}{}
+ bindings[port] = binding
+ }
+ }
+ }
+
+ container.NetworkSettings.PortMapping = nil
+
+ for port := range portSpecs {
+ binding := bindings[port]
+ if container.hostConfig.PublishAllPorts && len(binding) == 0 {
+ binding = append(binding, nat.PortBinding{})
+ }
+
+ for i := 0; i < len(binding); i++ {
+ b := binding[i]
+
+ portJob := eng.Job("allocate_port", container.ID)
+ portJob.Setenv("HostIP", b.HostIp)
+ portJob.Setenv("HostPort", b.HostPort)
+ portJob.Setenv("Proto", port.Proto())
+ portJob.Setenv("ContainerPort", port.Port())
+
+ portEnv, err := portJob.Stdout.AddEnv()
+ if err != nil {
+ return err
+ }
+ if err := portJob.Run(); err != nil {
+ eng.Job("release_interface", container.ID).Run()
+ return err
+ }
+ b.HostIp = portEnv.Get("HostIP")
+ b.HostPort = portEnv.Get("HostPort")
+
+ binding[i] = b
+ }
+ bindings[port] = binding
+ }
+ container.WriteHostConfig()
+
+ container.NetworkSettings.Ports = bindings
+
+ container.NetworkSettings.Bridge = env.Get("Bridge")
+ container.NetworkSettings.IPAddress = env.Get("IP")
+ container.NetworkSettings.IPPrefixLen = env.GetInt("IPPrefixLen")
+ container.NetworkSettings.Gateway = env.Get("Gateway")
+
+ return nil
+}
+
+func (container *Container) releaseNetwork() {
+ if container.Config.NetworkDisabled {
+ return
+ }
+ eng := container.runtime.eng
+
+ eng.Job("release_interface", container.ID).Run()
+ container.NetworkSettings = &NetworkSettings{}
+}
+
+func (container *Container) monitor(callback execdriver.StartCallback) error {
+ var (
+ err error
+ exitCode int
+ )
+
+ pipes := execdriver.NewPipes(container.stdin, container.stdout, container.stderr, container.Config.OpenStdin)
+ exitCode, err = container.runtime.Run(container, pipes, callback)
+ if err != nil {
+ utils.Errorf("Error running container: %s", err)
+ }
+
+ if container.runtime != nil && container.runtime.srv != nil && container.runtime.srv.IsRunning() {
+ container.State.SetStopped(exitCode)
+
+ // FIXME: there is a race condition here which causes this to fail during the unit tests.
+ // If another goroutine was waiting for Wait() to return before removing the container's root
+ // from the filesystem... At this point it may already have done so.
+ // This is because State.setStopped() has already been called, and has caused Wait()
+ // to return.
+ // FIXME: why are we serializing running state to disk in the first place?
+ //log.Printf("%s: Failed to dump configuration to the disk: %s", container.ID, err)
+ if err := container.ToDisk(); err != nil {
+ utils.Errorf("Error dumping container state to disk: %s\n", err)
+ }
+ }
+
+ // Cleanup
+ container.cleanup()
+
+ // Re-create a brand new stdin pipe once the container exited
+ if container.Config.OpenStdin {
+ container.stdin, container.stdinPipe = io.Pipe()
+ }
+
+ if container.runtime != nil && container.runtime.srv != nil {
+ container.runtime.srv.LogEvent("die", container.ID, container.runtime.repositories.ImageName(container.Image))
+ }
+
+ close(container.waitLock)
+
+ return err
+}
+
+func (container *Container) cleanup() {
+ container.releaseNetwork()
+
+ // Disable all active links
+ if container.activeLinks != nil {
+ for _, link := range container.activeLinks {
+ link.Disable()
+ }
+ }
+ if container.Config.OpenStdin {
+ if err := container.stdin.Close(); err != nil {
+ utils.Errorf("%s: Error close stdin: %s", container.ID, err)
+ }
+ }
+ if err := container.stdout.CloseWriters(); err != nil {
+ utils.Errorf("%s: Error close stdout: %s", container.ID, err)
+ }
+ if err := container.stderr.CloseWriters(); err != nil {
+ utils.Errorf("%s: Error close stderr: %s", container.ID, err)
+ }
+ if container.command != nil && container.command.Terminal != nil {
+ if err := container.command.Terminal.Close(); err != nil {
+ utils.Errorf("%s: Error closing terminal: %s", container.ID, err)
+ }
+ }
+
+ if err := container.Unmount(); err != nil {
+ log.Printf("%v: Failed to umount filesystem: %v", container.ID, err)
+ }
+}
+
+func (container *Container) KillSig(sig int) error {
+ container.Lock()
+ defer container.Unlock()
+
+ if !container.State.IsRunning() {
+ return nil
+ }
+ return container.runtime.Kill(container, sig)
+}
+
+func (container *Container) Kill() error {
+ if !container.State.IsRunning() {
+ return nil
+ }
+
+ // 1. Send SIGKILL
+ if err := container.KillSig(9); err != nil {
+ return err
+ }
+
+ // 2. Wait for the process to die, in last resort, try to kill the process directly
+ if err := container.WaitTimeout(10 * time.Second); err != nil {
+ log.Printf("Container %s failed to exit within 10 seconds of kill - trying direct SIGKILL", utils.TruncateID(container.ID))
+ if err := syscall.Kill(container.State.Pid, 9); err != nil {
+ return err
+ }
+ }
+
+ container.Wait()
+ return nil
+}
+
+func (container *Container) Stop(seconds int) error {
+ if !container.State.IsRunning() {
+ return nil
+ }
+
+ // 1. Send a SIGTERM
+ if err := container.KillSig(15); err != nil {
+ log.Print("Failed to send SIGTERM to the process, force killing")
+ if err := container.KillSig(9); err != nil {
+ return err
+ }
+ }
+
+ // 2. Wait for the process to exit on its own
+ if err := container.WaitTimeout(time.Duration(seconds) * time.Second); err != nil {
+ log.Printf("Container %v failed to exit within %d seconds of SIGTERM - using the force", container.ID, seconds)
+ // 3. If it doesn't, then send SIGKILL
+ if err := container.Kill(); err != nil {
+ return err
+ }
+ }
+ return nil
+}
+
+func (container *Container) Restart(seconds int) error {
+ // Avoid unnecessarily unmounting and then directly mounting
+ // the container when the container stops and then starts
+ // again
+ if err := container.Mount(); err == nil {
+ defer container.Unmount()
+ }
+
+ if err := container.Stop(seconds); err != nil {
+ return err
+ }
+ return container.Start()
+}
+
+// Wait blocks until the container stops running, then returns its exit code.
+func (container *Container) Wait() int {
+ <-container.waitLock
+ return container.State.GetExitCode()
+}
+
+func (container *Container) Resize(h, w int) error {
+ return container.command.Terminal.Resize(h, w)
+}
+
+func (container *Container) ExportRw() (archive.Archive, error) {
+ if err := container.Mount(); err != nil {
+ return nil, err
+ }
+ if container.runtime == nil {
+ return nil, fmt.Errorf("Can't load storage driver for unregistered container %s", container.ID)
+ }
+ archive, err := container.runtime.Diff(container)
+ if err != nil {
+ container.Unmount()
+ return nil, err
+ }
+ return utils.NewReadCloserWrapper(archive, func() error {
+ err := archive.Close()
+ container.Unmount()
+ return err
+ }),
+ nil
+}
+
+func (container *Container) Export() (archive.Archive, error) {
+ if err := container.Mount(); err != nil {
+ return nil, err
+ }
+
+ archive, err := archive.Tar(container.basefs, archive.Uncompressed)
+ if err != nil {
+ container.Unmount()
+ return nil, err
+ }
+ return utils.NewReadCloserWrapper(archive, func() error {
+ err := archive.Close()
+ container.Unmount()
+ return err
+ }),
+ nil
+}
+
+func (container *Container) WaitTimeout(timeout time.Duration) error {
+ done := make(chan bool)
+ go func() {
+ container.Wait()
+ done <- true
+ }()
+
+ select {
+ case <-time.After(timeout):
+ return fmt.Errorf("Timed Out")
+ case <-done:
+ return nil
+ }
+}
+
+func (container *Container) Mount() error {
+ return container.runtime.Mount(container)
+}
+
+func (container *Container) Changes() ([]archive.Change, error) {
+ return container.runtime.Changes(container)
+}
+
+func (container *Container) GetImage() (*image.Image, error) {
+ if container.runtime == nil {
+ return nil, fmt.Errorf("Can't get image of unregistered container")
+ }
+ return container.runtime.graph.Get(container.Image)
+}
+
+func (container *Container) Unmount() error {
+ return container.runtime.Unmount(container)
+}
+
+func (container *Container) logPath(name string) string {
+ return path.Join(container.root, fmt.Sprintf("%s-%s.log", container.ID, name))
+}
+
+func (container *Container) ReadLog(name string) (io.Reader, error) {
+ return os.Open(container.logPath(name))
+}
+
+func (container *Container) hostConfigPath() string {
+ return path.Join(container.root, "hostconfig.json")
+}
+
+func (container *Container) jsonPath() string {
+ return path.Join(container.root, "config.json")
+}
+
+func (container *Container) EnvConfigPath() (string, error) {
+ p := path.Join(container.root, "config.env")
+ if _, err := os.Stat(p); err != nil {
+ if os.IsNotExist(err) {
+ f, err := os.Create(p)
+ if err != nil {
+ return "", err
+ }
+ f.Close()
+ } else {
+ return "", err
+ }
+ }
+ return p, nil
+}
+
+// This method must be exported to be used from the lxc template
+// This directory is only usable when the container is running
+func (container *Container) RootfsPath() string {
+ return container.basefs
+}
+
+func validateID(id string) error {
+ if id == "" {
+ return fmt.Errorf("Invalid empty id")
+ }
+ return nil
+}
+
+// GetSize, return real size, virtual size
+func (container *Container) GetSize() (int64, int64) {
+ var (
+ sizeRw, sizeRootfs int64
+ err error
+ driver = container.runtime.driver
+ )
+
+ if err := container.Mount(); err != nil {
+ utils.Errorf("Warning: failed to compute size of container rootfs %s: %s", container.ID, err)
+ return sizeRw, sizeRootfs
+ }
+ defer container.Unmount()
+
+ if differ, ok := container.runtime.driver.(graphdriver.Differ); ok {
+ sizeRw, err = differ.DiffSize(container.ID)
+ if err != nil {
+ utils.Errorf("Warning: driver %s couldn't return diff size of container %s: %s", driver, container.ID, err)
+ // FIXME: GetSize should return an error. Not changing it now in case
+ // there is a side-effect.
+ sizeRw = -1
+ }
+ } else {
+ changes, _ := container.Changes()
+ if changes != nil {
+ sizeRw = archive.ChangesSize(container.basefs, changes)
+ } else {
+ sizeRw = -1
+ }
+ }
+
+ if _, err = os.Stat(container.basefs); err != nil {
+ if sizeRootfs, err = utils.TreeSize(container.basefs); err != nil {
+ sizeRootfs = -1
+ }
+ }
+ return sizeRw, sizeRootfs
+}
+
+func (container *Container) Copy(resource string) (io.ReadCloser, error) {
+ if err := container.Mount(); err != nil {
+ return nil, err
+ }
+ var filter []string
+ basePath := path.Join(container.basefs, resource)
+ stat, err := os.Stat(basePath)
+ if err != nil {
+ container.Unmount()
+ return nil, err
+ }
+ if !stat.IsDir() {
+ d, f := path.Split(basePath)
+ basePath = d
+ filter = []string{f}
+ } else {
+ filter = []string{path.Base(basePath)}
+ basePath = path.Dir(basePath)
+ }
+
+ archive, err := archive.TarFilter(basePath, &archive.TarOptions{
+ Compression: archive.Uncompressed,
+ Includes: filter,
+ })
+ if err != nil {
+ return nil, err
+ }
+ return utils.NewReadCloserWrapper(archive, func() error {
+ err := archive.Close()
+ container.Unmount()
+ return err
+ }),
+ nil
+}
+
+// Returns true if the container exposes a certain port
+func (container *Container) Exposes(p nat.Port) bool {
+ _, exists := container.Config.ExposedPorts[p]
+ return exists
+}
+
+func (container *Container) GetPtyMaster() (*os.File, error) {
+ ttyConsole, ok := container.command.Terminal.(execdriver.TtyTerminal)
+ if !ok {
+ return nil, ErrNoTTY
+ }
+ return ttyConsole.Master(), nil
+}
+
+func (container *Container) HostConfig() *runconfig.HostConfig {
+ return container.hostConfig
+}
+
+func (container *Container) SetHostConfig(hostConfig *runconfig.HostConfig) {
+ container.hostConfig = hostConfig
+}
+
+func (container *Container) DisableLink(name string) {
+ if container.activeLinks != nil {
+ if link, exists := container.activeLinks[name]; exists {
+ link.Disable()
+ } else {
+ utils.Debugf("Could not find active link for %s", name)
+ }
+ }
+}
+
+func (container *Container) setupContainerDns() error {
+ var (
+ config = container.hostConfig
+ runtime = container.runtime
+ )
+ resolvConf, err := utils.GetResolvConf()
+ if err != nil {
+ return err
+ }
+ // If custom dns exists, then create a resolv.conf for the container
+ if len(config.Dns) > 0 || len(runtime.config.Dns) > 0 || len(config.DnsSearch) > 0 || len(runtime.config.DnsSearch) > 0 {
+ var (
+ dns = utils.GetNameservers(resolvConf)
+ dnsSearch = utils.GetSearchDomains(resolvConf)
+ )
+ if len(config.Dns) > 0 {
+ dns = config.Dns
+ } else if len(runtime.config.Dns) > 0 {
+ dns = runtime.config.Dns
+ }
+ if len(config.DnsSearch) > 0 {
+ dnsSearch = config.DnsSearch
+ } else if len(runtime.config.DnsSearch) > 0 {
+ dnsSearch = runtime.config.DnsSearch
+ }
+ container.ResolvConfPath = path.Join(container.root, "resolv.conf")
+ f, err := os.Create(container.ResolvConfPath)
+ if err != nil {
+ return err
+ }
+ defer f.Close()
+ for _, dns := range dns {
+ if _, err := f.Write([]byte("nameserver " + dns + "\n")); err != nil {
+ return err
+ }
+ }
+ if len(dnsSearch) > 0 {
+ if _, err := f.Write([]byte("search " + strings.Join(dnsSearch, " ") + "\n")); err != nil {
+ return err
+ }
+ }
+ } else {
+ container.ResolvConfPath = "/etc/resolv.conf"
+ }
+ return nil
+}