mirror of https://github.com/containrrr/watchtower
				
				
				
			
			You cannot select more than 25 topics
			Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
		
		
		
		
		
			
		
			
				
	
	
		
			560 lines
		
	
	
		
			16 KiB
		
	
	
	
		
			Go
		
	
			
		
		
	
	
			560 lines
		
	
	
		
			16 KiB
		
	
	
	
		
			Go
		
	
package container
 | 
						|
 | 
						|
import (
 | 
						|
	"bytes"
 | 
						|
	"fmt"
 | 
						|
	"io/ioutil"
 | 
						|
	"strings"
 | 
						|
	"time"
 | 
						|
 | 
						|
	"github.com/containrrr/watchtower/pkg/registry"
 | 
						|
	"github.com/containrrr/watchtower/pkg/registry/digest"
 | 
						|
 | 
						|
	t "github.com/containrrr/watchtower/pkg/types"
 | 
						|
	"github.com/docker/docker/api/types"
 | 
						|
	"github.com/docker/docker/api/types/container"
 | 
						|
	"github.com/docker/docker/api/types/filters"
 | 
						|
	"github.com/docker/docker/api/types/network"
 | 
						|
	sdkClient "github.com/docker/docker/client"
 | 
						|
	log "github.com/sirupsen/logrus"
 | 
						|
	"golang.org/x/net/context"
 | 
						|
)
 | 
						|
 | 
						|
const defaultStopSignal = "SIGTERM"
 | 
						|
 | 
						|
// A Client is the interface through which watchtower interacts with the
 | 
						|
// Docker API.
 | 
						|
type Client interface {
 | 
						|
	ListContainers(t.Filter) ([]t.Container, error)
 | 
						|
	GetContainer(containerID t.ContainerID) (t.Container, error)
 | 
						|
	StopContainer(t.Container, time.Duration) error
 | 
						|
	StartContainer(t.Container) (t.ContainerID, error)
 | 
						|
	RenameContainer(t.Container, string) error
 | 
						|
	IsContainerStale(t.Container) (stale bool, latestImage t.ImageID, err error)
 | 
						|
	ExecuteCommand(containerID t.ContainerID, command string, timeout int) (SkipUpdate bool, err error)
 | 
						|
	RemoveImageByID(t.ImageID) error
 | 
						|
	WarnOnHeadPullFailed(container t.Container) bool
 | 
						|
	GetImage(imageID t.ImageID) (types.ImageInspect, error)
 | 
						|
}
 | 
						|
 | 
						|
// NewClient returns a new Client instance which can be used to interact with
 | 
						|
// the Docker API.
 | 
						|
// The client reads its configuration from the following environment variables:
 | 
						|
//   - DOCKER_HOST			the docker-engine host to send api requests to
 | 
						|
//   - DOCKER_TLS_VERIFY		whether to verify tls certificates
 | 
						|
//   - DOCKER_API_VERSION	the minimum docker api version to work with
 | 
						|
func NewClient(opts ClientOptions) Client {
 | 
						|
	cli, err := sdkClient.NewClientWithOpts(sdkClient.FromEnv)
 | 
						|
 | 
						|
	if err != nil {
 | 
						|
		log.Fatalf("Error instantiating Docker client: %s", err)
 | 
						|
	}
 | 
						|
 | 
						|
	return dockerClient{
 | 
						|
		api:           cli,
 | 
						|
		ClientOptions: opts,
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
// ClientOptions contains the options for how the docker client wrapper should behave
 | 
						|
type ClientOptions struct {
 | 
						|
	PullImages        bool
 | 
						|
	RemoveVolumes     bool
 | 
						|
	IncludeStopped    bool
 | 
						|
	ReviveStopped     bool
 | 
						|
	IncludeRestarting bool
 | 
						|
	WarnOnHeadFailed  WarningStrategy
 | 
						|
}
 | 
						|
 | 
						|
// WarningStrategy is a value determining when to show warnings
 | 
						|
type WarningStrategy string
 | 
						|
 | 
						|
const (
 | 
						|
	// WarnAlways warns whenever the problem occurs
 | 
						|
	WarnAlways WarningStrategy = "always"
 | 
						|
	// WarnNever never warns when the problem occurs
 | 
						|
	WarnNever WarningStrategy = "never"
 | 
						|
	// WarnAuto skips warning when the problem was expected
 | 
						|
	WarnAuto WarningStrategy = "auto"
 | 
						|
)
 | 
						|
 | 
						|
type dockerClient struct {
 | 
						|
	api sdkClient.CommonAPIClient
 | 
						|
	ClientOptions
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) WarnOnHeadPullFailed(container t.Container) bool {
 | 
						|
	if client.WarnOnHeadFailed == WarnAlways {
 | 
						|
		return true
 | 
						|
	}
 | 
						|
	if client.WarnOnHeadFailed == WarnNever {
 | 
						|
		return false
 | 
						|
	}
 | 
						|
 | 
						|
	return registry.WarnOnAPIConsumption(container)
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) ListContainers(fn t.Filter) ([]t.Container, error) {
 | 
						|
	cs := []t.Container{}
 | 
						|
	bg := context.Background()
 | 
						|
 | 
						|
	if client.IncludeStopped && client.IncludeRestarting {
 | 
						|
		log.Debug("Retrieving running, stopped, restarting and exited containers")
 | 
						|
	} else if client.IncludeStopped {
 | 
						|
		log.Debug("Retrieving running, stopped and exited containers")
 | 
						|
	} else if client.IncludeRestarting {
 | 
						|
		log.Debug("Retrieving running and restarting containers")
 | 
						|
	} else {
 | 
						|
		log.Debug("Retrieving running containers")
 | 
						|
	}
 | 
						|
 | 
						|
	filter := client.createListFilter()
 | 
						|
	containers, err := client.api.ContainerList(
 | 
						|
		bg,
 | 
						|
		types.ContainerListOptions{
 | 
						|
			Filters: filter,
 | 
						|
		})
 | 
						|
 | 
						|
	if err != nil {
 | 
						|
		return nil, err
 | 
						|
	}
 | 
						|
 | 
						|
	for _, runningContainer := range containers {
 | 
						|
 | 
						|
		c, err := client.GetContainer(t.ContainerID(runningContainer.ID))
 | 
						|
		if err != nil {
 | 
						|
			return nil, err
 | 
						|
		}
 | 
						|
 | 
						|
		if fn(c) {
 | 
						|
			cs = append(cs, c)
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	return cs, nil
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) createListFilter() filters.Args {
 | 
						|
	filterArgs := filters.NewArgs()
 | 
						|
	filterArgs.Add("status", "running")
 | 
						|
 | 
						|
	if client.IncludeStopped {
 | 
						|
		filterArgs.Add("status", "created")
 | 
						|
		filterArgs.Add("status", "exited")
 | 
						|
	}
 | 
						|
 | 
						|
	if client.IncludeRestarting {
 | 
						|
		filterArgs.Add("status", "restarting")
 | 
						|
	}
 | 
						|
 | 
						|
	return filterArgs
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) GetContainer(containerID t.ContainerID) (t.Container, error) {
 | 
						|
	bg := context.Background()
 | 
						|
 | 
						|
	containerInfo, err := client.api.ContainerInspect(bg, string(containerID))
 | 
						|
	if err != nil {
 | 
						|
		return &Container{}, err
 | 
						|
	}
 | 
						|
 | 
						|
	netType, netContainerId, found := strings.Cut(string(containerInfo.HostConfig.NetworkMode), ":")
 | 
						|
	if found && netType == "container" {
 | 
						|
		parentContainer, err := client.api.ContainerInspect(bg, netContainerId)
 | 
						|
		if err != nil {
 | 
						|
			log.WithFields(map[string]interface{}{
 | 
						|
				"container":         containerInfo.Name,
 | 
						|
				"error":             err,
 | 
						|
				"network-container": netContainerId,
 | 
						|
			}).Warnf("Unable to resolve network container: %v", err)
 | 
						|
 | 
						|
		} else {
 | 
						|
			// Replace the container ID with a container name to allow it to reference the re-created network container
 | 
						|
			containerInfo.HostConfig.NetworkMode = container.NetworkMode(fmt.Sprintf("container:%s", parentContainer.Name))
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	imageInfo, _, err := client.api.ImageInspectWithRaw(bg, containerInfo.Image)
 | 
						|
	if err != nil {
 | 
						|
		log.Warnf("Failed to retrieve container image info: %v", err)
 | 
						|
		return &Container{containerInfo: &containerInfo, imageInfo: nil}, nil
 | 
						|
	}
 | 
						|
 | 
						|
	return &Container{containerInfo: &containerInfo, imageInfo: &imageInfo}, nil
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) StopContainer(c t.Container, timeout time.Duration) error {
 | 
						|
	bg := context.Background()
 | 
						|
	signal := c.StopSignal()
 | 
						|
	if signal == "" {
 | 
						|
		signal = defaultStopSignal
 | 
						|
	}
 | 
						|
 | 
						|
	idStr := string(c.ID())
 | 
						|
	shortID := c.ID().ShortID()
 | 
						|
 | 
						|
	if c.IsRunning() {
 | 
						|
		log.Infof("Stopping %s (%s) with %s", c.Name(), shortID, signal)
 | 
						|
		if err := client.api.ContainerKill(bg, idStr, signal); err != nil {
 | 
						|
			return err
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	// TODO: This should probably be checked.
 | 
						|
	_ = client.waitForStopOrTimeout(c, timeout)
 | 
						|
 | 
						|
	if c.ContainerInfo().HostConfig.AutoRemove {
 | 
						|
		log.Debugf("AutoRemove container %s, skipping ContainerRemove call.", shortID)
 | 
						|
	} else {
 | 
						|
		log.Debugf("Removing container %s", shortID)
 | 
						|
 | 
						|
		if err := client.api.ContainerRemove(bg, idStr, types.ContainerRemoveOptions{Force: true, RemoveVolumes: client.RemoveVolumes}); err != nil {
 | 
						|
			if sdkClient.IsErrNotFound(err) {
 | 
						|
				log.Debugf("Container %s not found, skipping removal.", shortID)
 | 
						|
				return nil
 | 
						|
			}
 | 
						|
			return err
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	// Wait for container to be removed. In this case an error is a good thing
 | 
						|
	if err := client.waitForStopOrTimeout(c, timeout); err == nil {
 | 
						|
		return fmt.Errorf("container %s (%s) could not be removed", c.Name(), shortID)
 | 
						|
	}
 | 
						|
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) GetNetworkConfig(c t.Container) *network.NetworkingConfig {
 | 
						|
	config := &network.NetworkingConfig{
 | 
						|
		EndpointsConfig: c.ContainerInfo().NetworkSettings.Networks,
 | 
						|
	}
 | 
						|
 | 
						|
	for _, ep := range config.EndpointsConfig {
 | 
						|
		// This keeps accumulating across upgrades with no apparent added value
 | 
						|
		// so throwing the information away to prevent overflows.
 | 
						|
		ep.Aliases = nil
 | 
						|
	}
 | 
						|
	return config
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) StartContainer(c t.Container) (t.ContainerID, error) {
 | 
						|
	bg := context.Background()
 | 
						|
	config := c.GetCreateConfig()
 | 
						|
	hostConfig := c.GetCreateHostConfig()
 | 
						|
	networkConfig := client.GetNetworkConfig(c)
 | 
						|
 | 
						|
	// simpleNetworkConfig is a networkConfig with only 1 network.
 | 
						|
	// see: https://github.com/docker/docker/issues/29265
 | 
						|
	simpleNetworkConfig := func() *network.NetworkingConfig {
 | 
						|
		oneEndpoint := make(map[string]*network.EndpointSettings)
 | 
						|
		for k, v := range networkConfig.EndpointsConfig {
 | 
						|
			oneEndpoint[k] = v
 | 
						|
			// we only need 1
 | 
						|
			break
 | 
						|
		}
 | 
						|
		return &network.NetworkingConfig{EndpointsConfig: oneEndpoint}
 | 
						|
	}()
 | 
						|
 | 
						|
	name := c.Name()
 | 
						|
 | 
						|
	log.Infof("Creating %s", name)
 | 
						|
 | 
						|
	createdContainer, err := client.api.ContainerCreate(bg, config, hostConfig, simpleNetworkConfig, nil, name)
 | 
						|
	if err != nil {
 | 
						|
		return "", err
 | 
						|
	}
 | 
						|
 | 
						|
	if !(hostConfig.NetworkMode.IsHost()) {
 | 
						|
 | 
						|
		for k := range simpleNetworkConfig.EndpointsConfig {
 | 
						|
			err = client.api.NetworkDisconnect(bg, k, createdContainer.ID, true)
 | 
						|
			if err != nil {
 | 
						|
				return "", err
 | 
						|
			}
 | 
						|
		}
 | 
						|
 | 
						|
		for k, v := range networkConfig.EndpointsConfig {
 | 
						|
			err = client.api.NetworkConnect(bg, k, createdContainer.ID, v)
 | 
						|
			if err != nil {
 | 
						|
				return "", err
 | 
						|
			}
 | 
						|
		}
 | 
						|
 | 
						|
	}
 | 
						|
 | 
						|
	createdContainerID := t.ContainerID(createdContainer.ID)
 | 
						|
	if !c.IsRunning() && !client.ReviveStopped {
 | 
						|
		return createdContainerID, nil
 | 
						|
	}
 | 
						|
 | 
						|
	return createdContainerID, client.doStartContainer(bg, c, createdContainer)
 | 
						|
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) doStartContainer(bg context.Context, c t.Container, creation container.CreateResponse) error {
 | 
						|
	name := c.Name()
 | 
						|
 | 
						|
	log.Debugf("Starting container %s (%s)", name, t.ContainerID(creation.ID).ShortID())
 | 
						|
	err := client.api.ContainerStart(bg, creation.ID, types.ContainerStartOptions{})
 | 
						|
	if err != nil {
 | 
						|
		return err
 | 
						|
	}
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) RenameContainer(c t.Container, newName string) error {
 | 
						|
	bg := context.Background()
 | 
						|
	log.Debugf("Renaming container %s (%s) to %s", c.Name(), c.ID().ShortID(), newName)
 | 
						|
	return client.api.ContainerRename(bg, string(c.ID()), newName)
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) IsContainerStale(container t.Container) (stale bool, latestImage t.ImageID, err error) {
 | 
						|
	ctx := context.Background()
 | 
						|
 | 
						|
	if !client.PullImages || container.IsNoPull() {
 | 
						|
		log.Debugf("Skipping image pull.")
 | 
						|
	} else if err := client.PullImage(ctx, container); err != nil {
 | 
						|
		return false, container.SafeImageID(), err
 | 
						|
	}
 | 
						|
 | 
						|
	return client.HasNewImage(ctx, container)
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) HasNewImage(ctx context.Context, container t.Container) (hasNew bool, latestImage t.ImageID, err error) {
 | 
						|
	currentImageID := t.ImageID(container.ContainerInfo().ContainerJSONBase.Image)
 | 
						|
	imageName := container.ImageName()
 | 
						|
 | 
						|
	newImageInfo, _, err := client.api.ImageInspectWithRaw(ctx, imageName)
 | 
						|
	if err != nil {
 | 
						|
		return false, currentImageID, err
 | 
						|
	}
 | 
						|
 | 
						|
	newImageID := t.ImageID(newImageInfo.ID)
 | 
						|
	if newImageID == currentImageID {
 | 
						|
		log.Debugf("No new images found for %s", container.Name())
 | 
						|
		return false, currentImageID, nil
 | 
						|
	}
 | 
						|
 | 
						|
	log.Infof("Found new %s image (%s)", imageName, newImageID.ShortID())
 | 
						|
	return true, newImageID, nil
 | 
						|
}
 | 
						|
 | 
						|
// PullImage pulls the latest image for the supplied container, optionally skipping if it's digest can be confirmed
 | 
						|
// to match the one that the registry reports via a HEAD request
 | 
						|
func (client dockerClient) PullImage(ctx context.Context, container t.Container) error {
 | 
						|
	containerName := container.Name()
 | 
						|
	imageName := container.ImageName()
 | 
						|
 | 
						|
	fields := log.Fields{
 | 
						|
		"image":     imageName,
 | 
						|
		"container": containerName,
 | 
						|
	}
 | 
						|
 | 
						|
	if strings.HasPrefix(imageName, "sha256:") {
 | 
						|
		return fmt.Errorf("container uses a pinned image, and cannot be updated by watchtower")
 | 
						|
	}
 | 
						|
 | 
						|
	log.WithFields(fields).Debugf("Trying to load authentication credentials.")
 | 
						|
	opts, err := registry.GetPullOptions(imageName)
 | 
						|
	if err != nil {
 | 
						|
		log.Debugf("Error loading authentication credentials %s", err)
 | 
						|
		return err
 | 
						|
	}
 | 
						|
	if opts.RegistryAuth != "" {
 | 
						|
		log.Debug("Credentials loaded")
 | 
						|
	}
 | 
						|
 | 
						|
	log.WithFields(fields).Debugf("Checking if pull is needed")
 | 
						|
 | 
						|
	if match, err := digest.CompareDigest(container, opts.RegistryAuth); err != nil {
 | 
						|
		headLevel := log.DebugLevel
 | 
						|
		if client.WarnOnHeadPullFailed(container) {
 | 
						|
			headLevel = log.WarnLevel
 | 
						|
		}
 | 
						|
		log.WithFields(fields).Logf(headLevel, "Could not do a head request for %q, falling back to regular pull.", imageName)
 | 
						|
		log.WithFields(fields).Log(headLevel, "Reason: ", err)
 | 
						|
	} else if match {
 | 
						|
		log.Debug("No pull needed. Skipping image.")
 | 
						|
		return nil
 | 
						|
	} else {
 | 
						|
		log.Debug("Digests did not match, doing a pull.")
 | 
						|
	}
 | 
						|
 | 
						|
	log.WithFields(fields).Debugf("Pulling image")
 | 
						|
 | 
						|
	response, err := client.api.ImagePull(ctx, imageName, opts)
 | 
						|
	if err != nil {
 | 
						|
		log.Debugf("Error pulling image %s, %s", imageName, err)
 | 
						|
		return err
 | 
						|
	}
 | 
						|
 | 
						|
	defer response.Close()
 | 
						|
	// the pull request will be aborted prematurely unless the response is read
 | 
						|
	if _, err = ioutil.ReadAll(response); err != nil {
 | 
						|
		log.Error(err)
 | 
						|
		return err
 | 
						|
	}
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) GetImage(id t.ImageID) (types.ImageInspect, error) {
 | 
						|
	imageInfo, _, err := client.api.ImageInspectWithRaw(context.Background(), string(id))
 | 
						|
	if err != nil {
 | 
						|
		return types.ImageInspect{}, err
 | 
						|
	}
 | 
						|
	return imageInfo, nil
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) RemoveImageByID(id t.ImageID) error {
 | 
						|
	log.Infof("Removing image %s", id.ShortID())
 | 
						|
 | 
						|
	items, err := client.api.ImageRemove(
 | 
						|
		context.Background(),
 | 
						|
		string(id),
 | 
						|
		types.ImageRemoveOptions{
 | 
						|
			Force: true,
 | 
						|
		})
 | 
						|
 | 
						|
	if log.IsLevelEnabled(log.DebugLevel) {
 | 
						|
		deleted := strings.Builder{}
 | 
						|
		untagged := strings.Builder{}
 | 
						|
		for _, item := range items {
 | 
						|
			if item.Deleted != "" {
 | 
						|
				if deleted.Len() > 0 {
 | 
						|
					deleted.WriteString(`, `)
 | 
						|
				}
 | 
						|
				deleted.WriteString(t.ImageID(item.Deleted).ShortID())
 | 
						|
			}
 | 
						|
			if item.Untagged != "" {
 | 
						|
				if untagged.Len() > 0 {
 | 
						|
					untagged.WriteString(`, `)
 | 
						|
				}
 | 
						|
				untagged.WriteString(t.ImageID(item.Untagged).ShortID())
 | 
						|
			}
 | 
						|
		}
 | 
						|
		fields := log.Fields{`deleted`: deleted.String(), `untagged`: untagged.String()}
 | 
						|
		log.WithFields(fields).Debug("Image removal completed")
 | 
						|
	}
 | 
						|
 | 
						|
	return err
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) ExecuteCommand(containerID t.ContainerID, command string, timeout int) (SkipUpdate bool, err error) {
 | 
						|
	bg := context.Background()
 | 
						|
	clog := log.WithField("containerID", containerID)
 | 
						|
 | 
						|
	// Create the exec
 | 
						|
	execConfig := types.ExecConfig{
 | 
						|
		Tty:    true,
 | 
						|
		Detach: false,
 | 
						|
		Cmd:    []string{"sh", "-c", command},
 | 
						|
	}
 | 
						|
 | 
						|
	exec, err := client.api.ContainerExecCreate(bg, string(containerID), execConfig)
 | 
						|
	if err != nil {
 | 
						|
		return false, err
 | 
						|
	}
 | 
						|
 | 
						|
	response, attachErr := client.api.ContainerExecAttach(bg, exec.ID, types.ExecStartCheck{
 | 
						|
		Tty:    true,
 | 
						|
		Detach: false,
 | 
						|
	})
 | 
						|
	if attachErr != nil {
 | 
						|
		clog.Errorf("Failed to extract command exec logs: %v", attachErr)
 | 
						|
	}
 | 
						|
 | 
						|
	// Run the exec
 | 
						|
	execStartCheck := types.ExecStartCheck{Detach: false, Tty: true}
 | 
						|
	err = client.api.ContainerExecStart(bg, exec.ID, execStartCheck)
 | 
						|
	if err != nil {
 | 
						|
		return false, err
 | 
						|
	}
 | 
						|
 | 
						|
	var output string
 | 
						|
	if attachErr == nil {
 | 
						|
		defer response.Close()
 | 
						|
		var writer bytes.Buffer
 | 
						|
		written, err := writer.ReadFrom(response.Reader)
 | 
						|
		if err != nil {
 | 
						|
			clog.Error(err)
 | 
						|
		} else if written > 0 {
 | 
						|
			output = strings.TrimSpace(writer.String())
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	// Inspect the exec to get the exit code and print a message if the
 | 
						|
	// exit code is not success.
 | 
						|
	skipUpdate, err := client.waitForExecOrTimeout(bg, exec.ID, output, timeout)
 | 
						|
	if err != nil {
 | 
						|
		return true, err
 | 
						|
	}
 | 
						|
 | 
						|
	return skipUpdate, nil
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) waitForExecOrTimeout(bg context.Context, ID string, execOutput string, timeout int) (SkipUpdate bool, err error) {
 | 
						|
	const ExTempFail = 75
 | 
						|
	var ctx context.Context
 | 
						|
	var cancel context.CancelFunc
 | 
						|
 | 
						|
	if timeout > 0 {
 | 
						|
		ctx, cancel = context.WithTimeout(bg, time.Duration(timeout)*time.Minute)
 | 
						|
		defer cancel()
 | 
						|
	} else {
 | 
						|
		ctx = bg
 | 
						|
	}
 | 
						|
 | 
						|
	for {
 | 
						|
		execInspect, err := client.api.ContainerExecInspect(ctx, ID)
 | 
						|
 | 
						|
		//goland:noinspection GoNilness
 | 
						|
		log.WithFields(log.Fields{
 | 
						|
			"exit-code":    execInspect.ExitCode,
 | 
						|
			"exec-id":      execInspect.ExecID,
 | 
						|
			"running":      execInspect.Running,
 | 
						|
			"container-id": execInspect.ContainerID,
 | 
						|
		}).Debug("Awaiting timeout or completion")
 | 
						|
 | 
						|
		if err != nil {
 | 
						|
			return false, err
 | 
						|
		}
 | 
						|
		if execInspect.Running {
 | 
						|
			time.Sleep(1 * time.Second)
 | 
						|
			continue
 | 
						|
		}
 | 
						|
		if len(execOutput) > 0 {
 | 
						|
			log.Infof("Command output:\n%v", execOutput)
 | 
						|
		}
 | 
						|
 | 
						|
		if execInspect.ExitCode == ExTempFail {
 | 
						|
			return true, nil
 | 
						|
		}
 | 
						|
 | 
						|
		if execInspect.ExitCode > 0 {
 | 
						|
			return false, fmt.Errorf("command exited with code %v  %s", execInspect.ExitCode, execOutput)
 | 
						|
		}
 | 
						|
		break
 | 
						|
	}
 | 
						|
	return false, nil
 | 
						|
}
 | 
						|
 | 
						|
func (client dockerClient) waitForStopOrTimeout(c t.Container, waitTime time.Duration) error {
 | 
						|
	bg := context.Background()
 | 
						|
	timeout := time.After(waitTime)
 | 
						|
 | 
						|
	for {
 | 
						|
		select {
 | 
						|
		case <-timeout:
 | 
						|
			return nil
 | 
						|
		default:
 | 
						|
			if ci, err := client.api.ContainerInspect(bg, string(c.ID())); err != nil {
 | 
						|
				return err
 | 
						|
			} else if !ci.State.Running {
 | 
						|
				return nil
 | 
						|
			}
 | 
						|
		}
 | 
						|
		time.Sleep(1 * time.Second)
 | 
						|
	}
 | 
						|
}
 |