d222bf097c
The existing runtimes reload logic went to great lengths to replace the directory containing runtime wrapper scripts as atomically as possible within the limitations of the Linux filesystem ABI. Trouble is, atomically swapping the wrapper scripts directory solves the wrong problem! The runtime configuration is "locked in" when a container is started, including the path to the runC binary. If a container is started with a runtime which requires a daemon-managed wrapper script and then the daemon is reloaded with a config which no longer requires the wrapper script (i.e. some args -> no args, or the runtime is dropped from the config), that container would become unmanageable. Any attempts to stop, exec or otherwise perform lifecycle management operations on the container are likely to fail due to the wrapper script no longer existing at its original path. Atomically swapping the wrapper scripts is also incompatible with the read-copy-update paradigm for reloading configuration. A handler in the daemon could retain a reference to the pre-reload configuration for an indeterminate amount of time after the daemon configuration has been reloaded and updated. It is possible for the daemon to attempt to start a container using a deleted wrapper script if a request to run a container races a reload. Solve the problem of deleting referenced wrapper scripts by ensuring that all wrapper scripts are *immutable* for the lifetime of the daemon process. Any given runtime wrapper script must always exist with the same contents, no matter how many times the daemon config is reloaded, or what changes are made to the config. This is accomplished by using everyone's favourite design pattern: content-addressable storage. Each wrapper script file name is suffixed with the SHA-256 digest of its contents to (probabilistically) guarantee immutability without needing any concurrency control. Stale runtime wrapper scripts are only cleaned up on the next daemon restart. Split the derived runtimes configuration from the user-supplied configuration to have a place to store derived state without mutating the user-supplied configuration or exposing daemon internals in API struct types. Hold the derived state and the user-supplied configuration in a single struct value so that they can be updated as an atomic unit. Signed-off-by: Cory Snider <csnider@mirantis.com>
284 lines
9.2 KiB
Go
284 lines
9.2 KiB
Go
package daemon // import "github.com/docker/docker/daemon"
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"time"
|
|
|
|
"github.com/docker/docker/api/types"
|
|
"github.com/docker/docker/api/types/backend"
|
|
networktypes "github.com/docker/docker/api/types/network"
|
|
"github.com/docker/docker/api/types/versions"
|
|
"github.com/docker/docker/api/types/versions/v1p20"
|
|
"github.com/docker/docker/container"
|
|
"github.com/docker/docker/daemon/config"
|
|
"github.com/docker/docker/daemon/network"
|
|
"github.com/docker/docker/errdefs"
|
|
"github.com/docker/go-connections/nat"
|
|
)
|
|
|
|
// ContainerInspect returns low-level information about a
|
|
// container. Returns an error if the container cannot be found, or if
|
|
// there is an error getting the data.
|
|
func (daemon *Daemon) ContainerInspect(ctx context.Context, name string, size bool, version string) (interface{}, error) {
|
|
switch {
|
|
case versions.LessThan(version, "1.20"):
|
|
return daemon.containerInspectPre120(ctx, name)
|
|
case versions.Equal(version, "1.20"):
|
|
return daemon.containerInspect120(name)
|
|
}
|
|
return daemon.ContainerInspectCurrent(ctx, name, size)
|
|
}
|
|
|
|
// ContainerInspectCurrent returns low-level information about a
|
|
// container in a most recent api version.
|
|
func (daemon *Daemon) ContainerInspectCurrent(ctx context.Context, name string, size bool) (*types.ContainerJSON, error) {
|
|
ctr, err := daemon.GetContainer(name)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
ctr.Lock()
|
|
|
|
base, err := daemon.getInspectData(&daemon.config().Config, ctr)
|
|
if err != nil {
|
|
ctr.Unlock()
|
|
return nil, err
|
|
}
|
|
|
|
apiNetworks := make(map[string]*networktypes.EndpointSettings)
|
|
for nwName, epConf := range ctr.NetworkSettings.Networks {
|
|
if epConf.EndpointSettings != nil {
|
|
// We must make a copy of this pointer object otherwise it can race with other operations
|
|
apiNetworks[nwName] = epConf.EndpointSettings.Copy()
|
|
}
|
|
}
|
|
|
|
mountPoints := ctr.GetMountPoints()
|
|
networkSettings := &types.NetworkSettings{
|
|
NetworkSettingsBase: types.NetworkSettingsBase{
|
|
Bridge: ctr.NetworkSettings.Bridge,
|
|
SandboxID: ctr.NetworkSettings.SandboxID,
|
|
HairpinMode: ctr.NetworkSettings.HairpinMode,
|
|
LinkLocalIPv6Address: ctr.NetworkSettings.LinkLocalIPv6Address,
|
|
LinkLocalIPv6PrefixLen: ctr.NetworkSettings.LinkLocalIPv6PrefixLen,
|
|
SandboxKey: ctr.NetworkSettings.SandboxKey,
|
|
SecondaryIPAddresses: ctr.NetworkSettings.SecondaryIPAddresses,
|
|
SecondaryIPv6Addresses: ctr.NetworkSettings.SecondaryIPv6Addresses,
|
|
},
|
|
DefaultNetworkSettings: daemon.getDefaultNetworkSettings(ctr.NetworkSettings.Networks),
|
|
Networks: apiNetworks,
|
|
}
|
|
|
|
ports := make(nat.PortMap, len(ctr.NetworkSettings.Ports))
|
|
for k, pm := range ctr.NetworkSettings.Ports {
|
|
ports[k] = pm
|
|
}
|
|
networkSettings.NetworkSettingsBase.Ports = ports
|
|
|
|
ctr.Unlock()
|
|
|
|
if size {
|
|
sizeRw, sizeRootFs, err := daemon.imageService.GetContainerLayerSize(ctx, base.ID)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
base.SizeRw = &sizeRw
|
|
base.SizeRootFs = &sizeRootFs
|
|
}
|
|
|
|
return &types.ContainerJSON{
|
|
ContainerJSONBase: base,
|
|
Mounts: mountPoints,
|
|
Config: ctr.Config,
|
|
NetworkSettings: networkSettings,
|
|
}, nil
|
|
}
|
|
|
|
// containerInspect120 serializes the master version of a container into a json type.
|
|
func (daemon *Daemon) containerInspect120(name string) (*v1p20.ContainerJSON, error) {
|
|
ctr, err := daemon.GetContainer(name)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
ctr.Lock()
|
|
defer ctr.Unlock()
|
|
|
|
base, err := daemon.getInspectData(&daemon.config().Config, ctr)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return &v1p20.ContainerJSON{
|
|
ContainerJSONBase: base,
|
|
Mounts: ctr.GetMountPoints(),
|
|
Config: &v1p20.ContainerConfig{
|
|
Config: ctr.Config,
|
|
MacAddress: ctr.Config.MacAddress,
|
|
NetworkDisabled: ctr.Config.NetworkDisabled,
|
|
ExposedPorts: ctr.Config.ExposedPorts,
|
|
VolumeDriver: ctr.HostConfig.VolumeDriver,
|
|
},
|
|
NetworkSettings: daemon.getBackwardsCompatibleNetworkSettings(ctr.NetworkSettings),
|
|
}, nil
|
|
}
|
|
|
|
func (daemon *Daemon) getInspectData(daemonCfg *config.Config, container *container.Container) (*types.ContainerJSONBase, error) {
|
|
// make a copy to play with
|
|
hostConfig := *container.HostConfig
|
|
|
|
children := daemon.children(container)
|
|
hostConfig.Links = nil // do not expose the internal structure
|
|
for linkAlias, child := range children {
|
|
hostConfig.Links = append(hostConfig.Links, fmt.Sprintf("%s:%s", child.Name, linkAlias))
|
|
}
|
|
|
|
// We merge the Ulimits from hostConfig with daemon default
|
|
daemon.mergeUlimits(&hostConfig, daemonCfg)
|
|
|
|
var containerHealth *types.Health
|
|
if container.State.Health != nil {
|
|
containerHealth = &types.Health{
|
|
Status: container.State.Health.Status(),
|
|
FailingStreak: container.State.Health.FailingStreak,
|
|
Log: append([]*types.HealthcheckResult{}, container.State.Health.Log...),
|
|
}
|
|
}
|
|
|
|
containerState := &types.ContainerState{
|
|
Status: container.State.StateString(),
|
|
Running: container.State.Running,
|
|
Paused: container.State.Paused,
|
|
Restarting: container.State.Restarting,
|
|
OOMKilled: container.State.OOMKilled,
|
|
Dead: container.State.Dead,
|
|
Pid: container.State.Pid,
|
|
ExitCode: container.State.ExitCode(),
|
|
Error: container.State.ErrorMsg,
|
|
StartedAt: container.State.StartedAt.Format(time.RFC3339Nano),
|
|
FinishedAt: container.State.FinishedAt.Format(time.RFC3339Nano),
|
|
Health: containerHealth,
|
|
}
|
|
|
|
contJSONBase := &types.ContainerJSONBase{
|
|
ID: container.ID,
|
|
Created: container.Created.Format(time.RFC3339Nano),
|
|
Path: container.Path,
|
|
Args: container.Args,
|
|
State: containerState,
|
|
Image: container.ImageID.String(),
|
|
LogPath: container.LogPath,
|
|
Name: container.Name,
|
|
RestartCount: container.RestartCount,
|
|
Driver: container.Driver,
|
|
Platform: container.OS,
|
|
MountLabel: container.MountLabel,
|
|
ProcessLabel: container.ProcessLabel,
|
|
ExecIDs: container.GetExecIDs(),
|
|
HostConfig: &hostConfig,
|
|
}
|
|
|
|
// Now set any platform-specific fields
|
|
contJSONBase = setPlatformSpecificContainerFields(container, contJSONBase)
|
|
|
|
contJSONBase.GraphDriver.Name = container.Driver
|
|
|
|
if daemon.UsesSnapshotter() {
|
|
// Additional information only applies to graphDrivers, so we're done.
|
|
return contJSONBase, nil
|
|
}
|
|
|
|
if container.RWLayer == nil {
|
|
if container.Dead {
|
|
return contJSONBase, nil
|
|
}
|
|
return nil, errdefs.System(errors.New("RWLayer of container " + container.ID + " is unexpectedly nil"))
|
|
}
|
|
|
|
graphDriverData, err := container.RWLayer.Metadata()
|
|
if err != nil {
|
|
if container.Dead {
|
|
// container is marked as Dead, and its graphDriver metadata may
|
|
// have been removed; we can ignore errors.
|
|
return contJSONBase, nil
|
|
}
|
|
return nil, errdefs.System(err)
|
|
}
|
|
|
|
contJSONBase.GraphDriver.Data = graphDriverData
|
|
return contJSONBase, nil
|
|
}
|
|
|
|
// ContainerExecInspect returns low-level information about the exec
|
|
// command. An error is returned if the exec cannot be found.
|
|
func (daemon *Daemon) ContainerExecInspect(id string) (*backend.ExecInspect, error) {
|
|
e := daemon.execCommands.Get(id)
|
|
if e == nil {
|
|
return nil, errExecNotFound(id)
|
|
}
|
|
|
|
if ctr := daemon.containers.Get(e.Container.ID); ctr == nil {
|
|
return nil, errExecNotFound(id)
|
|
}
|
|
|
|
e.Lock()
|
|
defer e.Unlock()
|
|
pc := inspectExecProcessConfig(e)
|
|
var pid int
|
|
if e.Process != nil {
|
|
pid = int(e.Process.Pid())
|
|
}
|
|
|
|
return &backend.ExecInspect{
|
|
ID: e.ID,
|
|
Running: e.Running,
|
|
ExitCode: e.ExitCode,
|
|
ProcessConfig: pc,
|
|
OpenStdin: e.OpenStdin,
|
|
OpenStdout: e.OpenStdout,
|
|
OpenStderr: e.OpenStderr,
|
|
CanRemove: e.CanRemove,
|
|
ContainerID: e.Container.ID,
|
|
DetachKeys: e.DetachKeys,
|
|
Pid: pid,
|
|
}, nil
|
|
}
|
|
|
|
func (daemon *Daemon) getBackwardsCompatibleNetworkSettings(settings *network.Settings) *v1p20.NetworkSettings {
|
|
result := &v1p20.NetworkSettings{
|
|
NetworkSettingsBase: types.NetworkSettingsBase{
|
|
Bridge: settings.Bridge,
|
|
SandboxID: settings.SandboxID,
|
|
HairpinMode: settings.HairpinMode,
|
|
LinkLocalIPv6Address: settings.LinkLocalIPv6Address,
|
|
LinkLocalIPv6PrefixLen: settings.LinkLocalIPv6PrefixLen,
|
|
Ports: settings.Ports,
|
|
SandboxKey: settings.SandboxKey,
|
|
SecondaryIPAddresses: settings.SecondaryIPAddresses,
|
|
SecondaryIPv6Addresses: settings.SecondaryIPv6Addresses,
|
|
},
|
|
DefaultNetworkSettings: daemon.getDefaultNetworkSettings(settings.Networks),
|
|
}
|
|
|
|
return result
|
|
}
|
|
|
|
// getDefaultNetworkSettings creates the deprecated structure that holds the information
|
|
// about the bridge network for a container.
|
|
func (daemon *Daemon) getDefaultNetworkSettings(networks map[string]*network.EndpointSettings) types.DefaultNetworkSettings {
|
|
var settings types.DefaultNetworkSettings
|
|
|
|
if defaultNetwork, ok := networks["bridge"]; ok && defaultNetwork.EndpointSettings != nil {
|
|
settings.EndpointID = defaultNetwork.EndpointID
|
|
settings.Gateway = defaultNetwork.Gateway
|
|
settings.GlobalIPv6Address = defaultNetwork.GlobalIPv6Address
|
|
settings.GlobalIPv6PrefixLen = defaultNetwork.GlobalIPv6PrefixLen
|
|
settings.IPAddress = defaultNetwork.IPAddress
|
|
settings.IPPrefixLen = defaultNetwork.IPPrefixLen
|
|
settings.IPv6Gateway = defaultNetwork.IPv6Gateway
|
|
settings.MacAddress = defaultNetwork.MacAddress
|
|
}
|
|
return settings
|
|
}
|