edad52707c
Reuse existing structures and rely on json serialization to deep copy Container objects. Also consolidate all "save" operations on container.CheckpointTo, which now both saves a serialized json to disk, and replicates state to the ACID in-memory store. Signed-off-by: Fabio Kung <fabio.kung@gmail.com>
86 lines
2.4 KiB
Go
86 lines
2.4 KiB
Go
package daemon
|
|
|
|
import (
|
|
"fmt"
|
|
|
|
"github.com/docker/docker/api/types/container"
|
|
)
|
|
|
|
// ContainerUpdate updates configuration of the container
|
|
func (daemon *Daemon) ContainerUpdate(name string, hostConfig *container.HostConfig) (container.ContainerUpdateOKBody, error) {
|
|
var warnings []string
|
|
|
|
warnings, err := daemon.verifyContainerSettings(hostConfig, nil, true)
|
|
if err != nil {
|
|
return container.ContainerUpdateOKBody{Warnings: warnings}, err
|
|
}
|
|
|
|
if err := daemon.update(name, hostConfig); err != nil {
|
|
return container.ContainerUpdateOKBody{Warnings: warnings}, err
|
|
}
|
|
|
|
return container.ContainerUpdateOKBody{Warnings: warnings}, nil
|
|
}
|
|
|
|
func (daemon *Daemon) update(name string, hostConfig *container.HostConfig) error {
|
|
if hostConfig == nil {
|
|
return nil
|
|
}
|
|
|
|
container, err := daemon.GetContainer(name)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
restoreConfig := false
|
|
backupHostConfig := *container.HostConfig
|
|
defer func() {
|
|
if restoreConfig {
|
|
container.Lock()
|
|
container.HostConfig = &backupHostConfig
|
|
container.CheckpointTo(daemon.containersReplica)
|
|
container.Unlock()
|
|
}
|
|
}()
|
|
|
|
if container.RemovalInProgress || container.Dead {
|
|
return errCannotUpdate(container.ID, fmt.Errorf("Container is marked for removal and cannot be \"update\"."))
|
|
}
|
|
|
|
container.Lock()
|
|
if err := container.UpdateContainer(hostConfig); err != nil {
|
|
restoreConfig = true
|
|
container.Unlock()
|
|
return errCannotUpdate(container.ID, err)
|
|
}
|
|
if err := container.CheckpointTo(daemon.containersReplica); err != nil {
|
|
restoreConfig = true
|
|
container.Unlock()
|
|
return errCannotUpdate(container.ID, err)
|
|
}
|
|
container.Unlock()
|
|
|
|
// if Restart Policy changed, we need to update container monitor
|
|
if hostConfig.RestartPolicy.Name != "" {
|
|
container.UpdateMonitor(hostConfig.RestartPolicy)
|
|
}
|
|
|
|
// If container is not running, update hostConfig struct is enough,
|
|
// resources will be updated when the container is started again.
|
|
// If container is running (including paused), we need to update configs
|
|
// to the real world.
|
|
if container.IsRunning() && !container.IsRestarting() {
|
|
if err := daemon.containerd.UpdateResources(container.ID, toContainerdResources(hostConfig.Resources)); err != nil {
|
|
restoreConfig = true
|
|
return errCannotUpdate(container.ID, err)
|
|
}
|
|
}
|
|
|
|
daemon.LogContainerEvent(container, "update")
|
|
|
|
return nil
|
|
}
|
|
|
|
func errCannotUpdate(containerID string, err error) error {
|
|
return fmt.Errorf("Cannot update container %s: %v", containerID, err)
|
|
}
|