e7f4963e73
Terminating the exec process when the context is canceled has been
broken since Docker v17.11 so nobody has been able to depend upon that
behaviour in five years of releases. We are thus free from backwards-
compatibility constraints.
Co-authored-by: Nicolas De Loof <nicolas.deloof@gmail.com>
Co-authored-by: Sebastiaan van Stijn <github@gone.nl>
Signed-off-by: Nicolas De Loof <nicolas.deloof@gmail.com>
Signed-off-by: Cory Snider <csnider@mirantis.com>
Signed-off-by: Sebastiaan van Stijn <github@gone.nl>
(cherry picked from commit 4b84a33217
)
Signed-off-by: Sebastiaan van Stijn <github@gone.nl>
146 lines
4.6 KiB
Go
146 lines
4.6 KiB
Go
package container // import "github.com/docker/docker/integration/container"
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/docker/docker/api/types"
|
|
containertypes "github.com/docker/docker/api/types/container"
|
|
"github.com/docker/docker/client"
|
|
"github.com/docker/docker/integration/internal/container"
|
|
"gotest.tools/v3/assert"
|
|
"gotest.tools/v3/poll"
|
|
"gotest.tools/v3/skip"
|
|
)
|
|
|
|
// TestHealthCheckWorkdir verifies that health-checks inherit the containers'
|
|
// working-dir.
|
|
func TestHealthCheckWorkdir(t *testing.T) {
|
|
skip.If(t, testEnv.OSType == "windows", "FIXME")
|
|
defer setupTest(t)()
|
|
ctx := context.Background()
|
|
client := testEnv.APIClient()
|
|
|
|
cID := container.Run(ctx, t, client, container.WithTty(true), container.WithWorkingDir("/foo"), func(c *container.TestContainerConfig) {
|
|
c.Config.Healthcheck = &containertypes.HealthConfig{
|
|
Test: []string{"CMD-SHELL", "if [ \"$PWD\" = \"/foo\" ]; then exit 0; else exit 1; fi;"},
|
|
Interval: 50 * time.Millisecond,
|
|
Retries: 3,
|
|
}
|
|
})
|
|
|
|
poll.WaitOn(t, pollForHealthStatus(ctx, client, cID, types.Healthy), poll.WithDelay(100*time.Millisecond))
|
|
}
|
|
|
|
// GitHub #37263
|
|
// Do not stop healthchecks just because we sent a signal to the container
|
|
func TestHealthKillContainer(t *testing.T) {
|
|
skip.If(t, testEnv.OSType == "windows", "Windows only supports SIGKILL and SIGTERM? See https://github.com/moby/moby/issues/39574")
|
|
defer setupTest(t)()
|
|
|
|
ctx := context.Background()
|
|
client := testEnv.APIClient()
|
|
|
|
id := container.Run(ctx, t, client, func(c *container.TestContainerConfig) {
|
|
cmd := `
|
|
# Set the initial HEALTH value so the healthcheck passes
|
|
HEALTH="1"
|
|
echo $HEALTH > /health
|
|
|
|
# Any time doHealth is run we flip the value
|
|
# This lets us use kill signals to determine when healtchecks have run.
|
|
doHealth() {
|
|
case "$HEALTH" in
|
|
"0")
|
|
HEALTH="1"
|
|
;;
|
|
"1")
|
|
HEALTH="0"
|
|
;;
|
|
esac
|
|
echo $HEALTH > /health
|
|
}
|
|
|
|
trap 'doHealth' USR1
|
|
|
|
while true; do sleep 1; done
|
|
`
|
|
c.Config.Cmd = []string{"/bin/sh", "-c", cmd}
|
|
c.Config.Healthcheck = &containertypes.HealthConfig{
|
|
Test: []string{"CMD-SHELL", `[ "$(cat /health)" = "1" ]`},
|
|
Interval: time.Second,
|
|
Retries: 5,
|
|
}
|
|
})
|
|
|
|
ctxPoll, cancel := context.WithTimeout(ctx, 30*time.Second)
|
|
defer cancel()
|
|
poll.WaitOn(t, pollForHealthStatus(ctxPoll, client, id, "healthy"), poll.WithDelay(100*time.Millisecond))
|
|
|
|
err := client.ContainerKill(ctx, id, "SIGUSR1")
|
|
assert.NilError(t, err)
|
|
|
|
ctxPoll, cancel = context.WithTimeout(ctx, 30*time.Second)
|
|
defer cancel()
|
|
poll.WaitOn(t, pollForHealthStatus(ctxPoll, client, id, "unhealthy"), poll.WithDelay(100*time.Millisecond))
|
|
|
|
err = client.ContainerKill(ctx, id, "SIGUSR1")
|
|
assert.NilError(t, err)
|
|
|
|
ctxPoll, cancel = context.WithTimeout(ctx, 30*time.Second)
|
|
defer cancel()
|
|
poll.WaitOn(t, pollForHealthStatus(ctxPoll, client, id, "healthy"), poll.WithDelay(100*time.Millisecond))
|
|
}
|
|
|
|
// TestHealthCheckProcessKilled verifies that health-checks exec get killed on time-out.
|
|
func TestHealthCheckProcessKilled(t *testing.T) {
|
|
skip.If(t, testEnv.RuntimeIsWindowsContainerd(), "FIXME: Broken on Windows + containerd combination")
|
|
defer setupTest(t)()
|
|
ctx := context.Background()
|
|
apiClient := testEnv.APIClient()
|
|
|
|
cID := container.Run(ctx, t, apiClient, func(c *container.TestContainerConfig) {
|
|
c.Config.Healthcheck = &containertypes.HealthConfig{
|
|
Test: []string{"CMD", "sh", "-c", "sleep 60"},
|
|
Interval: 100 * time.Millisecond,
|
|
Timeout: 50 * time.Millisecond,
|
|
Retries: 1,
|
|
}
|
|
})
|
|
poll.WaitOn(t, pollForHealthCheckLog(ctx, apiClient, cID, "Health check exceeded timeout (50ms)"))
|
|
}
|
|
|
|
func pollForHealthCheckLog(ctx context.Context, client client.APIClient, containerID string, expected string) func(log poll.LogT) poll.Result {
|
|
return func(log poll.LogT) poll.Result {
|
|
inspect, err := client.ContainerInspect(ctx, containerID)
|
|
if err != nil {
|
|
return poll.Error(err)
|
|
}
|
|
healthChecksTotal := len(inspect.State.Health.Log)
|
|
if healthChecksTotal > 0 {
|
|
output := inspect.State.Health.Log[healthChecksTotal-1].Output
|
|
if output == expected {
|
|
return poll.Success()
|
|
}
|
|
return poll.Error(fmt.Errorf("expected %q, got %q", expected, output))
|
|
}
|
|
return poll.Continue("waiting for container healthcheck logs")
|
|
}
|
|
}
|
|
|
|
func pollForHealthStatus(ctx context.Context, client client.APIClient, containerID string, healthStatus string) func(log poll.LogT) poll.Result {
|
|
return func(log poll.LogT) poll.Result {
|
|
inspect, err := client.ContainerInspect(ctx, containerID)
|
|
|
|
switch {
|
|
case err != nil:
|
|
return poll.Error(err)
|
|
case inspect.State.Health.Status == healthStatus:
|
|
return poll.Success()
|
|
default:
|
|
return poll.Continue("waiting for container to become %s", healthStatus)
|
|
}
|
|
}
|
|
}
|