//go:build !windows package daemon // import "github.com/docker/docker/daemon" import ( "bufio" "context" "fmt" "os" "strconv" "strings" statsV1 "github.com/containerd/cgroups/v3/cgroup1/stats" statsV2 "github.com/containerd/cgroups/v3/cgroup2/stats" "github.com/docker/docker/api/types" "github.com/docker/docker/container" "github.com/pkg/errors" ) func copyBlkioEntry(entries []*statsV1.BlkIOEntry) []types.BlkioStatEntry { out := make([]types.BlkioStatEntry, len(entries)) for i, re := range entries { out[i] = types.BlkioStatEntry{ Major: re.Major, Minor: re.Minor, Op: re.Op, Value: re.Value, } } return out } func (daemon *Daemon) stats(c *container.Container) (*types.StatsJSON, error) { c.Lock() task, err := c.GetRunningTask() c.Unlock() if err != nil { return nil, err } cs, err := task.Stats(context.Background()) if err != nil { if strings.Contains(err.Error(), "container not found") { return nil, containerNotFound(c.ID) } return nil, err } s := &types.StatsJSON{} s.Read = cs.Read stats := cs.Metrics switch t := stats.(type) { case *statsV1.Metrics: return daemon.statsV1(s, t) case *statsV2.Metrics: return daemon.statsV2(s, t) default: return nil, errors.Errorf("unexpected type of metrics %+v", t) } } func (daemon *Daemon) statsV1(s *types.StatsJSON, stats *statsV1.Metrics) (*types.StatsJSON, error) { if stats.Blkio != nil { s.BlkioStats = types.BlkioStats{ IoServiceBytesRecursive: copyBlkioEntry(stats.Blkio.IoServiceBytesRecursive), IoServicedRecursive: copyBlkioEntry(stats.Blkio.IoServicedRecursive), IoQueuedRecursive: copyBlkioEntry(stats.Blkio.IoQueuedRecursive), IoServiceTimeRecursive: copyBlkioEntry(stats.Blkio.IoServiceTimeRecursive), IoWaitTimeRecursive: copyBlkioEntry(stats.Blkio.IoWaitTimeRecursive), IoMergedRecursive: copyBlkioEntry(stats.Blkio.IoMergedRecursive), IoTimeRecursive: copyBlkioEntry(stats.Blkio.IoTimeRecursive), SectorsRecursive: copyBlkioEntry(stats.Blkio.SectorsRecursive), } } if stats.CPU != nil { s.CPUStats = types.CPUStats{ CPUUsage: types.CPUUsage{ TotalUsage: stats.CPU.Usage.Total, PercpuUsage: stats.CPU.Usage.PerCPU, UsageInKernelmode: stats.CPU.Usage.Kernel, UsageInUsermode: stats.CPU.Usage.User, }, ThrottlingData: types.ThrottlingData{ Periods: stats.CPU.Throttling.Periods, ThrottledPeriods: stats.CPU.Throttling.ThrottledPeriods, ThrottledTime: stats.CPU.Throttling.ThrottledTime, }, } } if stats.Memory != nil { raw := map[string]uint64{ "cache": stats.Memory.Cache, "rss": stats.Memory.RSS, "rss_huge": stats.Memory.RSSHuge, "mapped_file": stats.Memory.MappedFile, "dirty": stats.Memory.Dirty, "writeback": stats.Memory.Writeback, "pgpgin": stats.Memory.PgPgIn, "pgpgout": stats.Memory.PgPgOut, "pgfault": stats.Memory.PgFault, "pgmajfault": stats.Memory.PgMajFault, "inactive_anon": stats.Memory.InactiveAnon, "active_anon": stats.Memory.ActiveAnon, "inactive_file": stats.Memory.InactiveFile, "active_file": stats.Memory.ActiveFile, "unevictable": stats.Memory.Unevictable, "hierarchical_memory_limit": stats.Memory.HierarchicalMemoryLimit, "hierarchical_memsw_limit": stats.Memory.HierarchicalSwapLimit, "total_cache": stats.Memory.TotalCache, "total_rss": stats.Memory.TotalRSS, "total_rss_huge": stats.Memory.TotalRSSHuge, "total_mapped_file": stats.Memory.TotalMappedFile, "total_dirty": stats.Memory.TotalDirty, "total_writeback": stats.Memory.TotalWriteback, "total_pgpgin": stats.Memory.TotalPgPgIn, "total_pgpgout": stats.Memory.TotalPgPgOut, "total_pgfault": stats.Memory.TotalPgFault, "total_pgmajfault": stats.Memory.TotalPgMajFault, "total_inactive_anon": stats.Memory.TotalInactiveAnon, "total_active_anon": stats.Memory.TotalActiveAnon, "total_inactive_file": stats.Memory.TotalInactiveFile, "total_active_file": stats.Memory.TotalActiveFile, "total_unevictable": stats.Memory.TotalUnevictable, } if stats.Memory.Usage != nil { s.MemoryStats = types.MemoryStats{ Stats: raw, Usage: stats.Memory.Usage.Usage, MaxUsage: stats.Memory.Usage.Max, Limit: stats.Memory.Usage.Limit, Failcnt: stats.Memory.Usage.Failcnt, } } else { s.MemoryStats = types.MemoryStats{ Stats: raw, } } // if the container does not set memory limit, use the machineMemory if s.MemoryStats.Limit > daemon.machineMemory && daemon.machineMemory > 0 { s.MemoryStats.Limit = daemon.machineMemory } } if stats.Pids != nil { s.PidsStats = types.PidsStats{ Current: stats.Pids.Current, Limit: stats.Pids.Limit, } } return s, nil } func (daemon *Daemon) statsV2(s *types.StatsJSON, stats *statsV2.Metrics) (*types.StatsJSON, error) { if stats.Io != nil { var isbr []types.BlkioStatEntry for _, re := range stats.Io.Usage { isbr = append(isbr, types.BlkioStatEntry{ Major: re.Major, Minor: re.Minor, Op: "read", Value: re.Rbytes, }, types.BlkioStatEntry{ Major: re.Major, Minor: re.Minor, Op: "write", Value: re.Wbytes, }, ) } s.BlkioStats = types.BlkioStats{ IoServiceBytesRecursive: isbr, // Other fields are unsupported } } if stats.CPU != nil { s.CPUStats = types.CPUStats{ CPUUsage: types.CPUUsage{ TotalUsage: stats.CPU.UsageUsec * 1000, // PercpuUsage is not supported UsageInKernelmode: stats.CPU.SystemUsec * 1000, UsageInUsermode: stats.CPU.UserUsec * 1000, }, ThrottlingData: types.ThrottlingData{ Periods: stats.CPU.NrPeriods, ThrottledPeriods: stats.CPU.NrThrottled, ThrottledTime: stats.CPU.ThrottledUsec * 1000, }, } } if stats.Memory != nil { s.MemoryStats = types.MemoryStats{ // Stats is not compatible with v1 Stats: map[string]uint64{ "anon": stats.Memory.Anon, "file": stats.Memory.File, "kernel_stack": stats.Memory.KernelStack, "slab": stats.Memory.Slab, "sock": stats.Memory.Sock, "shmem": stats.Memory.Shmem, "file_mapped": stats.Memory.FileMapped, "file_dirty": stats.Memory.FileDirty, "file_writeback": stats.Memory.FileWriteback, "anon_thp": stats.Memory.AnonThp, "inactive_anon": stats.Memory.InactiveAnon, "active_anon": stats.Memory.ActiveAnon, "inactive_file": stats.Memory.InactiveFile, "active_file": stats.Memory.ActiveFile, "unevictable": stats.Memory.Unevictable, "slab_reclaimable": stats.Memory.SlabReclaimable, "slab_unreclaimable": stats.Memory.SlabUnreclaimable, "pgfault": stats.Memory.Pgfault, "pgmajfault": stats.Memory.Pgmajfault, "workingset_refault": stats.Memory.WorkingsetRefault, "workingset_activate": stats.Memory.WorkingsetActivate, "workingset_nodereclaim": stats.Memory.WorkingsetNodereclaim, "pgrefill": stats.Memory.Pgrefill, "pgscan": stats.Memory.Pgscan, "pgsteal": stats.Memory.Pgsteal, "pgactivate": stats.Memory.Pgactivate, "pgdeactivate": stats.Memory.Pgdeactivate, "pglazyfree": stats.Memory.Pglazyfree, "pglazyfreed": stats.Memory.Pglazyfreed, "thp_fault_alloc": stats.Memory.ThpFaultAlloc, "thp_collapse_alloc": stats.Memory.ThpCollapseAlloc, }, Usage: stats.Memory.Usage, // MaxUsage is not supported Limit: stats.Memory.UsageLimit, } // if the container does not set memory limit, use the machineMemory if s.MemoryStats.Limit > daemon.machineMemory && daemon.machineMemory > 0 { s.MemoryStats.Limit = daemon.machineMemory } if stats.MemoryEvents != nil { // Failcnt is set to the "oom" field of the "memory.events" file. // See https://www.kernel.org/doc/html/latest/admin-guide/cgroup-v2.html s.MemoryStats.Failcnt = stats.MemoryEvents.Oom } } if stats.Pids != nil { s.PidsStats = types.PidsStats{ Current: stats.Pids.Current, Limit: stats.Pids.Limit, } } return s, nil } // Resolve Network SandboxID in case the container reuse another container's network stack func (daemon *Daemon) getNetworkSandboxID(c *container.Container) (string, error) { curr := c for curr.HostConfig.NetworkMode.IsContainer() { containerID := curr.HostConfig.NetworkMode.ConnectedContainer() connected, err := daemon.GetContainer(containerID) if err != nil { return "", errors.Wrapf(err, "Could not get container for %s", containerID) } curr = connected } return curr.NetworkSettings.SandboxID, nil } func (daemon *Daemon) getNetworkStats(c *container.Container) (map[string]types.NetworkStats, error) { sandboxID, err := daemon.getNetworkSandboxID(c) if err != nil { return nil, err } sb, err := daemon.netController.SandboxByID(sandboxID) if err != nil { return nil, err } lnstats, err := sb.Statistics() if err != nil { return nil, err } stats := make(map[string]types.NetworkStats) // Convert libnetwork nw stats into api stats for ifName, ifStats := range lnstats { stats[ifName] = types.NetworkStats{ RxBytes: ifStats.RxBytes, RxPackets: ifStats.RxPackets, RxErrors: ifStats.RxErrors, RxDropped: ifStats.RxDropped, TxBytes: ifStats.TxBytes, TxPackets: ifStats.TxPackets, TxErrors: ifStats.TxErrors, TxDropped: ifStats.TxDropped, } } return stats, nil } const ( // The value comes from `C.sysconf(C._SC_CLK_TCK)`, and // on Linux it's a constant which is safe to be hard coded, // so we can avoid using cgo here. For details, see: // https://github.com/containerd/cgroups/pull/12 clockTicksPerSecond = 100 nanoSecondsPerSecond = 1e9 ) // getSystemCPUUsage returns the host system's cpu usage in // nanoseconds and number of online CPUs. An error is returned // if the format of the underlying file does not match. // // Uses /proc/stat defined by POSIX. Looks for the cpu // statistics line and then sums up the first seven fields // provided. See `man 5 proc` for details on specific field // information. func getSystemCPUUsage() (cpuUsage uint64, cpuNum uint32, err error) { f, err := os.Open("/proc/stat") if err != nil { return 0, 0, err } defer f.Close() scanner := bufio.NewScanner(f) for scanner.Scan() { line := scanner.Text() if len(line) < 4 || line[:3] != "cpu" { break // Assume all cpu* records are at the front, like glibc https://github.com/bminor/glibc/blob/5d00c201b9a2da768a79ea8d5311f257871c0b43/sysdeps/unix/sysv/linux/getsysstats.c#L108-L135 } if line[3] == ' ' { parts := strings.Fields(line) if len(parts) < 8 { return 0, 0, fmt.Errorf("invalid number of cpu fields") } var totalClockTicks uint64 for _, i := range parts[1:8] { v, err := strconv.ParseUint(i, 10, 64) if err != nil { return 0, 0, fmt.Errorf("Unable to convert value %s to int: %w", i, err) } totalClockTicks += v } cpuUsage = (totalClockTicks * nanoSecondsPerSecond) / clockTicksPerSecond } if '0' <= line[3] && line[3] <= '9' { cpuNum++ } } if err := scanner.Err(); err != nil { return 0, 0, fmt.Errorf("error scanning '/proc/stat' file: %w", err) } return }