// Copyright 2018 Google Inc. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package libcontainer import ( "bufio" "fmt" "io" "io/ioutil" "os" "path" "strconv" "strings" "time" "github.com/google/cadvisor/container" info "github.com/google/cadvisor/info/v1" "bytes" "github.com/opencontainers/runc/libcontainer" "github.com/opencontainers/runc/libcontainer/cgroups" "k8s.io/klog" ) /* #include */ import "C" type Handler struct { cgroupManager cgroups.Manager rootFs string pid int includedMetrics container.MetricSet pidMetricsCache map[int]*info.CpuSchedstat } func NewHandler(cgroupManager cgroups.Manager, rootFs string, pid int, includedMetrics container.MetricSet) *Handler { return &Handler{ cgroupManager: cgroupManager, rootFs: rootFs, pid: pid, includedMetrics: includedMetrics, pidMetricsCache: make(map[int]*info.CpuSchedstat), } } // Get cgroup and networking stats of the specified container func (h *Handler) GetStats() (*info.ContainerStats, error) { cgroupStats, err := h.cgroupManager.GetStats() if err != nil { return nil, err } libcontainerStats := &libcontainer.Stats{ CgroupStats: cgroupStats, } stats := newContainerStats(libcontainerStats, h.includedMetrics) if h.includedMetrics.Has(container.ProcessSchedulerMetrics) { pids, err := h.cgroupManager.GetAllPids() if err != nil { klog.V(4).Infof("Could not get PIDs for container %d: %v", h.pid, err) } else { stats.Cpu.Schedstat, err = schedulerStatsFromProcs(h.rootFs, pids, h.pidMetricsCache) if err != nil { klog.V(4).Infof("Unable to get Process Scheduler Stats: %v", err) } } } // If we know the pid then get network stats from /proc//net/dev if h.pid == 0 { return stats, nil } if h.includedMetrics.Has(container.NetworkUsageMetrics) { netStats, err := networkStatsFromProc(h.rootFs, h.pid) if err != nil { klog.V(4).Infof("Unable to get network stats from pid %d: %v", h.pid, err) } else { stats.Network.Interfaces = append(stats.Network.Interfaces, netStats...) } } if h.includedMetrics.Has(container.NetworkTcpUsageMetrics) { t, err := tcpStatsFromProc(h.rootFs, h.pid, "net/tcp") if err != nil { klog.V(4).Infof("Unable to get tcp stats from pid %d: %v", h.pid, err) } else { stats.Network.Tcp = t } t6, err := tcpStatsFromProc(h.rootFs, h.pid, "net/tcp6") if err != nil { klog.V(4).Infof("Unable to get tcp6 stats from pid %d: %v", h.pid, err) } else { stats.Network.Tcp6 = t6 } } if h.includedMetrics.Has(container.NetworkUdpUsageMetrics) { u, err := udpStatsFromProc(h.rootFs, h.pid, "net/udp") if err != nil { klog.V(4).Infof("Unable to get udp stats from pid %d: %v", h.pid, err) } else { stats.Network.Udp = u } u6, err := udpStatsFromProc(h.rootFs, h.pid, "net/udp6") if err != nil { klog.V(4).Infof("Unable to get udp6 stats from pid %d: %v", h.pid, err) } else { stats.Network.Udp6 = u6 } } if h.includedMetrics.Has(container.ProcessMetrics) { paths := h.cgroupManager.GetPaths() path, ok := paths["cpu"] if !ok { klog.V(4).Infof("Could not find cgroups CPU for container %d", h.pid) } else { stats.Processes, err = processStatsFromProcs(h.rootFs, path) if err != nil { klog.V(4).Infof("Unable to get Process Stats: %v", err) } } } // For backwards compatibility. if len(stats.Network.Interfaces) > 0 { stats.Network.InterfaceStats = stats.Network.Interfaces[0] } return stats, nil } func processStatsFromProcs(rootFs string, cgroupPath string) (info.ProcessStats, error) { var fdCount uint64 filePath := path.Join(cgroupPath, "cgroup.procs") out, err := ioutil.ReadFile(filePath) if err != nil { return info.ProcessStats{}, fmt.Errorf("couldn't open cpu cgroup procs file %v : %v", filePath, err) } pids := strings.Split(string(out), "\n") // EOL is also treated as a new line while reading "cgroup.procs" file with ioutil.ReadFile. // The last value is an empty string "". Ex: pids = ["22", "1223", ""] // Trim the last value if len(pids) != 0 && pids[len(pids)-1] == "" { pids = pids[:len(pids)-1] } for _, pid := range pids { dirPath := path.Join(rootFs, "/proc", pid, "fd") fds, err := ioutil.ReadDir(dirPath) if err != nil { klog.V(4).Infof("error while listing directory %q to measure fd count: %v", dirPath, err) continue } fdCount += uint64(len(fds)) } processStats := info.ProcessStats{ ProcessCount: uint64(len(pids)), FdCount: fdCount, } return processStats, nil } func schedulerStatsFromProcs(rootFs string, pids []int, pidMetricsCache map[int]*info.CpuSchedstat) (info.CpuSchedstat, error) { for _, pid := range pids { f, err := os.Open(path.Join(rootFs, "proc", strconv.Itoa(pid), "schedstat")) if err != nil { return info.CpuSchedstat{}, fmt.Errorf("couldn't open scheduler statistics for process %d: %v", pid, err) } defer f.Close() contents, err := ioutil.ReadAll(f) if err != nil { return info.CpuSchedstat{}, fmt.Errorf("couldn't read scheduler statistics for process %d: %v", pid, err) } rawMetrics := bytes.Split(bytes.TrimRight(contents, "\n"), []byte(" ")) if len(rawMetrics) != 3 { return info.CpuSchedstat{}, fmt.Errorf("unexpected number of metrics in schedstat file for process %d", pid) } cacheEntry, ok := pidMetricsCache[pid] if !ok { cacheEntry = &info.CpuSchedstat{} pidMetricsCache[pid] = cacheEntry } for i, rawMetric := range rawMetrics { metric, err := strconv.ParseUint(string(rawMetric), 10, 64) if err != nil { return info.CpuSchedstat{}, fmt.Errorf("parsing error while reading scheduler statistics for process: %d: %v", pid, err) } switch i { case 0: cacheEntry.RunTime = metric case 1: cacheEntry.RunqueueTime = metric case 2: cacheEntry.RunPeriods = metric } } } schedstats := info.CpuSchedstat{} for _, v := range pidMetricsCache { schedstats.RunPeriods += v.RunPeriods schedstats.RunqueueTime += v.RunqueueTime schedstats.RunTime += v.RunTime } return schedstats, nil } func networkStatsFromProc(rootFs string, pid int) ([]info.InterfaceStats, error) { netStatsFile := path.Join(rootFs, "proc", strconv.Itoa(pid), "/net/dev") ifaceStats, err := scanInterfaceStats(netStatsFile) if err != nil { return []info.InterfaceStats{}, fmt.Errorf("couldn't read network stats: %v", err) } return ifaceStats, nil } var ( ignoredDevicePrefixes = []string{"lo", "veth", "docker"} ) func isIgnoredDevice(ifName string) bool { for _, prefix := range ignoredDevicePrefixes { if strings.HasPrefix(strings.ToLower(ifName), prefix) { return true } } return false } func scanInterfaceStats(netStatsFile string) ([]info.InterfaceStats, error) { file, err := os.Open(netStatsFile) if err != nil { return nil, fmt.Errorf("failure opening %s: %v", netStatsFile, err) } defer file.Close() scanner := bufio.NewScanner(file) // Discard header lines for i := 0; i < 2; i++ { if b := scanner.Scan(); !b { return nil, scanner.Err() } } stats := []info.InterfaceStats{} for scanner.Scan() { line := scanner.Text() line = strings.Replace(line, ":", "", -1) fields := strings.Fields(line) // If the format of the line is invalid then don't trust any of the stats // in this file. if len(fields) != 17 { return nil, fmt.Errorf("invalid interface stats line: %v", line) } devName := fields[0] if isIgnoredDevice(devName) { continue } i := info.InterfaceStats{ Name: devName, } statFields := append(fields[1:5], fields[9:13]...) statPointers := []*uint64{ &i.RxBytes, &i.RxPackets, &i.RxErrors, &i.RxDropped, &i.TxBytes, &i.TxPackets, &i.TxErrors, &i.TxDropped, } err := setInterfaceStatValues(statFields, statPointers) if err != nil { return nil, fmt.Errorf("cannot parse interface stats (%v): %v", err, line) } stats = append(stats, i) } return stats, nil } func setInterfaceStatValues(fields []string, pointers []*uint64) error { for i, v := range fields { val, err := strconv.ParseUint(v, 10, 64) if err != nil { return err } *pointers[i] = val } return nil } func tcpStatsFromProc(rootFs string, pid int, file string) (info.TcpStat, error) { tcpStatsFile := path.Join(rootFs, "proc", strconv.Itoa(pid), file) tcpStats, err := scanTcpStats(tcpStatsFile) if err != nil { return tcpStats, fmt.Errorf("couldn't read tcp stats: %v", err) } return tcpStats, nil } func scanTcpStats(tcpStatsFile string) (info.TcpStat, error) { var stats info.TcpStat data, err := ioutil.ReadFile(tcpStatsFile) if err != nil { return stats, fmt.Errorf("failure opening %s: %v", tcpStatsFile, err) } tcpStateMap := map[string]uint64{ "01": 0, //ESTABLISHED "02": 0, //SYN_SENT "03": 0, //SYN_RECV "04": 0, //FIN_WAIT1 "05": 0, //FIN_WAIT2 "06": 0, //TIME_WAIT "07": 0, //CLOSE "08": 0, //CLOSE_WAIT "09": 0, //LAST_ACK "0A": 0, //LISTEN "0B": 0, //CLOSING } reader := strings.NewReader(string(data)) scanner := bufio.NewScanner(reader) scanner.Split(bufio.ScanLines) // Discard header line if b := scanner.Scan(); !b { return stats, scanner.Err() } for scanner.Scan() { line := scanner.Text() state := strings.Fields(line) // TCP state is the 4th field. // Format: sl local_address rem_address st tx_queue rx_queue tr tm->when retrnsmt uid timeout inode tcpState := state[3] _, ok := tcpStateMap[tcpState] if !ok { return stats, fmt.Errorf("invalid TCP stats line: %v", line) } tcpStateMap[tcpState]++ } stats = info.TcpStat{ Established: tcpStateMap["01"], SynSent: tcpStateMap["02"], SynRecv: tcpStateMap["03"], FinWait1: tcpStateMap["04"], FinWait2: tcpStateMap["05"], TimeWait: tcpStateMap["06"], Close: tcpStateMap["07"], CloseWait: tcpStateMap["08"], LastAck: tcpStateMap["09"], Listen: tcpStateMap["0A"], Closing: tcpStateMap["0B"], } return stats, nil } func udpStatsFromProc(rootFs string, pid int, file string) (info.UdpStat, error) { var err error var udpStats info.UdpStat udpStatsFile := path.Join(rootFs, "proc", strconv.Itoa(pid), file) r, err := os.Open(udpStatsFile) if err != nil { return udpStats, fmt.Errorf("failure opening %s: %v", udpStatsFile, err) } udpStats, err = scanUdpStats(r) if err != nil { return udpStats, fmt.Errorf("couldn't read udp stats: %v", err) } return udpStats, nil } func scanUdpStats(r io.Reader) (info.UdpStat, error) { var stats info.UdpStat scanner := bufio.NewScanner(r) scanner.Split(bufio.ScanLines) // Discard header line if b := scanner.Scan(); !b { return stats, scanner.Err() } listening := uint64(0) dropped := uint64(0) rxQueued := uint64(0) txQueued := uint64(0) for scanner.Scan() { line := scanner.Text() // Format: sl local_address rem_address st tx_queue rx_queue tr tm->when retrnsmt uid timeout inode ref pointer drops listening++ fs := strings.Fields(line) if len(fs) != 13 { continue } rx, tx := uint64(0), uint64(0) fmt.Sscanf(fs[4], "%X:%X", &rx, &tx) rxQueued += rx txQueued += tx d, err := strconv.Atoi(string(fs[12])) if err != nil { continue } dropped += uint64(d) } stats = info.UdpStat{ Listen: listening, Dropped: dropped, RxQueued: rxQueued, TxQueued: txQueued, } return stats, nil } func (h *Handler) GetProcesses() ([]int, error) { pids, err := h.cgroupManager.GetPids() if err != nil { return nil, err } return pids, nil } func minUint32(x, y uint32) uint32 { if x < y { return x } return y } // var to allow unit tests to stub it out var numCpusFunc = getNumberOnlineCPUs // Convert libcontainer stats to info.ContainerStats. func setCpuStats(s *cgroups.Stats, ret *info.ContainerStats, withPerCPU bool) { ret.Cpu.Usage.User = s.CpuStats.CpuUsage.UsageInUsermode ret.Cpu.Usage.System = s.CpuStats.CpuUsage.UsageInKernelmode ret.Cpu.Usage.Total = s.CpuStats.CpuUsage.TotalUsage ret.Cpu.CFS.Periods = s.CpuStats.ThrottlingData.Periods ret.Cpu.CFS.ThrottledPeriods = s.CpuStats.ThrottlingData.ThrottledPeriods ret.Cpu.CFS.ThrottledTime = s.CpuStats.ThrottlingData.ThrottledTime if !withPerCPU { return } if len(s.CpuStats.CpuUsage.PercpuUsage) == 0 { // libcontainer's 'GetStats' can leave 'PercpuUsage' nil if it skipped the // cpuacct subsystem. return } numPossible := uint32(len(s.CpuStats.CpuUsage.PercpuUsage)) // Note that as of https://patchwork.kernel.org/patch/8607101/ (kernel v4.7), // the percpu usage information includes extra zero values for all additional // possible CPUs. This is to allow statistic collection after CPU-hotplug. // We intentionally ignore these extra zeroes. numActual, err := numCpusFunc() if err != nil { klog.Errorf("unable to determine number of actual cpus; defaulting to maximum possible number: errno %v", err) numActual = numPossible } if numActual > numPossible { // The real number of cores should never be greater than the number of // datapoints reported in cpu usage. klog.Errorf("PercpuUsage had %v cpus, but the actual number is %v; ignoring extra CPUs", numPossible, numActual) } numActual = minUint32(numPossible, numActual) ret.Cpu.Usage.PerCpu = make([]uint64, numActual) for i := uint32(0); i < numActual; i++ { ret.Cpu.Usage.PerCpu[i] = s.CpuStats.CpuUsage.PercpuUsage[i] } } // Copied from // https://github.com/moby/moby/blob/8b1adf55c2af329a4334f21d9444d6a169000c81/daemon/stats/collector_unix.go#L73 // Apache 2.0, Copyright Docker, Inc. func getNumberOnlineCPUs() (uint32, error) { i, err := C.sysconf(C._SC_NPROCESSORS_ONLN) // According to POSIX - errno is undefined after successful // sysconf, and can be non-zero in several cases, so look for // error in returned value not in errno. // (https://sourceware.org/bugzilla/show_bug.cgi?id=21536) if i == -1 { return 0, err } return uint32(i), nil } func setDiskIoStats(s *cgroups.Stats, ret *info.ContainerStats) { ret.DiskIo.IoServiceBytes = DiskStatsCopy(s.BlkioStats.IoServiceBytesRecursive) ret.DiskIo.IoServiced = DiskStatsCopy(s.BlkioStats.IoServicedRecursive) ret.DiskIo.IoQueued = DiskStatsCopy(s.BlkioStats.IoQueuedRecursive) ret.DiskIo.Sectors = DiskStatsCopy(s.BlkioStats.SectorsRecursive) ret.DiskIo.IoServiceTime = DiskStatsCopy(s.BlkioStats.IoServiceTimeRecursive) ret.DiskIo.IoWaitTime = DiskStatsCopy(s.BlkioStats.IoWaitTimeRecursive) ret.DiskIo.IoMerged = DiskStatsCopy(s.BlkioStats.IoMergedRecursive) ret.DiskIo.IoTime = DiskStatsCopy(s.BlkioStats.IoTimeRecursive) } func setMemoryStats(s *cgroups.Stats, ret *info.ContainerStats) { ret.Memory.Usage = s.MemoryStats.Usage.Usage ret.Memory.MaxUsage = s.MemoryStats.Usage.MaxUsage ret.Memory.Failcnt = s.MemoryStats.Usage.Failcnt if s.MemoryStats.UseHierarchy { ret.Memory.Cache = s.MemoryStats.Stats["total_cache"] ret.Memory.RSS = s.MemoryStats.Stats["total_rss"] ret.Memory.Swap = s.MemoryStats.Stats["total_swap"] ret.Memory.MappedFile = s.MemoryStats.Stats["total_mapped_file"] } else { ret.Memory.Cache = s.MemoryStats.Stats["cache"] ret.Memory.RSS = s.MemoryStats.Stats["rss"] ret.Memory.Swap = s.MemoryStats.Stats["swap"] ret.Memory.MappedFile = s.MemoryStats.Stats["mapped_file"] } if v, ok := s.MemoryStats.Stats["pgfault"]; ok { ret.Memory.ContainerData.Pgfault = v ret.Memory.HierarchicalData.Pgfault = v } if v, ok := s.MemoryStats.Stats["pgmajfault"]; ok { ret.Memory.ContainerData.Pgmajfault = v ret.Memory.HierarchicalData.Pgmajfault = v } workingSet := ret.Memory.Usage if v, ok := s.MemoryStats.Stats["total_inactive_file"]; ok { if workingSet < v { workingSet = 0 } else { workingSet -= v } } ret.Memory.WorkingSet = workingSet } func setNetworkStats(libcontainerStats *libcontainer.Stats, ret *info.ContainerStats) { ret.Network.Interfaces = make([]info.InterfaceStats, len(libcontainerStats.Interfaces)) for i := range libcontainerStats.Interfaces { ret.Network.Interfaces[i] = info.InterfaceStats{ Name: libcontainerStats.Interfaces[i].Name, RxBytes: libcontainerStats.Interfaces[i].RxBytes, RxPackets: libcontainerStats.Interfaces[i].RxPackets, RxErrors: libcontainerStats.Interfaces[i].RxErrors, RxDropped: libcontainerStats.Interfaces[i].RxDropped, TxBytes: libcontainerStats.Interfaces[i].TxBytes, TxPackets: libcontainerStats.Interfaces[i].TxPackets, TxErrors: libcontainerStats.Interfaces[i].TxErrors, TxDropped: libcontainerStats.Interfaces[i].TxDropped, } } // Add to base struct for backwards compatibility. if len(ret.Network.Interfaces) > 0 { ret.Network.InterfaceStats = ret.Network.Interfaces[0] } } func newContainerStats(libcontainerStats *libcontainer.Stats, includedMetrics container.MetricSet) *info.ContainerStats { ret := &info.ContainerStats{ Timestamp: time.Now(), } if s := libcontainerStats.CgroupStats; s != nil { setCpuStats(s, ret, includedMetrics.Has(container.PerCpuUsageMetrics)) if includedMetrics.Has(container.DiskIOMetrics) { setDiskIoStats(s, ret) } setMemoryStats(s, ret) } if len(libcontainerStats.Interfaces) > 0 { setNetworkStats(libcontainerStats, ret) } return ret }