Browse Source

successfully exporting one of 2 cpus

pull/310/head
stuart nelson 8 years ago
parent
commit
03da1ea516
  1. 154
      collector/cpu_dragonfly.go

154
collector/cpu_dragonfly.go

@ -17,64 +17,113 @@ package collector
import (
"errors"
"strconv"
"unsafe"
"fmt"
"github.com/prometheus/client_golang/prometheus"
)
/*
#cgo LDFLAGS:
#include <stdio.h>
#include <kinfo.h>
#include <fcntl.h>
#include <stdlib.h>
#include <sys/param.h>
#include <sys/resource.h>
#include <sys/time.h>
#include <sys/sysctl.h>
#include <kinfo.h>
static int mibs_set_up = 0;
static int mib_kern_cp_times[2];
static size_t mib_kern_cp_times_len = 2;
static const int mib_hw_ncpu[] = {CTL_HW, HW_NCPU};
static const size_t mib_hw_ncpu_len = 2;
int
getCPUTimes(int *ncpu, double **cpu_times, size_t *cp_times_length)
{
int cpu, mib[2];
struct kinfo_cputime cp_t[ncpu];
uint64_t user, nice, sys, intr, idle;
static const int mib_kern_clockrate[] = {CTL_KERN, KERN_CLOCKRATE};
static size_t mib_kern_clockrate_len = 2;
// Setup method for MIBs not available as constants.
// Calls to this method must be synchronized externally.
int setupSysctlMIBs() {
int ret = sysctlnametomib("kern.cputime", mib_kern_cp_times, &mib_kern_cp_times_len);
if (ret == 0) mibs_set_up = 1;
return ret;
}
int getCPUTimes(int *ncpu, struct kinfo_cputime *cputime, uint64_t *cpu_user) {
// // Assert that mibs are set up through setupSysctlMIBs
// if (!mibs_set_up) {
// return -1;
// }
// // Retrieve number of cpu cores
// size_t ncpu_size = sizeof(*ncpu);
// if (sysctl(mib_hw_ncpu, mib_hw_ncpu_len, ncpu, &ncpu_size, NULL, 0) == -1 ||
// sizeof(*ncpu) != ncpu_size) {
// return -1;
// }
size_t len;
// Get number of cpu cores.
int mib[2];
mib[0] = CTL_HW;
mib[1] = HW_NCPU;
len = sizeof(*ncpu);
if (-1 == sysctl(mib, 2, &ncpu, &len, NULL, 0))
return -1;
bzero(cp_t, sizeof(struct kinfo_cputime)*(*ncpu));
len = sizeof(cp_t[0])*(*ncpu);
if (sysctlbyname("kern.cputime", &cp_t, &len, NULL, 0))
if (sysctl(mib, 2, ncpu, &len, NULL, 0)) {
return -1;
}
// Retrieve clockrate
struct clockinfo clockrate;
int mib_kern_clockrate[] = {CTL_KERN, KERN_CLOCKRATE};
size_t mib_kern_clockrate_len = 2;
size_t clockrate_size = sizeof(clockrate);
if (sysctl(mib_kern_clockrate, mib_kern_clockrate_len, &clockrate, &clockrate_size, NULL, 0) == -1)
if (sysctl(mib_kern_clockrate, mib_kern_clockrate_len, &clockrate, &clockrate_size, NULL, 0) == -1 ||
sizeof(clockrate) != clockrate_size) {
return -1;
}
// // Retrieve cp_times values
// *cp_times_length = (*ncpu) * CPUSTATES;
//
// long cp_times[*cp_times_length];
// size_t cp_times_size = sizeof(cp_times);
long cpufreq = clockrate.stathz > 0 ? clockrate.stathz : clockrate.hz;
cpu_times = (double *) malloc(sizeof(double)*(*cp_times_length));
for (int i = 0; i < (*cp_times_length); i++) {
(*cpu_times)[i] = ((double) cp_times[i]) / cpufreq;
// Get the cpu times.
struct kinfo_cputime cp_t[*ncpu];
bzero(cp_t, sizeof(struct kinfo_cputime)*(*ncpu));
len = sizeof(cp_t[0])*(*ncpu);
if (sysctlbyname("kern.cputime", &cp_t, &len, NULL, 0)) {
return -1;
}
*cpu_user = cp_t[0].cp_user;
*cputime = cp_t[0];
// This results in outputting:
// {1362514572 273421 845667973 12986861 17529717536 0 0 0 0 [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]}
// So, we have the first 5 numbers from the first cpu.
// Need to figure out how to get the second cpu, i.e. cputime needs to be created with [2]
// Compute absolute time for different CPU states
// long cpufreq = clockrate.stathz > 0 ? clockrate.stathz : clockrate.hz;
// *cpu_times = (double *) malloc(sizeof(double)*(len));
// for (int i = 0; i < (len); i++) {
// (*cpu_times)[i] = ((double) cp_t[i]) / cpufreq;
// }
return 0;
}
void freeCPUTimes(double *cpu_times) {
free(cpu_times);
}
*/
*/
import "C"
const maxCPUTimesLen = C.MAXCPU * C.CPUSTATES
type statCollector struct {
cpu *prometheus.CounterVec
}
@ -86,6 +135,9 @@ func init() {
// Takes a prometheus registry and returns a new Collector exposing
// CPU stats.
func NewStatCollector() (Collector, error) {
if C.setupSysctlMIBs() == -1 {
return nil, errors.New("could not initialize sysctl MIBs")
}
return &statCollector{
cpu: prometheus.NewCounterVec(
prometheus.CounterOpts{
@ -98,30 +150,50 @@ func NewStatCollector() (Collector, error) {
}, nil
}
type kinfoCPUTime struct {
cp_user, cp_nice, cp_sys, cp_intr, cp_idle uint64
}
// Expose CPU stats using sysctl.
func (c *statCollector) Update(ch chan<- prometheus.Metric) (err error) {
// Adapted from
// https://www.dragonflybsd.org/mailarchive/users/2010-04/msg00056.html
// We want time spent per-cpu per CPUSTATE.
// CPUSTATES (number of CPUSTATES) is defined as 5U.
// Order: CP_USER | CP_NICE | CP_SYS | CP_IDLE | CP_INTR
// sysctl kern.cp_times provides hw.ncpu * CPUSTATES long integers:
// hw.ncpu * (space-separated list of the above variables)
//
// Each value is a counter incremented at frequency
// kern.clockrate.(stathz | hz)
//
// Look into sys/kern/kern_clock.c for details.
var ncpu C.int
var cpuTimesC *C.double
var cpuTimesLength C.size_t
var cpuTimesC C.struct_kinfo_cputime
var cpuTimesLength C.uint64_t
if C.getCPUTimes(&ncpu, &cpuTimesC, &cpuTimesLength) == -1 {
return errors.New("could not retrieve CPU times")
}
defer C.freeCPUTimes(cpuTimesC)
// defer C.freeCPUTimes(cpuTimesC)
fmt.Println(cpuTimesC)
fmt.Println(uint64(cpuTimesLength))
return errors.New("early kill")
if cpuTimesLength > maxCPUTimesLen {
return errors.New("more CPU's than MAXCPU?")
}
// Convert C.double array to Go array (https://github.com/golang/go/wiki/cgo#turning-c-arrays-into-go-slices).
cpuTimes := (*[maxCPUTimesLen]C.double)(unsafe.Pointer(cpuTimesC))[:cpuTimesLength:cpuTimesLength]
for cpu := 0; cpu < int(ncpu); cpu++ {
base_idx := C.CPUSTATES * cpu
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "user"}).Set(float64(cpuTimes[base_idx+C.CP_USER]))
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "nice"}).Set(float64(cpuTimes[base_idx+C.CP_NICE]))
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "system"}).Set(float64(cpuTimes[base_idx+C.CP_SYS]))
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "interrupt"}).Set(float64(cpuTimes[base_idx+C.CP_INTR]))
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "idle"}).Set(float64(cpuTimes[base_idx+C.CP_IDLE]))
}
// cpuTimes := (*[maxCPUTimesLen]C.double)(unsafe.Pointer(cpuTimesC))[:cpuTimesLength:cpuTimesLength]
//
// for cpu := 0; cpu < int(ncpu); cpu++ {
// base_idx := C.CPUSTATES * cpu
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "user"}).Set(float64(cpuTimes[base_idx+C.CP_USER]))
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "nice"}).Set(float64(cpuTimes[base_idx+C.CP_NICE]))
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "system"}).Set(float64(cpuTimes[base_idx+C.CP_SYS]))
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "interrupt"}).Set(float64(cpuTimes[base_idx+C.CP_INTR]))
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "idle"}).Set(float64(cpuTimes[base_idx+C.CP_IDLE]))
// }
c.cpu.Collect(ch)
return err

Loading…
Cancel
Save