mirror of https://github.com/statping/statping
647 lines
17 KiB
Go
647 lines
17 KiB
Go
package services
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"crypto/tls"
|
|
"fmt"
|
|
"github.com/statping/statping/types/downtimes"
|
|
"net"
|
|
"net/http"
|
|
"net/url"
|
|
"regexp"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/prometheus/client_golang/prometheus"
|
|
"github.com/statping/statping/types/metrics"
|
|
"google.golang.org/grpc"
|
|
"google.golang.org/grpc/credentials"
|
|
|
|
"github.com/statping/statping/types/failures"
|
|
"github.com/statping/statping/types/hits"
|
|
"github.com/statping/statping/utils"
|
|
healthpb "google.golang.org/grpc/health/grpc_health_v1"
|
|
)
|
|
|
|
// checkServices will start the checking go routine for each service
|
|
func CheckServices() {
|
|
go refreshAllServices()
|
|
log.Infoln(fmt.Sprintf("Starting monitoring process for %v Services", len(allServices)))
|
|
for _, s := range allServices {
|
|
time.Sleep(50 * time.Millisecond)
|
|
go ServiceCheckQueue(s, true)
|
|
}
|
|
}
|
|
|
|
func refreshAllServices() {
|
|
for {
|
|
time.Sleep(time.Duration(time.Second * 20))
|
|
|
|
newList := all()
|
|
|
|
for _, s := range newList {
|
|
if _, found := allServices[s.Id]; !found {
|
|
allServices[s.Id] = s
|
|
go ServiceCheckQueue(s, true)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// CheckQueue is the main go routine for checking a service
|
|
func ServiceCheckQueue(s *Service, record bool) {
|
|
s.Start()
|
|
s.SleepDuration = (time.Duration(s.Id) * 100) * time.Millisecond
|
|
|
|
CheckLoop:
|
|
for {
|
|
s.Checkpoint = utils.Now()
|
|
select {
|
|
case <-s.Running:
|
|
log.Infoln(fmt.Sprintf("Stopping service: %v", s.Name))
|
|
break CheckLoop
|
|
case <-time.After(s.SleepDuration):
|
|
|
|
s.SleepDuration = s.Duration()
|
|
|
|
sid := s.Id
|
|
|
|
err := s.acquireServiceRun()
|
|
s, er := Find(s.Id)
|
|
|
|
if er == nil {
|
|
if err == nil && !s.ManualDowntime {
|
|
|
|
log.Infof("Service Run Started : %s %s", s.Id, s.Name)
|
|
|
|
ce := s.CheckService(record)
|
|
//s.UpdateStats()
|
|
s.HandleDowntime(ce, record)
|
|
s.markServiceRunProcessed()
|
|
}
|
|
|
|
s.Checkpoint = s.Checkpoint.Add(s.Duration())
|
|
if s.Online && err == nil {
|
|
s.SleepDuration = s.Checkpoint.Sub(time.Now())
|
|
}
|
|
} else {
|
|
delete(allServices, sid)
|
|
break CheckLoop
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
func parseHost(s *Service) string {
|
|
if s.Type == "tcp" || s.Type == "udp" || s.Type == "grpc" {
|
|
return s.Domain
|
|
} else {
|
|
u, err := url.Parse(s.Domain)
|
|
if err != nil {
|
|
return s.Domain
|
|
}
|
|
return strings.Split(u.Host, ":")[0]
|
|
}
|
|
}
|
|
|
|
// dnsCheck will check the domain name and return a float64 for the amount of time the DNS check took
|
|
func dnsCheck(s *Service) (int64, error) {
|
|
var err error
|
|
t1 := utils.Now()
|
|
host := parseHost(s)
|
|
if s.Type == "tcp" || s.Type == "udp" || s.Type == "grpc" {
|
|
_, err = net.LookupHost(host)
|
|
} else {
|
|
_, err = net.LookupIP(host)
|
|
}
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
return utils.Now().Sub(t1).Microseconds(), err
|
|
}
|
|
|
|
func isIPv6(address string) bool {
|
|
return strings.Count(address, ":") >= 2
|
|
}
|
|
|
|
// checkIcmp will send a ICMP ping packet to the service
|
|
func CheckIcmp(s *Service, record bool) (*Service, error) {
|
|
defer s.updateLastCheck()
|
|
timer := prometheus.NewTimer(metrics.ServiceTimer(s.Name))
|
|
defer timer.ObserveDuration()
|
|
|
|
dur, err := utils.Ping(s.Domain, s.Timeout)
|
|
if err != nil {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("Could not send ICMP to service %v, %v", s.Domain, err), "lookup")
|
|
}
|
|
return s, err
|
|
}
|
|
|
|
s.PingTime = dur
|
|
s.Latency = dur
|
|
s.LastResponse = ""
|
|
//s.Online = true
|
|
if record {
|
|
RecordSuccess(s)
|
|
}
|
|
return s, nil
|
|
}
|
|
|
|
// CheckGrpc will check a gRPC service
|
|
func CheckGrpc(s *Service, record bool) (*Service, error) {
|
|
defer s.updateLastCheck()
|
|
timer := prometheus.NewTimer(metrics.ServiceTimer(s.Name))
|
|
defer timer.ObserveDuration()
|
|
|
|
// Strip URL scheme if present. Eg: https:// , http://
|
|
if strings.Contains(s.Domain, "://") {
|
|
u, err := url.Parse(s.Domain)
|
|
if err != nil {
|
|
// Unable to parse.
|
|
log.Warnln(fmt.Sprintf("GRPC Service: '%s', Unable to parse URL: '%v'", s.Name, s.Domain))
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("Unable to parse GRPC domain %v, %v", s.Domain, err), "parse_domain")
|
|
}
|
|
}
|
|
|
|
// Set domain as hostname without port number.
|
|
s.Domain = u.Hostname()
|
|
}
|
|
|
|
// Calculate DNS check time
|
|
dnsLookup, err := dnsCheck(s)
|
|
if err != nil {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("Could not get IP address for GRPC service %v, %v", s.Domain, err), "lookup")
|
|
}
|
|
return s, err
|
|
}
|
|
|
|
// Connect to grpc service without TLS certs.
|
|
grpcOption := grpc.WithInsecure()
|
|
|
|
// Check if TLS is enabled
|
|
// Upgrade GRPC connection if using TLS
|
|
// Force to connect on HTTP2 with TLS. Needed when using a reverse proxy such as nginx.
|
|
if s.VerifySSL.Bool {
|
|
h2creds := credentials.NewTLS(&tls.Config{NextProtos: []string{"h2"}})
|
|
grpcOption = grpc.WithTransportCredentials(h2creds)
|
|
}
|
|
|
|
s.PingTime = dnsLookup
|
|
t1 := utils.Now()
|
|
domain := fmt.Sprintf("%v", s.Domain)
|
|
if s.Port != 0 {
|
|
domain = fmt.Sprintf("%v:%v", s.Domain, s.Port)
|
|
if isIPv6(s.Domain) {
|
|
domain = fmt.Sprintf("[%v]:%v", s.Domain, s.Port)
|
|
}
|
|
}
|
|
|
|
// Context will cancel the request when timeout is exceeded.
|
|
// Cancel the context when request is served within the timeout limit.
|
|
timeout := time.Duration(s.Timeout) * time.Second
|
|
ctx, cancel := context.WithTimeout(context.Background(), timeout)
|
|
defer cancel()
|
|
|
|
conn, err := grpc.DialContext(ctx, domain, grpcOption, grpc.WithBlock())
|
|
if err != nil {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("Dial Error %v", err), "connection")
|
|
}
|
|
return s, err
|
|
}
|
|
|
|
if s.GrpcHealthCheck.Bool {
|
|
// Create a new health check client
|
|
c := healthpb.NewHealthClient(conn)
|
|
in := &healthpb.HealthCheckRequest{}
|
|
res, err := c.Check(ctx, in)
|
|
if err != nil {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("GRPC Error %v", err), "healthcheck")
|
|
}
|
|
return s, nil
|
|
}
|
|
|
|
// Record responses
|
|
s.LastResponse = strings.TrimSpace(res.String())
|
|
s.LastStatusCode = int(res.GetStatus())
|
|
}
|
|
|
|
if err := conn.Close(); err != nil {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("%v Socket Close Error %v", strings.ToUpper(s.Type), err), "close")
|
|
}
|
|
return s, err
|
|
}
|
|
|
|
// Record latency
|
|
s.Latency = utils.Now().Sub(t1).Microseconds()
|
|
//s.Online = true
|
|
|
|
if s.GrpcHealthCheck.Bool {
|
|
if s.ExpectedStatus != s.LastStatusCode {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("GRPC Service: '%s', Status Code: expected '%v', got '%v'", s.Name, s.ExpectedStatus, s.LastStatusCode), "response_code")
|
|
}
|
|
return s, nil
|
|
}
|
|
|
|
if s.Expected.String != s.LastResponse {
|
|
log.Warnln(fmt.Sprintf("GRPC Service: '%s', Response: expected '%v', got '%v'", s.Name, s.Expected.String, s.LastResponse))
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("GRPC Response Body '%v' did not match '%v'", s.LastResponse, s.Expected.String), "response_body")
|
|
}
|
|
return s, nil
|
|
}
|
|
}
|
|
|
|
if record {
|
|
RecordSuccess(s)
|
|
}
|
|
|
|
return s, nil
|
|
}
|
|
|
|
// checkTcp will check a TCP service
|
|
func CheckTcp(s *Service, record bool) (*Service, error) {
|
|
defer s.updateLastCheck()
|
|
timer := prometheus.NewTimer(metrics.ServiceTimer(s.Name))
|
|
defer timer.ObserveDuration()
|
|
|
|
dnsLookup, err := dnsCheck(s)
|
|
if err != nil {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("Could not get IP address for TCP service %v, %v", s.Domain, err), "lookup")
|
|
}
|
|
return s, err
|
|
}
|
|
s.PingTime = dnsLookup
|
|
t1 := utils.Now()
|
|
domain := fmt.Sprintf("%v", s.Domain)
|
|
if s.Port != 0 {
|
|
domain = fmt.Sprintf("%v:%v", s.Domain, s.Port)
|
|
if isIPv6(s.Domain) {
|
|
domain = fmt.Sprintf("[%v]:%v", s.Domain, s.Port)
|
|
}
|
|
}
|
|
|
|
tlsConfig, err := s.LoadTLSCert()
|
|
if err != nil {
|
|
log.Errorln(err)
|
|
}
|
|
|
|
// test TCP connection if there is no TLS Certificate set
|
|
if s.TLSCert.String == "" {
|
|
conn, err := net.DialTimeout(s.Type, domain, time.Duration(s.Timeout)*time.Second)
|
|
if err != nil {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("Dial Error: %v", err), "tls")
|
|
}
|
|
return s, err
|
|
}
|
|
defer conn.Close()
|
|
} else {
|
|
// test TCP connection if TLS Certificate was set
|
|
dialer := &net.Dialer{
|
|
KeepAlive: time.Duration(s.Timeout) * time.Second,
|
|
Timeout: time.Duration(s.Timeout) * time.Second,
|
|
}
|
|
conn, err := tls.DialWithDialer(dialer, s.Type, domain, tlsConfig)
|
|
if err != nil {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("Dial Error: %v", err), "tls")
|
|
}
|
|
return s, err
|
|
}
|
|
defer conn.Close()
|
|
}
|
|
|
|
s.Latency = utils.Now().Sub(t1).Microseconds()
|
|
s.LastResponse = ""
|
|
//s.Online = true
|
|
if record {
|
|
RecordSuccess(s)
|
|
}
|
|
return s, nil
|
|
}
|
|
|
|
func (s *Service) updateLastCheck() {
|
|
s.LastCheck = time.Now()
|
|
}
|
|
|
|
// checkHttp will check a HTTP service
|
|
func CheckHttp(s *Service, record bool) (*Service, error) {
|
|
defer s.updateLastCheck()
|
|
timer := prometheus.NewTimer(metrics.ServiceTimer(s.Name))
|
|
defer timer.ObserveDuration()
|
|
|
|
dnsLookup, err := dnsCheck(s)
|
|
if err != nil {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("Could not get IP address for domain %v, %v", s.Domain, err), "lookup")
|
|
}
|
|
return s, err
|
|
}
|
|
s.PingTime = dnsLookup
|
|
t1 := utils.Now()
|
|
|
|
timeout := time.Duration(s.Timeout) * time.Second
|
|
var content []byte
|
|
var res *http.Response
|
|
var data *bytes.Buffer
|
|
var headers []string
|
|
contentType := "application/json" // default Content-Type
|
|
|
|
if s.Headers.Valid {
|
|
headers = strings.Split(s.Headers.String, ",")
|
|
} else {
|
|
headers = nil
|
|
}
|
|
|
|
// check if 'Content-Type' header was defined
|
|
for _, header := range headers {
|
|
if strings.Split(header, "=")[0] == "Content-Type" {
|
|
contentType = strings.Split(header, "=")[1]
|
|
break
|
|
}
|
|
}
|
|
|
|
if s.Redirect.Bool {
|
|
headers = append(headers, "Redirect=true")
|
|
}
|
|
|
|
if s.PostData.String != "" {
|
|
data = bytes.NewBuffer([]byte(s.PostData.String))
|
|
} else {
|
|
data = bytes.NewBuffer(nil)
|
|
}
|
|
|
|
// force set Content-Type to 'application/json' if requests are made
|
|
// with POST method
|
|
if s.Method == "POST" && contentType != "application/json" {
|
|
contentType = "application/json"
|
|
}
|
|
|
|
customTLS, err := s.LoadTLSCert()
|
|
if err != nil {
|
|
log.Errorln(err)
|
|
}
|
|
|
|
content, res, err = utils.HttpRequest(s.Domain, s.Method, contentType, headers, data, timeout, s.VerifySSL.Bool, customTLS)
|
|
if err != nil {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("HTTP Error %v", err), "request")
|
|
}
|
|
return s, err
|
|
}
|
|
s.Latency = utils.Now().Sub(t1).Microseconds()
|
|
s.LastResponse = string(content)
|
|
s.LastStatusCode = res.StatusCode
|
|
|
|
metrics.Gauge("status_code", float64(res.StatusCode), s.Name)
|
|
|
|
if s.Expected.String != "" {
|
|
match, err := regexp.MatchString(s.Expected.String, string(content))
|
|
if err != nil {
|
|
log.Warnln(fmt.Sprintf("Service %v expected: %v to match %v", s.Name, string(content), s.Expected.String))
|
|
}
|
|
if !match {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("HTTP Response Body did not match '%v'", s.Expected), "regex")
|
|
}
|
|
return s, fmt.Errorf("HTTP Response Body did not match '%v'", s.Expected)
|
|
}
|
|
}
|
|
if s.ExpectedStatus != res.StatusCode {
|
|
if record {
|
|
RecordFailure(s, fmt.Sprintf("HTTP Status Code %v did not match %v", res.StatusCode, s.ExpectedStatus), "status_code")
|
|
}
|
|
return s, fmt.Errorf("HTTP Status Code %v did not match %v", res.StatusCode, s.ExpectedStatus)
|
|
}
|
|
if record {
|
|
RecordSuccess(s)
|
|
}
|
|
//s.Online = true
|
|
return s, nil
|
|
}
|
|
|
|
func CheckCollection(s *Service, record bool) (*Service, error) {
|
|
defer s.updateLastCheck()
|
|
timer := prometheus.NewTimer(metrics.ServiceTimer(s.Name))
|
|
defer timer.ObserveDuration()
|
|
|
|
hcStartTime := time.Now()
|
|
|
|
combinedStatus := STATUS_UP
|
|
var impactedSubService SubService
|
|
downCount := 0
|
|
|
|
for id, subServiceDetail := range s.SubServicesDetails {
|
|
if subService, err := FindFirstFromDB(id); err != nil {
|
|
log.Errorf("[Ignored]Failed to find Sub Service : %s %s %s %s", s.Id, s.Name, id, subServiceDetail.DisplayName)
|
|
continue
|
|
} else {
|
|
if !subService.Online && subService.CurrentDowntime > 0 {
|
|
downtimeType := STATUS_DOWN
|
|
if d, de := downtimes.Find(subService.CurrentDowntime); de != nil {
|
|
log.Errorf("[Ignored]Failed to find Sub Service Downtime : %s %s %s %s", s.Id, s.Name, id, subServiceDetail.DisplayName)
|
|
continue
|
|
} else {
|
|
downtimeType = d.SubStatus
|
|
}
|
|
|
|
if combinedStatus != STATUS_DOWN {
|
|
switch subServiceDetail.DependencyType {
|
|
case CRITICAL:
|
|
combinedStatus = HandleEmptyStatus(downtimeType)
|
|
impactedSubService = subServiceDetail
|
|
case DELAYED, PARTIAL:
|
|
combinedStatus = STATUS_DEGRADED
|
|
if downtimeType == STATUS_DOWN {
|
|
downCount++
|
|
}
|
|
impactedSubService = subServiceDetail
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if combinedStatus == STATUS_DEGRADED && downCount == len(s.SubServicesDetails) {
|
|
combinedStatus = STATUS_DOWN
|
|
}
|
|
|
|
s.Latency = time.Now().Sub(hcStartTime).Milliseconds()
|
|
s.PingTime = time.Now().Sub(hcStartTime).Milliseconds()
|
|
s.LastFailureType = combinedStatus
|
|
if combinedStatus == STATUS_DOWN || combinedStatus == STATUS_DEGRADED {
|
|
if record {
|
|
RecordFailureWithType(s, fmt.Sprintf("Sub Service Impacted : %s", impactedSubService.DisplayName), "", combinedStatus)
|
|
}
|
|
return s, fmt.Errorf("Sub Service Impacted: %s %s %s", s.Id, s.Name, impactedSubService.DisplayName)
|
|
}
|
|
|
|
if record {
|
|
RecordSuccess(s)
|
|
}
|
|
log.Infof("Collection Check Done : %s %s %s %s", s.Id, s.Name, s.LastFailureType)
|
|
//s.Online = true
|
|
return s, nil
|
|
}
|
|
|
|
// RecordSuccess will create a new 'hit' record in the database for a successful/online service
|
|
func RecordSuccess(s *Service) {
|
|
s.LastOnline = utils.Now()
|
|
//s.Online = true
|
|
hit := &hits.Hit{
|
|
Service: s.Id,
|
|
Latency: s.Latency,
|
|
PingTime: s.PingTime,
|
|
CreatedAt: utils.Now(),
|
|
}
|
|
if err := hit.Create(); err != nil {
|
|
log.Error(err)
|
|
}
|
|
log.WithFields(utils.ToFields(hit, s)).Infoln(
|
|
fmt.Sprintf("Service #%d '%v' Successful Response: %s | Lookup in: %s | Online: %v | Interval: %d seconds", s.Id, s.Name, humanMicro(hit.Latency), humanMicro(hit.PingTime), s.Online, s.Interval))
|
|
s.LastLookupTime = hit.PingTime
|
|
s.LastLatency = hit.Latency
|
|
metrics.Gauge("online", 1., s.Name, s.Type)
|
|
metrics.Inc("success", s.Name)
|
|
sendSuccess(s)
|
|
}
|
|
|
|
func RecordFailure(s *Service, issue, reason string) {
|
|
RecordFailureWithType(s, issue, reason, "")
|
|
}
|
|
|
|
// RecordFailure will create a new 'Failure' record in the database for a offline service
|
|
func RecordFailureWithType(s *Service, issue, reason string, failureType string) {
|
|
s.LastOffline = utils.Now()
|
|
|
|
fail := &failures.Failure{
|
|
Service: s.Id,
|
|
//Issue: issue,
|
|
PingTime: s.PingTime,
|
|
CreatedAt: utils.Now(),
|
|
ErrorCode: s.LastStatusCode,
|
|
Reason: reason,
|
|
Type: failureType,
|
|
}
|
|
log.WithFields(utils.ToFields(fail, s)).
|
|
Warnln(fmt.Sprintf("Service %v Failing: %v | Lookup in: %v", s.Name, issue, humanMicro(fail.PingTime)))
|
|
|
|
if err := fail.Create(); err != nil {
|
|
log.Error(err)
|
|
}
|
|
//s.Online = false
|
|
s.DownText = s.DowntimeText()
|
|
|
|
limitOffset := len(s.Failures)
|
|
if len(s.Failures) >= limitFailures {
|
|
limitOffset = limitFailures - 1
|
|
}
|
|
|
|
s.Failures = append([]*failures.Failure{fail}, s.Failures[:limitOffset]...)
|
|
|
|
metrics.Gauge("online", 0., s.Name, s.Type)
|
|
metrics.Inc("failure", s.Name)
|
|
sendFailure(s, fail)
|
|
}
|
|
|
|
// Check will run checkHttp for HTTP services and checkTcp for TCP services
|
|
// if record param is set to true, it will add a record into the database.
|
|
func (s *Service) CheckService(record bool) (err error) {
|
|
switch s.Type {
|
|
case "http":
|
|
_, err = CheckHttp(s, record)
|
|
case "tcp", "udp":
|
|
_, err = CheckTcp(s, record)
|
|
case "grpc":
|
|
_, err = CheckGrpc(s, record)
|
|
case "icmp":
|
|
_, err = CheckIcmp(s, record)
|
|
case "collection":
|
|
_, err = CheckCollection(s, record)
|
|
}
|
|
log.Infof("Health Check Executed : %s %s %s %s %s", s.Id, s.Name, s.Type, s.Online, err)
|
|
return
|
|
}
|
|
|
|
func (s *Service) HandleDowntime(err error, record bool) {
|
|
if err != nil {
|
|
s.FailureCounter++
|
|
if s.FailureCounter >= s.GetFtc() || s.Type == "collection" {
|
|
|
|
s.Online = false
|
|
|
|
t := time.Now().Add(time.Duration(-s.FailureCounter*s.Interval) * (time.Second))
|
|
|
|
downtime := &downtimes.Downtime{
|
|
Start: &t,
|
|
ServiceId: s.Id,
|
|
}
|
|
|
|
if s.CurrentDowntime > 0 {
|
|
if downtime, err = downtimes.Find(s.CurrentDowntime); err != nil {
|
|
log.Errorf("[Failure]Failed to find downtime : %s %s", s.Id, s.CurrentDowntime)
|
|
s.LastFailureType = ""
|
|
s.CurrentDowntime = 0
|
|
s.FailureCounter = 0
|
|
return
|
|
}
|
|
}
|
|
|
|
now := time.Now()
|
|
|
|
downtime.End = &now
|
|
newStatus := HandleEmptyStatus(s.LastFailureType)
|
|
|
|
start := time.Now().Add(time.Duration(-s.Interval) * (time.Second))
|
|
|
|
if downtime.SubStatus != "" && downtime.SubStatus != newStatus {
|
|
downtime.Id = 0
|
|
downtime.Start = &start
|
|
}
|
|
|
|
downtime.SubStatus = newStatus
|
|
|
|
downtime.Failures = s.FailureCounter
|
|
|
|
if downtime.Id > 0 {
|
|
if e := downtime.Update(); e != nil {
|
|
log.Errorf("Failed to update downtime : %s %s %s", s.Id, s.Name, e)
|
|
}
|
|
} else {
|
|
if e := downtime.Create(); e != nil {
|
|
log.Errorf("Failed to create downtime : %s %s %s", s.Id, s.Name, e)
|
|
}
|
|
}
|
|
s.CurrentDowntime = downtime.Id
|
|
}
|
|
} else {
|
|
s.Online = true
|
|
/*if s.CurrentDowntime > 0 {
|
|
if downtime, err := downtimes.Find(s.CurrentDowntime); err != nil {
|
|
log.Errorf("[Success]Failed to find downtime : %s %s", s.Id, s.CurrentDowntime)
|
|
} else {
|
|
downtime.End = time.Now()
|
|
downtime.SubStatus = ApplyStatus(downtime.SubStatus, HandleEmptyStatus(s.LastFailureType), STATUS_DEGRADED)
|
|
downtime.Failures = s.FailureCounter
|
|
if e := downtime.Update(); e != nil {
|
|
log.Errorf("Failed to close downtime : %s %s %s", s.Id, s.Name, e)
|
|
}
|
|
}
|
|
}*/
|
|
s.LastFailureType = ""
|
|
s.FailureCounter = 0
|
|
s.CurrentDowntime = 0
|
|
}
|
|
|
|
return
|
|
}
|