mirror of https://github.com/hashicorp/consul
413 lines
12 KiB
Go
413 lines
12 KiB
Go
// Copyright (c) HashiCorp, Inc.
|
|
// SPDX-License-Identifier: BUSL-1.1
|
|
|
|
package controller
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"time"
|
|
|
|
"golang.org/x/sync/errgroup"
|
|
|
|
"github.com/hashicorp/go-hclog"
|
|
|
|
"github.com/hashicorp/consul/agent/consul/controller/queue"
|
|
"github.com/hashicorp/consul/internal/controller/cache"
|
|
"github.com/hashicorp/consul/internal/protoutil"
|
|
"github.com/hashicorp/consul/internal/resource"
|
|
"github.com/hashicorp/consul/internal/storage"
|
|
"github.com/hashicorp/consul/proto-public/pbresource"
|
|
)
|
|
|
|
// Runtime contains the dependencies required by reconcilers.
|
|
type Runtime struct {
|
|
Client pbresource.ResourceServiceClient
|
|
Logger hclog.Logger
|
|
Cache cache.ReadOnlyCache
|
|
}
|
|
|
|
// controllerRunner contains the actual implementation of running a controller
|
|
// including creating watches, calling the reconciler, handling retries, etc.
|
|
type controllerRunner struct {
|
|
ctrl *Controller
|
|
// watchClient will be used by the controller infrastructure internally to
|
|
// perform watches and maintain the cache. On servers, this client will use
|
|
// the in-memory gRPC clients which DO NOT cause cloning of data returned by
|
|
// the resource service. This is desirable so that our cache doesn't incur
|
|
// the overhead of duplicating all resources that are watched. Generally
|
|
// dependency mappers and reconcilers should not be given this client so
|
|
// that they can be free to modify the data they are returned.
|
|
watchClient pbresource.ResourceServiceClient
|
|
// runtimeClient will be used by dependency mappers and reconcilers to
|
|
// access the resource service. On servers, this client will use the in-memory
|
|
// gRPC client wrapped with the cloning client to force cloning of protobuf
|
|
// messages as they pass through the client. This is desirable so that
|
|
// controllers and their mappers can be free to modify the data returned
|
|
// to them without having to think about the fact that the data should
|
|
// be immutable as it is shared with the controllers cache as well as the
|
|
// resource service itself.
|
|
runtimeClient pbresource.ResourceServiceClient
|
|
logger hclog.Logger
|
|
cache cache.Cache
|
|
}
|
|
|
|
func newControllerRunner(c *Controller, client pbresource.ResourceServiceClient, defaultLogger hclog.Logger) *controllerRunner {
|
|
return &controllerRunner{
|
|
ctrl: c,
|
|
watchClient: client,
|
|
runtimeClient: pbresource.NewCloningResourceServiceClient(client),
|
|
logger: c.buildLogger(defaultLogger),
|
|
// Do not build the cache here. If we build/set it when the controller runs
|
|
// then if a controller is restarted it will invalidate the previous cache automatically.
|
|
}
|
|
}
|
|
|
|
func (c *controllerRunner) run(ctx context.Context) error {
|
|
c.logger.Debug("controller running")
|
|
defer c.logger.Debug("controller stopping")
|
|
|
|
// Initialize the controller if required
|
|
if c.ctrl.initializer != nil {
|
|
c.logger.Debug("controller initializing")
|
|
err := c.ctrl.initializer.Initialize(ctx, c.runtime(c.logger))
|
|
if err != nil {
|
|
return err
|
|
}
|
|
c.logger.Debug("controller initialized")
|
|
}
|
|
|
|
c.cache = c.ctrl.buildCache()
|
|
defer func() {
|
|
// once no longer running we should nil out the cache
|
|
// so that we don't hold pointers to resources which may
|
|
// become out of date in the future.
|
|
c.cache = nil
|
|
}()
|
|
|
|
if c.ctrl.startCb != nil {
|
|
c.ctrl.startCb(ctx, c.runtime(c.logger))
|
|
}
|
|
|
|
if c.ctrl.stopCb != nil {
|
|
defer c.ctrl.stopCb(ctx, c.runtime(c.logger))
|
|
}
|
|
|
|
// Before we launch the reconciler or the dependency mappers, ensure the
|
|
// cache is properly primed to avoid errant reconciles.
|
|
//
|
|
// Without doing this the cache is unsafe for general use without causing
|
|
// reconcile regressions in certain cases.
|
|
{
|
|
c.logger.Debug("priming caches")
|
|
primeGroup, primeGroupCtx := errgroup.WithContext(ctx)
|
|
// Managed Type Events
|
|
primeGroup.Go(func() error {
|
|
return c.primeCache(primeGroupCtx, c.ctrl.managedTypeWatch.watchedType)
|
|
})
|
|
for _, w := range c.ctrl.watches {
|
|
watcher := w
|
|
// Watched Type Events
|
|
primeGroup.Go(func() error {
|
|
return c.primeCache(primeGroupCtx, watcher.watchedType)
|
|
})
|
|
}
|
|
|
|
if err := primeGroup.Wait(); err != nil {
|
|
return err
|
|
}
|
|
c.logger.Debug("priming caches complete")
|
|
}
|
|
|
|
group, groupCtx := errgroup.WithContext(ctx)
|
|
recQueue := runQueue[Request](groupCtx, c.ctrl)
|
|
|
|
// Managed Type Events → Reconciliation Queue
|
|
group.Go(func() error {
|
|
return c.watch(groupCtx, c.ctrl.managedTypeWatch.watchedType, func(res *pbresource.Resource) {
|
|
recQueue.Add(Request{ID: res.Id})
|
|
})
|
|
})
|
|
|
|
for _, w := range c.ctrl.watches {
|
|
mapQueue := runQueue[mapperRequest](groupCtx, c.ctrl)
|
|
watcher := w
|
|
|
|
// Watched Type Events → Mapper Queue
|
|
group.Go(func() error {
|
|
return c.watch(groupCtx, watcher.watchedType, func(res *pbresource.Resource) {
|
|
mapQueue.Add(mapperRequest{res: res})
|
|
})
|
|
})
|
|
|
|
// Mapper Queue → Mapper → Reconciliation Queue
|
|
group.Go(func() error {
|
|
return c.runMapper(groupCtx, watcher, mapQueue, recQueue, func(ctx context.Context, runtime Runtime, itemType queue.ItemType) ([]Request, error) {
|
|
return watcher.mapper(ctx, runtime, itemType.(mapperRequest).res)
|
|
})
|
|
})
|
|
}
|
|
|
|
for _, cw := range c.ctrl.customWatches {
|
|
customMapQueue := runQueue[Event](groupCtx, c.ctrl)
|
|
watcher := cw
|
|
// Custom Events → Mapper Queue
|
|
group.Go(func() error {
|
|
return watcher.source.Watch(groupCtx, func(e Event) {
|
|
customMapQueue.Add(e)
|
|
})
|
|
})
|
|
|
|
// Mapper Queue → Mapper → Reconciliation Queue
|
|
group.Go(func() error {
|
|
return c.runCustomMapper(groupCtx, watcher, customMapQueue, recQueue, func(ctx context.Context, runtime Runtime, itemType queue.ItemType) ([]Request, error) {
|
|
return watcher.mapper(ctx, runtime, itemType.(Event))
|
|
})
|
|
})
|
|
}
|
|
|
|
// Reconciliation Queue → Reconciler
|
|
group.Go(func() error {
|
|
return c.runReconciler(groupCtx, recQueue)
|
|
})
|
|
|
|
return group.Wait()
|
|
}
|
|
|
|
func runQueue[T queue.ItemType](ctx context.Context, ctrl *Controller) queue.WorkQueue[T] {
|
|
base, max := ctrl.backoff()
|
|
return queue.RunWorkQueue[T](ctx, base, max)
|
|
}
|
|
|
|
func (c *controllerRunner) primeCache(ctx context.Context, typ *pbresource.Type) error {
|
|
wl, err := c.watchClient.WatchList(ctx, &pbresource.WatchListRequest{
|
|
Type: typ,
|
|
Tenancy: &pbresource.Tenancy{
|
|
Partition: storage.Wildcard,
|
|
Namespace: storage.Wildcard,
|
|
},
|
|
})
|
|
if err != nil {
|
|
c.logger.Error("failed to create cache priming watch", "error", err)
|
|
return err
|
|
}
|
|
|
|
for {
|
|
event, err := wl.Recv()
|
|
if err != nil {
|
|
c.logger.Warn("error received from cache priming watch", "error", err)
|
|
return err
|
|
}
|
|
|
|
switch {
|
|
case event.GetUpsert() != nil:
|
|
c.cache.Insert(event.GetUpsert().Resource)
|
|
case event.GetDelete() != nil:
|
|
c.cache.Delete(event.GetDelete().Resource)
|
|
case event.GetEndOfSnapshot() != nil:
|
|
// This concludes the initial snapshot. The cache is primed.
|
|
return nil
|
|
default:
|
|
c.logger.Warn("skipping unexpected event type", "type", hclog.Fmt("%T", event.GetEvent()))
|
|
continue
|
|
}
|
|
}
|
|
}
|
|
|
|
func (c *controllerRunner) watch(ctx context.Context, typ *pbresource.Type, add func(*pbresource.Resource)) error {
|
|
wl, err := c.watchClient.WatchList(ctx, &pbresource.WatchListRequest{
|
|
Type: typ,
|
|
Tenancy: &pbresource.Tenancy{
|
|
Partition: storage.Wildcard,
|
|
Namespace: storage.Wildcard,
|
|
},
|
|
})
|
|
if err != nil {
|
|
c.logger.Error("failed to create watch", "error", err)
|
|
return err
|
|
}
|
|
|
|
for {
|
|
event, err := wl.Recv()
|
|
if err != nil {
|
|
c.logger.Warn("error received from watch", "error", err)
|
|
return err
|
|
}
|
|
|
|
// Keep the cache up to date. There main reason to do this here is
|
|
// to ensure that any mapper/reconciliation queue deduping wont
|
|
// hide events from being observed and updating the cache state.
|
|
// Therefore we should do this before any queueing.
|
|
var resource *pbresource.Resource
|
|
switch {
|
|
case event.GetUpsert() != nil:
|
|
resource = event.GetUpsert().GetResource()
|
|
c.cache.Insert(resource)
|
|
case event.GetDelete() != nil:
|
|
resource = event.GetDelete().GetResource()
|
|
c.cache.Delete(resource)
|
|
case event.GetEndOfSnapshot() != nil:
|
|
continue // ignore
|
|
default:
|
|
c.logger.Warn("skipping unexpected event type", "type", hclog.Fmt("%T", event.GetEvent()))
|
|
continue
|
|
}
|
|
|
|
// Before adding the resource into the queue we must clone it.
|
|
// While we want the cache to not have duplicate copies of all the
|
|
// data, we do want downstream consumers like dependency mappers and
|
|
// controllers to be able to freely modify the data they are given.
|
|
// Therefore we clone the resource here to prevent any accidental
|
|
// mutation of data held by the cache (and presumably by the resource
|
|
// service assuming that the regular client we were given is the inmem
|
|
// variant)
|
|
add(protoutil.Clone(resource))
|
|
}
|
|
}
|
|
|
|
func (c *controllerRunner) runMapper(
|
|
ctx context.Context,
|
|
w *watch,
|
|
from queue.WorkQueue[mapperRequest],
|
|
to queue.WorkQueue[Request],
|
|
mapper func(ctx context.Context, runtime Runtime, itemType queue.ItemType) ([]Request, error),
|
|
) error {
|
|
logger := c.logger.With("watched_resource_type", resource.ToGVK(w.watchedType))
|
|
|
|
for {
|
|
item, shutdown := from.Get()
|
|
if shutdown {
|
|
return nil
|
|
}
|
|
|
|
if err := c.doMap(ctx, mapper, to, item, logger); err != nil {
|
|
from.AddRateLimited(item)
|
|
from.Done(item)
|
|
continue
|
|
}
|
|
|
|
from.Forget(item)
|
|
from.Done(item)
|
|
}
|
|
}
|
|
|
|
func (c *controllerRunner) runCustomMapper(
|
|
ctx context.Context,
|
|
cw customWatch,
|
|
from queue.WorkQueue[Event],
|
|
to queue.WorkQueue[Request],
|
|
mapper func(ctx context.Context, runtime Runtime, itemType queue.ItemType) ([]Request, error),
|
|
) error {
|
|
logger := c.logger.With("watched_event", cw.source)
|
|
|
|
for {
|
|
item, shutdown := from.Get()
|
|
if shutdown {
|
|
return nil
|
|
}
|
|
|
|
if err := c.doMap(ctx, mapper, to, item, logger); err != nil {
|
|
from.AddRateLimited(item)
|
|
from.Done(item)
|
|
continue
|
|
}
|
|
|
|
from.Forget(item)
|
|
from.Done(item)
|
|
}
|
|
}
|
|
|
|
func (c *controllerRunner) doMap(ctx context.Context, mapper func(ctx context.Context, runtime Runtime, itemType queue.ItemType) ([]Request, error), to queue.WorkQueue[Request], item queue.ItemType, logger hclog.Logger) error {
|
|
var reqs []Request
|
|
if err := c.handlePanic(func() error {
|
|
var err error
|
|
reqs, err = mapper(ctx, c.runtime(logger.With("map-request-key", item.Key())), item)
|
|
return err
|
|
}); err != nil {
|
|
return err
|
|
}
|
|
|
|
for _, r := range reqs {
|
|
if !resource.EqualType(r.ID.Type, c.ctrl.managedTypeWatch.watchedType) {
|
|
logger.Error("dependency mapper returned request for a resource of the wrong type",
|
|
"type_expected", resource.ToGVK(c.ctrl.managedTypeWatch.watchedType),
|
|
"type_got", resource.ToGVK(r.ID.Type),
|
|
)
|
|
continue
|
|
}
|
|
to.Add(r)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (c *controllerRunner) runReconciler(ctx context.Context, queue queue.WorkQueue[Request]) error {
|
|
for {
|
|
req, shutdown := queue.Get()
|
|
if shutdown {
|
|
return nil
|
|
}
|
|
|
|
c.logger.Trace("handling request", "request", req)
|
|
err := c.handlePanic(func() error {
|
|
return c.ctrl.reconciler.Reconcile(ctx, c.runtime(c.logger.With("resource-id", req.ID.String())), req)
|
|
})
|
|
if err == nil {
|
|
queue.Forget(req)
|
|
} else {
|
|
var requeueAfter RequeueAfterError
|
|
if errors.As(err, &requeueAfter) {
|
|
queue.Forget(req)
|
|
queue.AddAfter(req, time.Duration(requeueAfter))
|
|
} else {
|
|
queue.AddRateLimited(req)
|
|
}
|
|
}
|
|
queue.Done(req)
|
|
}
|
|
}
|
|
|
|
func (c *controllerRunner) handlePanic(fn func() error) (err error) {
|
|
defer func() {
|
|
if r := recover(); r != nil {
|
|
stack := hclog.Stacktrace()
|
|
c.logger.Error("controller panic",
|
|
"panic", r,
|
|
"stack", stack,
|
|
)
|
|
err = fmt.Errorf("panic [recovered]: %v", r)
|
|
return
|
|
}
|
|
}()
|
|
|
|
return fn()
|
|
}
|
|
|
|
func (c *controllerRunner) runtime(logger hclog.Logger) Runtime {
|
|
return Runtime{
|
|
// dependency mappers and controllers are always given the cloning client
|
|
// so that they do not have to care about mutating values that they read
|
|
// through the client.
|
|
Client: c.runtimeClient,
|
|
Logger: logger,
|
|
// ensure that resources queried via the cache get cloned so that the
|
|
// dependency mapper or reconciler is free to modify them.
|
|
Cache: cache.NewCloningReadOnlyCache(c.cache),
|
|
}
|
|
}
|
|
|
|
type mapperRequest struct{ res *pbresource.Resource }
|
|
|
|
// Key satisfies the queue.ItemType interface. It returns a string which will be
|
|
// used to de-duplicate requests in the queue.
|
|
func (i mapperRequest) Key() string {
|
|
return fmt.Sprintf(
|
|
"type=%q,part=%q,ns=%q,name=%q,uid=%q",
|
|
resource.ToGVK(i.res.Id.Type),
|
|
i.res.Id.Tenancy.Partition,
|
|
i.res.Id.Tenancy.Namespace,
|
|
i.res.Id.Name,
|
|
i.res.Id.Uid,
|
|
)
|
|
}
|