mirror of https://github.com/hashicorp/consul
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
732 lines
26 KiB
732 lines
26 KiB
// Copyright (c) HashiCorp, Inc. |
|
// SPDX-License-Identifier: MPL-2.0 |
|
|
|
package proxycfg |
|
|
|
import ( |
|
"context" |
|
"crypto/sha1" |
|
"encoding/base64" |
|
"fmt" |
|
"path" |
|
"strings" |
|
|
|
"github.com/mitchellh/mapstructure" |
|
|
|
"github.com/hashicorp/consul/acl" |
|
cachetype "github.com/hashicorp/consul/agent/cache-types" |
|
"github.com/hashicorp/consul/agent/leafcert" |
|
"github.com/hashicorp/consul/agent/proxycfg/internal/watch" |
|
"github.com/hashicorp/consul/agent/structs" |
|
"github.com/hashicorp/consul/api" |
|
"github.com/hashicorp/consul/proto/private/pbpeering" |
|
) |
|
|
|
type handlerConnectProxy struct { |
|
handlerState |
|
} |
|
|
|
// initialize sets up the watches needed based on current proxy registration |
|
// state. |
|
func (s *handlerConnectProxy) initialize(ctx context.Context) (ConfigSnapshot, error) { |
|
snap := newConfigSnapshotFromServiceInstance(s.serviceInstance, s.stateConfig) |
|
snap.ConnectProxy.DiscoveryChain = make(map[UpstreamID]*structs.CompiledDiscoveryChain) |
|
snap.ConnectProxy.WatchedDiscoveryChains = make(map[UpstreamID]context.CancelFunc) |
|
snap.ConnectProxy.WatchedUpstreams = make(map[UpstreamID]map[string]context.CancelFunc) |
|
snap.ConnectProxy.WatchedUpstreamEndpoints = make(map[UpstreamID]map[string]structs.CheckServiceNodes) |
|
snap.ConnectProxy.UpstreamPeerTrustBundles = watch.NewMap[string, *pbpeering.PeeringTrustBundle]() |
|
snap.ConnectProxy.WatchedGateways = make(map[UpstreamID]map[string]context.CancelFunc) |
|
snap.ConnectProxy.WatchedGatewayEndpoints = make(map[UpstreamID]map[string]structs.CheckServiceNodes) |
|
snap.ConnectProxy.WatchedLocalGWEndpoints = watch.NewMap[string, structs.CheckServiceNodes]() |
|
snap.ConnectProxy.WatchedServiceChecks = make(map[structs.ServiceID][]structs.CheckType) |
|
snap.ConnectProxy.PreparedQueryEndpoints = make(map[UpstreamID]structs.CheckServiceNodes) |
|
snap.ConnectProxy.DestinationsUpstream = watch.NewMap[UpstreamID, *structs.ServiceConfigEntry]() |
|
snap.ConnectProxy.UpstreamConfig = make(map[UpstreamID]*structs.Upstream) |
|
snap.ConnectProxy.PassthroughUpstreams = make(map[UpstreamID]map[string]map[string]struct{}) |
|
snap.ConnectProxy.PassthroughIndices = make(map[string]indexedTarget) |
|
snap.ConnectProxy.PeerUpstreamEndpoints = watch.NewMap[UpstreamID, structs.CheckServiceNodes]() |
|
snap.ConnectProxy.DestinationGateways = watch.NewMap[UpstreamID, structs.CheckServiceNodes]() |
|
snap.ConnectProxy.PeerUpstreamEndpointsUseHostnames = make(map[UpstreamID]struct{}) |
|
|
|
// Watch for root changes |
|
err := s.dataSources.CARoots.Notify(ctx, &structs.DCSpecificRequest{ |
|
Datacenter: s.source.Datacenter, |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
Source: *s.source, |
|
}, rootsWatchID, s.ch) |
|
if err != nil { |
|
return snap, err |
|
} |
|
|
|
err = s.dataSources.TrustBundleList.Notify(ctx, &cachetype.TrustBundleListRequest{ |
|
Request: &pbpeering.TrustBundleListByServiceRequest{ |
|
ServiceName: s.proxyCfg.DestinationServiceName, |
|
Namespace: s.proxyID.NamespaceOrDefault(), |
|
Partition: s.proxyID.PartitionOrDefault(), |
|
}, |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
}, peeringTrustBundlesWatchID, s.ch) |
|
if err != nil { |
|
return snap, err |
|
} |
|
|
|
// Watch the leaf cert |
|
err = s.dataSources.LeafCertificate.Notify(ctx, &leafcert.ConnectCALeafRequest{ |
|
Datacenter: s.source.Datacenter, |
|
Token: s.token, |
|
Service: s.proxyCfg.DestinationServiceName, |
|
EnterpriseMeta: s.proxyID.EnterpriseMeta, |
|
}, leafWatchID, s.ch) |
|
if err != nil { |
|
return snap, err |
|
} |
|
|
|
// Watch for intention updates |
|
err = s.dataSources.Intentions.Notify(ctx, &structs.ServiceSpecificRequest{ |
|
Datacenter: s.source.Datacenter, |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
EnterpriseMeta: s.proxyID.EnterpriseMeta, |
|
ServiceName: s.proxyCfg.DestinationServiceName, |
|
}, intentionsWatchID, s.ch) |
|
if err != nil { |
|
return snap, err |
|
} |
|
|
|
// Watch for JWT provider updates. |
|
// While we could optimize by only watching providers referenced by intentions, |
|
// this should be okay because we expect few JWT providers and infrequent JWT |
|
// provider updates. |
|
err = s.dataSources.ConfigEntryList.Notify(ctx, &structs.ConfigEntryQuery{ |
|
Kind: structs.JWTProvider, |
|
Datacenter: s.source.Datacenter, |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
EnterpriseMeta: *structs.DefaultEnterpriseMetaInPartition(s.proxyID.PartitionOrDefault()), |
|
}, jwtProviderID, s.ch) |
|
if err != nil { |
|
return snap, err |
|
} |
|
|
|
// Get information about the entire service mesh. |
|
err = s.dataSources.ConfigEntry.Notify(ctx, &structs.ConfigEntryQuery{ |
|
Kind: structs.MeshConfig, |
|
Name: structs.MeshConfigMesh, |
|
Datacenter: s.source.Datacenter, |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
EnterpriseMeta: *structs.DefaultEnterpriseMetaInPartition(s.proxyID.PartitionOrDefault()), |
|
}, meshConfigEntryID, s.ch) |
|
if err != nil { |
|
return snap, err |
|
} |
|
|
|
// Watch for service check updates |
|
err = s.dataSources.HTTPChecks.Notify(ctx, &cachetype.ServiceHTTPChecksRequest{ |
|
ServiceID: s.proxyCfg.DestinationServiceID, |
|
NodeName: s.source.Node, |
|
EnterpriseMeta: s.proxyID.EnterpriseMeta, |
|
}, svcChecksWatchIDPrefix+structs.ServiceIDString(s.proxyCfg.DestinationServiceID, &s.proxyID.EnterpriseMeta), s.ch) |
|
if err != nil { |
|
return snap, err |
|
} |
|
|
|
if err := s.maybeInitializeTelemetryCollectorWatches(ctx, snap); err != nil { |
|
return snap, fmt.Errorf("failed to initialize telemetry collector watches: %w", err) |
|
} |
|
|
|
if s.proxyCfg.Mode == structs.ProxyModeTransparent { |
|
// When in transparent proxy we will infer upstreams from intentions with this source |
|
err := s.dataSources.IntentionUpstreams.Notify(ctx, &structs.ServiceSpecificRequest{ |
|
Datacenter: s.source.Datacenter, |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
ServiceName: s.proxyCfg.DestinationServiceName, |
|
EnterpriseMeta: s.proxyID.EnterpriseMeta, |
|
}, intentionUpstreamsID, s.ch) |
|
if err != nil { |
|
return snap, err |
|
} |
|
err = s.dataSources.PeeredUpstreams.Notify(ctx, &structs.PartitionSpecificRequest{ |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
Datacenter: s.source.Datacenter, |
|
EnterpriseMeta: s.proxyID.EnterpriseMeta, |
|
}, peeredUpstreamsID, s.ch) |
|
if err != nil { |
|
return snap, err |
|
} |
|
// We also infer upstreams from destinations (egress points) |
|
err = s.dataSources.IntentionUpstreamsDestination.Notify(ctx, &structs.ServiceSpecificRequest{ |
|
Datacenter: s.source.Datacenter, |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
ServiceName: s.proxyCfg.DestinationServiceName, |
|
EnterpriseMeta: s.proxyID.EnterpriseMeta, |
|
}, intentionUpstreamsDestinationID, s.ch) |
|
if err != nil { |
|
return snap, err |
|
} |
|
} |
|
|
|
// Watch for updates to service endpoints for all upstreams |
|
for i := range s.proxyCfg.Upstreams { |
|
u := s.proxyCfg.Upstreams[i] |
|
|
|
uid := NewUpstreamID(&u) |
|
|
|
// Store defaults keyed under wildcard so they can be applied to centrally configured upstreams |
|
if u.DestinationName == structs.WildcardSpecifier { |
|
snap.ConnectProxy.UpstreamConfig[uid] = &u |
|
continue |
|
} |
|
|
|
snap.ConnectProxy.UpstreamConfig[uid] = &u |
|
// This can be true if the upstream is a synthetic entry populated from centralized upstream config. |
|
// Watches should not be created for them. |
|
if u.CentrallyConfigured { |
|
continue |
|
} |
|
|
|
dc := s.source.Datacenter |
|
if u.Datacenter != "" { |
|
dc = u.Datacenter |
|
} |
|
|
|
// Default the partition and namespace to the namespace of this proxy service. |
|
partition := s.proxyID.PartitionOrDefault() |
|
if u.DestinationPartition != "" { |
|
partition = u.DestinationPartition |
|
} |
|
ns := s.proxyID.NamespaceOrDefault() |
|
if u.DestinationNamespace != "" { |
|
ns = u.DestinationNamespace |
|
} |
|
|
|
cfg, err := parseReducedUpstreamConfig(u.Config) |
|
if err != nil { |
|
// Don't hard fail on a config typo, just warn. We'll fall back on |
|
// the plain discovery chain if there is an error so it's safe to |
|
// continue. |
|
s.logger.Warn("failed to parse upstream config", |
|
"upstream", uid.String(), |
|
"error", err, |
|
) |
|
} |
|
|
|
switch u.DestinationType { |
|
case structs.UpstreamDestTypePreparedQuery: |
|
err := s.dataSources.PreparedQuery.Notify(ctx, &structs.PreparedQueryExecuteRequest{ |
|
Datacenter: dc, |
|
QueryOptions: structs.QueryOptions{Token: s.token, MaxAge: defaultPreparedQueryPollInterval}, |
|
QueryIDOrName: u.DestinationName, |
|
Connect: true, |
|
Source: *s.source, |
|
}, "upstream:"+uid.String(), s.ch) |
|
if err != nil { |
|
return snap, err |
|
} |
|
|
|
case structs.UpstreamDestTypeService: |
|
fallthrough |
|
|
|
case "": |
|
if u.DestinationPeer != "" { |
|
err := s.setupWatchesForPeeredUpstream(ctx, snap.ConnectProxy, NewUpstreamID(&u), dc) |
|
if err != nil { |
|
return snap, fmt.Errorf("failed to setup watches for peered upstream %q: %w", uid.String(), err) |
|
} |
|
continue |
|
} |
|
|
|
err := s.dataSources.CompiledDiscoveryChain.Notify(ctx, &structs.DiscoveryChainRequest{ |
|
Datacenter: s.source.Datacenter, |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
Name: u.DestinationName, |
|
EvaluateInDatacenter: dc, |
|
EvaluateInNamespace: ns, |
|
EvaluateInPartition: partition, |
|
OverrideMeshGateway: u.MeshGateway, |
|
OverrideProtocol: cfg.Protocol, |
|
OverrideConnectTimeout: cfg.ConnectTimeout(), |
|
}, "discovery-chain:"+uid.String(), s.ch) |
|
if err != nil { |
|
return snap, fmt.Errorf("failed to watch discovery chain for %s: %v", uid.String(), err) |
|
} |
|
|
|
default: |
|
return snap, fmt.Errorf("unknown upstream type: %q", u.DestinationType) |
|
} |
|
} |
|
|
|
return snap, nil |
|
} |
|
|
|
func (s *handlerConnectProxy) setupWatchesForPeeredUpstream( |
|
ctx context.Context, |
|
snapConnectProxy configSnapshotConnectProxy, |
|
uid UpstreamID, |
|
dc string, |
|
) error { |
|
s.logger.Trace("initializing watch of peered upstream", "upstream", uid) |
|
|
|
// NOTE: An upstream that points to a peer by definition will |
|
// only ever watch a single catalog query, so a map key of just |
|
// "UID" is sufficient to cover the peer data watches here. |
|
err := s.dataSources.Health.Notify(ctx, &structs.ServiceSpecificRequest{ |
|
PeerName: uid.Peer, |
|
Datacenter: dc, |
|
QueryOptions: structs.QueryOptions{ |
|
Token: s.token, |
|
}, |
|
ServiceName: uid.Name, |
|
Connect: true, |
|
Source: *s.source, |
|
EnterpriseMeta: uid.EnterpriseMeta, |
|
}, upstreamPeerWatchIDPrefix+uid.String(), s.ch) |
|
if err != nil { |
|
return fmt.Errorf("failed to watch health for %s: %v", uid, err) |
|
} |
|
snapConnectProxy.PeerUpstreamEndpoints.InitWatch(uid, nil) |
|
|
|
// Check whether a watch for this peer exists to avoid duplicates. |
|
if ok := snapConnectProxy.UpstreamPeerTrustBundles.IsWatched(uid.Peer); !ok { |
|
peerCtx, cancel := context.WithCancel(ctx) |
|
if err := s.dataSources.TrustBundle.Notify(peerCtx, &cachetype.TrustBundleReadRequest{ |
|
Request: &pbpeering.TrustBundleReadRequest{ |
|
Name: uid.Peer, |
|
Partition: uid.PartitionOrDefault(), |
|
}, |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
}, peerTrustBundleIDPrefix+uid.Peer, s.ch); err != nil { |
|
cancel() |
|
return fmt.Errorf("error while watching trust bundle for peer %q: %w", uid.Peer, err) |
|
} |
|
|
|
snapConnectProxy.UpstreamPeerTrustBundles.InitWatch(uid.Peer, cancel) |
|
} |
|
|
|
// Always watch local GW endpoints for peer upstreams so that we don't have to worry about |
|
// the timing on whether the wildcard upstream config was fetched yet. |
|
up := &handlerUpstreams{handlerState: s.handlerState} |
|
up.setupWatchForLocalGWEndpoints(ctx, &snapConnectProxy.ConfigSnapshotUpstreams) |
|
return nil |
|
} |
|
|
|
func (s *handlerConnectProxy) handleUpdate(ctx context.Context, u UpdateEvent, snap *ConfigSnapshot) error { |
|
if u.Err != nil { |
|
return fmt.Errorf("error filling agent cache: %v", u.Err) |
|
} |
|
|
|
switch { |
|
case u.CorrelationID == rootsWatchID: |
|
roots, ok := u.Result.(*structs.IndexedCARoots) |
|
if !ok { |
|
return fmt.Errorf("invalid type for response: %T", u.Result) |
|
} |
|
snap.Roots = roots |
|
|
|
case u.CorrelationID == peeringTrustBundlesWatchID: |
|
resp, ok := u.Result.(*pbpeering.TrustBundleListByServiceResponse) |
|
if !ok { |
|
return fmt.Errorf("invalid type for response: %T", u.Result) |
|
} |
|
if len(resp.Bundles) > 0 { |
|
snap.ConnectProxy.InboundPeerTrustBundles = resp.Bundles |
|
} |
|
snap.ConnectProxy.InboundPeerTrustBundlesSet = true |
|
|
|
case u.CorrelationID == intentionsWatchID: |
|
resp, ok := u.Result.(structs.SimplifiedIntentions) |
|
if !ok { |
|
return fmt.Errorf("invalid type for response: %T", u.Result) |
|
} |
|
snap.ConnectProxy.Intentions = resp |
|
snap.ConnectProxy.IntentionsSet = true |
|
|
|
case u.CorrelationID == jwtProviderID: |
|
resp, ok := u.Result.(*structs.IndexedConfigEntries) |
|
|
|
if !ok { |
|
return fmt.Errorf("invalid type for response: %T", u.Result) |
|
} |
|
|
|
providers := make(map[string]*structs.JWTProviderConfigEntry, len(resp.Entries)) |
|
for _, entry := range resp.Entries { |
|
jwtEntry, ok := entry.(*structs.JWTProviderConfigEntry) |
|
if !ok { |
|
return fmt.Errorf("invalid type for response: %T", entry) |
|
} |
|
providers[jwtEntry.Name] = jwtEntry |
|
} |
|
|
|
snap.JWTProviders = providers |
|
case u.CorrelationID == peeredUpstreamsID: |
|
resp, ok := u.Result.(*structs.IndexedPeeredServiceList) |
|
if !ok { |
|
return fmt.Errorf("invalid type for response %T", u.Result) |
|
} |
|
|
|
seenUpstreams := make(map[UpstreamID]struct{}) |
|
for _, psn := range resp.Services { |
|
uid := NewUpstreamIDFromPeeredServiceName(psn) |
|
|
|
if _, ok := seenUpstreams[uid]; ok { |
|
continue |
|
} |
|
seenUpstreams[uid] = struct{}{} |
|
|
|
err := s.setupWatchesForPeeredUpstream(ctx, snap.ConnectProxy, uid, s.source.Datacenter) |
|
if err != nil { |
|
return fmt.Errorf("failed to setup watches for peered upstream %q: %w", uid.String(), err) |
|
} |
|
} |
|
snap.ConnectProxy.PeeredUpstreams = seenUpstreams |
|
|
|
// |
|
// Clean up data |
|
// |
|
|
|
peeredChainTargets := make(map[UpstreamID]struct{}) |
|
for _, discoChain := range snap.ConnectProxy.DiscoveryChain { |
|
for _, target := range discoChain.Targets { |
|
if target.Peer == "" { |
|
continue |
|
} |
|
uid := NewUpstreamIDFromTargetID(target.ID) |
|
peeredChainTargets[uid] = struct{}{} |
|
} |
|
} |
|
|
|
validPeerNames := make(map[string]struct{}) |
|
|
|
// Iterate through all known endpoints and remove references to upstream IDs that weren't in the update |
|
snap.ConnectProxy.PeerUpstreamEndpoints.ForEachKey(func(uid UpstreamID) bool { |
|
// Peered upstream is explicitly defined in upstream config |
|
if _, ok := snap.ConnectProxy.UpstreamConfig[uid]; ok { |
|
validPeerNames[uid.Peer] = struct{}{} |
|
return true |
|
} |
|
// Peered upstream came from dynamic source of imported services |
|
if _, ok := seenUpstreams[uid]; ok { |
|
validPeerNames[uid.Peer] = struct{}{} |
|
return true |
|
} |
|
// Peered upstream came from a discovery chain target |
|
if _, ok := peeredChainTargets[uid]; ok { |
|
validPeerNames[uid.Peer] = struct{}{} |
|
return true |
|
} |
|
snap.ConnectProxy.PeerUpstreamEndpoints.CancelWatch(uid) |
|
return true |
|
}) |
|
|
|
// Iterate through all known trust bundles and remove references to any unseen peer names |
|
snap.ConnectProxy.UpstreamPeerTrustBundles.ForEachKey(func(peerName PeerName) bool { |
|
if _, ok := validPeerNames[peerName]; !ok { |
|
snap.ConnectProxy.UpstreamPeerTrustBundles.CancelWatch(peerName) |
|
} |
|
return true |
|
}) |
|
|
|
case u.CorrelationID == intentionUpstreamsID: |
|
resp, ok := u.Result.(*structs.IndexedServiceList) |
|
if !ok { |
|
return fmt.Errorf("invalid type for response %T", u.Result) |
|
} |
|
|
|
seenUpstreams := make(map[UpstreamID]struct{}) |
|
for _, svc := range resp.Services { |
|
uid := NewUpstreamIDFromServiceName(svc) |
|
|
|
seenUpstreams[uid] = struct{}{} |
|
|
|
cfgMap := make(map[string]interface{}) |
|
u, ok := snap.ConnectProxy.UpstreamConfig[uid] |
|
if ok { |
|
cfgMap = u.Config |
|
} else { |
|
// Use the centralized upstream defaults if they exist and there isn't specific configuration for this upstream |
|
// This is only relevant to upstreams from intentions because for explicit upstreams the defaulting is handled |
|
// by the ResolveServiceConfig endpoint. |
|
wildcardUID := NewWildcardUID(&s.proxyID.EnterpriseMeta) |
|
defaults, ok := snap.ConnectProxy.UpstreamConfig[wildcardUID] |
|
if ok { |
|
u = defaults |
|
cfgMap = defaults.Config |
|
snap.ConnectProxy.UpstreamConfig[uid] = defaults |
|
} |
|
} |
|
|
|
cfg, err := parseReducedUpstreamConfig(cfgMap) |
|
if err != nil { |
|
// Don't hard fail on a config typo, just warn. We'll fall back on |
|
// the plain discovery chain if there is an error so it's safe to |
|
// continue. |
|
s.logger.Warn("failed to parse upstream config", |
|
"upstream", uid, |
|
"error", err, |
|
) |
|
} |
|
|
|
meshGateway := s.proxyCfg.MeshGateway |
|
if u != nil { |
|
meshGateway = u.MeshGateway |
|
} |
|
watchOpts := discoveryChainWatchOpts{ |
|
id: NewUpstreamIDFromServiceName(svc), |
|
name: svc.Name, |
|
namespace: svc.NamespaceOrDefault(), |
|
partition: svc.PartitionOrDefault(), |
|
datacenter: s.source.Datacenter, |
|
cfg: cfg, |
|
meshGateway: meshGateway, |
|
} |
|
up := &handlerUpstreams{handlerState: s.handlerState} |
|
err = up.watchDiscoveryChain(ctx, snap, watchOpts) |
|
if err != nil { |
|
return fmt.Errorf("failed to watch discovery chain for %s: %v", uid, err) |
|
} |
|
} |
|
snap.ConnectProxy.IntentionUpstreams = seenUpstreams |
|
|
|
// Clean up data from services that were not in the update |
|
for uid, targets := range snap.ConnectProxy.WatchedUpstreams { |
|
if upstream, ok := snap.ConnectProxy.UpstreamConfig[uid]; ok && !upstream.CentrallyConfigured { |
|
continue |
|
} |
|
if _, ok := seenUpstreams[uid]; !ok { |
|
for targetID, cancelFn := range targets { |
|
cancelFn() |
|
|
|
targetUID := NewUpstreamIDFromTargetID(targetID) |
|
if targetUID.Peer != "" { |
|
snap.ConnectProxy.PeerUpstreamEndpoints.CancelWatch(targetUID) |
|
snap.ConnectProxy.UpstreamPeerTrustBundles.CancelWatch(targetUID.Peer) |
|
} |
|
} |
|
delete(snap.ConnectProxy.WatchedUpstreams, uid) |
|
} |
|
} |
|
for uid := range snap.ConnectProxy.WatchedUpstreamEndpoints { |
|
if upstream, ok := snap.ConnectProxy.UpstreamConfig[uid]; ok && !upstream.CentrallyConfigured { |
|
continue |
|
} |
|
if _, ok := seenUpstreams[uid]; !ok { |
|
delete(snap.ConnectProxy.WatchedUpstreamEndpoints, uid) |
|
} |
|
} |
|
for uid, cancelMap := range snap.ConnectProxy.WatchedGateways { |
|
if upstream, ok := snap.ConnectProxy.UpstreamConfig[uid]; ok && !upstream.CentrallyConfigured { |
|
continue |
|
} |
|
if _, ok := seenUpstreams[uid]; !ok { |
|
for _, cancelFn := range cancelMap { |
|
cancelFn() |
|
} |
|
delete(snap.ConnectProxy.WatchedGateways, uid) |
|
} |
|
} |
|
for uid := range snap.ConnectProxy.WatchedGatewayEndpoints { |
|
if upstream, ok := snap.ConnectProxy.UpstreamConfig[uid]; ok && !upstream.CentrallyConfigured { |
|
continue |
|
} |
|
if _, ok := seenUpstreams[uid]; !ok { |
|
delete(snap.ConnectProxy.WatchedGatewayEndpoints, uid) |
|
} |
|
} |
|
for uid, cancelFn := range snap.ConnectProxy.WatchedDiscoveryChains { |
|
if upstream, ok := snap.ConnectProxy.UpstreamConfig[uid]; ok && !upstream.CentrallyConfigured { |
|
continue |
|
} |
|
if _, ok := seenUpstreams[uid]; !ok { |
|
cancelFn() |
|
delete(snap.ConnectProxy.WatchedDiscoveryChains, uid) |
|
} |
|
} |
|
for uid := range snap.ConnectProxy.PassthroughUpstreams { |
|
if _, ok := seenUpstreams[uid]; !ok { |
|
delete(snap.ConnectProxy.PassthroughUpstreams, uid) |
|
} |
|
} |
|
for addr, indexed := range snap.ConnectProxy.PassthroughIndices { |
|
if _, ok := seenUpstreams[indexed.upstreamID]; !ok { |
|
delete(snap.ConnectProxy.PassthroughIndices, addr) |
|
} |
|
} |
|
|
|
// These entries are intentionally handled separately from the WatchedDiscoveryChains above. |
|
// There have been situations where a discovery watch was cancelled, then fired. |
|
// That update event then re-populated the DiscoveryChain map entry, which wouldn't get cleaned up |
|
// since there was no known watch for it. |
|
for uid := range snap.ConnectProxy.DiscoveryChain { |
|
if upstream, ok := snap.ConnectProxy.UpstreamConfig[uid]; ok && !upstream.CentrallyConfigured { |
|
continue |
|
} |
|
if _, ok := seenUpstreams[uid]; !ok { |
|
delete(snap.ConnectProxy.DiscoveryChain, uid) |
|
} |
|
} |
|
case u.CorrelationID == intentionUpstreamsDestinationID: |
|
resp, ok := u.Result.(*structs.IndexedServiceList) |
|
if !ok { |
|
return fmt.Errorf("invalid type for response %T", u.Result) |
|
} |
|
seenUpstreams := make(map[UpstreamID]struct{}) |
|
for _, svc := range resp.Services { |
|
uid := NewUpstreamIDFromServiceName(svc) |
|
seenUpstreams[uid] = struct{}{} |
|
{ |
|
childCtx, cancel := context.WithCancel(ctx) |
|
err := s.dataSources.ConfigEntry.Notify(childCtx, &structs.ConfigEntryQuery{ |
|
Kind: structs.ServiceDefaults, |
|
Name: svc.Name, |
|
Datacenter: s.source.Datacenter, |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
EnterpriseMeta: svc.EnterpriseMeta, |
|
}, DestinationConfigEntryID+svc.String(), s.ch) |
|
if err != nil { |
|
cancel() |
|
return err |
|
} |
|
snap.ConnectProxy.DestinationsUpstream.InitWatch(uid, cancel) |
|
} |
|
{ |
|
childCtx, cancel := context.WithCancel(ctx) |
|
err := s.dataSources.ServiceGateways.Notify(childCtx, &structs.ServiceSpecificRequest{ |
|
ServiceName: svc.Name, |
|
Datacenter: s.source.Datacenter, |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
EnterpriseMeta: svc.EnterpriseMeta, |
|
ServiceKind: structs.ServiceKindTerminatingGateway, |
|
}, DestinationGatewayID+svc.String(), s.ch) |
|
if err != nil { |
|
cancel() |
|
return err |
|
} |
|
snap.ConnectProxy.DestinationGateways.InitWatch(uid, cancel) |
|
} |
|
} |
|
|
|
snap.ConnectProxy.DestinationsUpstream.ForEachKey(func(uid UpstreamID) bool { |
|
if _, ok := seenUpstreams[uid]; !ok { |
|
snap.ConnectProxy.DestinationsUpstream.CancelWatch(uid) |
|
} |
|
return true |
|
}) |
|
|
|
snap.ConnectProxy.DestinationGateways.ForEachKey(func(uid UpstreamID) bool { |
|
if _, ok := seenUpstreams[uid]; !ok { |
|
snap.ConnectProxy.DestinationGateways.CancelWatch(uid) |
|
} |
|
return true |
|
}) |
|
case strings.HasPrefix(u.CorrelationID, DestinationConfigEntryID): |
|
resp, ok := u.Result.(*structs.ConfigEntryResponse) |
|
if !ok { |
|
return fmt.Errorf("invalid type for response: %T", u.Result) |
|
} |
|
|
|
pq := strings.TrimPrefix(u.CorrelationID, DestinationConfigEntryID) |
|
uid := UpstreamIDFromString(pq) |
|
serviceConf, ok := resp.Entry.(*structs.ServiceConfigEntry) |
|
if !ok { |
|
return fmt.Errorf("invalid type for service default: %T", resp.Entry.GetName()) |
|
} |
|
|
|
snap.ConnectProxy.DestinationsUpstream.Set(uid, serviceConf) |
|
case strings.HasPrefix(u.CorrelationID, DestinationGatewayID): |
|
resp, ok := u.Result.(*structs.IndexedCheckServiceNodes) |
|
if !ok { |
|
return fmt.Errorf("invalid type for response: %T", u.Result) |
|
} |
|
|
|
pq := strings.TrimPrefix(u.CorrelationID, DestinationGatewayID) |
|
uid := UpstreamIDFromString(pq) |
|
snap.ConnectProxy.DestinationGateways.Set(uid, resp.Nodes) |
|
case strings.HasPrefix(u.CorrelationID, "upstream:"+preparedQueryIDPrefix): |
|
resp, ok := u.Result.(*structs.PreparedQueryExecuteResponse) |
|
if !ok { |
|
return fmt.Errorf("invalid type for response: %T", u.Result) |
|
} |
|
pq := strings.TrimPrefix(u.CorrelationID, "upstream:") |
|
uid := UpstreamIDFromString(pq) |
|
snap.ConnectProxy.PreparedQueryEndpoints[uid] = resp.Nodes |
|
|
|
case strings.HasPrefix(u.CorrelationID, svcChecksWatchIDPrefix): |
|
resp, ok := u.Result.([]structs.CheckType) |
|
if !ok { |
|
return fmt.Errorf("invalid type for service checks response: %T, want: []structs.CheckType", u.Result) |
|
} |
|
svcID := structs.ServiceIDFromString(strings.TrimPrefix(u.CorrelationID, svcChecksWatchIDPrefix)) |
|
snap.ConnectProxy.WatchedServiceChecks[svcID] = resp |
|
|
|
default: |
|
return (*handlerUpstreams)(s).handleUpdateUpstreams(ctx, u, snap) |
|
} |
|
return nil |
|
} |
|
|
|
// telemetryCollectorConfig represents the basic opaque config values for pushing telemetry to |
|
// a consul telemetry collector. |
|
type telemetryCollectorConfig struct { |
|
// TelemetryCollectorBindSocketDir is a string that configures the directory for a |
|
// unix socket where Envoy will forward metrics. These metrics get pushed to |
|
// the Consul Telemetry collector. |
|
TelemetryCollectorBindSocketDir string `mapstructure:"envoy_telemetry_collector_bind_socket_dir"` |
|
} |
|
|
|
func parseTelemetryCollectorConfig(m map[string]interface{}) (telemetryCollectorConfig, error) { |
|
var cfg telemetryCollectorConfig |
|
err := mapstructure.WeakDecode(m, &cfg) |
|
|
|
if err != nil { |
|
return cfg, fmt.Errorf("failed to decode: %w", err) |
|
} |
|
|
|
return cfg, nil |
|
} |
|
|
|
// maybeInitializeTelemetryCollectorWatches will initialize a synthetic upstream and discovery chain |
|
// watch for the consul telemetry collector, if telemetry data collection is enabled on the proxy registration. |
|
func (s *handlerConnectProxy) maybeInitializeTelemetryCollectorWatches(ctx context.Context, snap ConfigSnapshot) error { |
|
cfg, err := parseTelemetryCollectorConfig(s.proxyCfg.Config) |
|
if err != nil { |
|
s.logger.Error("failed to parse connect.proxy.config", "error", err) |
|
} |
|
|
|
if cfg.TelemetryCollectorBindSocketDir == "" { |
|
// telemetry collection is not enabled, return early. |
|
return nil |
|
} |
|
|
|
// The path includes the proxy ID so that when multiple proxies are on the same host |
|
// they each have a distinct path to send their telemetry data. |
|
id := s.proxyID.NamespaceOrDefault() + "_" + s.proxyID.ID |
|
|
|
// UNIX domain sockets paths have a max length of 108, so we take a hash of the compound ID |
|
// to limit the length of the socket path. |
|
h := sha1.New() |
|
h.Write([]byte(id)) |
|
hash := base64.RawURLEncoding.EncodeToString(h.Sum(nil)) |
|
path := path.Join(cfg.TelemetryCollectorBindSocketDir, hash+".sock") |
|
|
|
upstream := structs.Upstream{ |
|
DestinationNamespace: acl.DefaultNamespaceName, |
|
DestinationPartition: s.proxyID.PartitionOrDefault(), |
|
DestinationName: api.TelemetryCollectorName, |
|
LocalBindSocketPath: path, |
|
Config: map[string]interface{}{ |
|
"protocol": "grpc", |
|
}, |
|
} |
|
uid := NewUpstreamID(&upstream) |
|
snap.ConnectProxy.UpstreamConfig[uid] = &upstream |
|
|
|
err = s.dataSources.CompiledDiscoveryChain.Notify(ctx, &structs.DiscoveryChainRequest{ |
|
Datacenter: s.source.Datacenter, |
|
QueryOptions: structs.QueryOptions{Token: s.token}, |
|
Name: upstream.DestinationName, |
|
EvaluateInDatacenter: s.source.Datacenter, |
|
EvaluateInNamespace: uid.NamespaceOrDefault(), |
|
EvaluateInPartition: uid.PartitionOrDefault(), |
|
}, "discovery-chain:"+uid.String(), s.ch) |
|
if err != nil { |
|
return fmt.Errorf("failed to watch discovery chain for %s: %v", uid.String(), err) |
|
} |
|
return nil |
|
}
|
|
|