mirror of https://github.com/hashicorp/consul
672 lines
20 KiB
Go
672 lines
20 KiB
Go
// Copyright (c) HashiCorp, Inc.
|
|
// SPDX-License-Identifier: BUSL-1.1
|
|
|
|
package proxystateconverter
|
|
|
|
import (
|
|
"errors"
|
|
"fmt"
|
|
|
|
"github.com/hashicorp/consul/agent/connect"
|
|
"github.com/hashicorp/consul/agent/proxycfg"
|
|
"github.com/hashicorp/consul/agent/structs"
|
|
"github.com/hashicorp/consul/agent/xds/response"
|
|
"github.com/hashicorp/consul/api"
|
|
"github.com/hashicorp/consul/proto-public/pbmesh/v2beta1/pbproxystate"
|
|
"github.com/hashicorp/go-bexpr"
|
|
|
|
"google.golang.org/protobuf/types/known/wrapperspb"
|
|
)
|
|
|
|
func makeLbEndpoint(addr string, port int, health pbproxystate.HealthStatus, weight int) *pbproxystate.Endpoint {
|
|
ep := &pbproxystate.Endpoint{
|
|
Address: &pbproxystate.Endpoint_HostPort{
|
|
HostPort: &pbproxystate.HostPortAddress{
|
|
Host: addr,
|
|
Port: uint32(port),
|
|
},
|
|
},
|
|
}
|
|
ep.HealthStatus = health
|
|
ep.LoadBalancingWeight = &wrapperspb.UInt32Value{Value: uint32(weight)}
|
|
return ep
|
|
}
|
|
|
|
// endpointsFromSnapshot returns the mesh API representation of the "routes" in the snapshot.
|
|
func (s *Converter) endpointsFromSnapshot(cfgSnap *proxycfg.ConfigSnapshot) error {
|
|
|
|
if cfgSnap == nil {
|
|
return errors.New("nil config given")
|
|
}
|
|
|
|
switch cfgSnap.Kind {
|
|
case structs.ServiceKindConnectProxy:
|
|
return s.endpointsFromSnapshotConnectProxy(cfgSnap)
|
|
//case structs.ServiceKindTerminatingGateway:
|
|
// return s.endpointsFromSnapshotTerminatingGateway(cfgSnap)
|
|
//case structs.ServiceKindMeshGateway:
|
|
// return s.endpointsFromSnapshotMeshGateway(cfgSnap)
|
|
//case structs.ServiceKindIngressGateway:
|
|
// return s.endpointsFromSnapshotIngressGateway(cfgSnap)
|
|
//case structs.ServiceKindAPIGateway:
|
|
// return s.endpointsFromSnapshotAPIGateway(cfgSnap)
|
|
default:
|
|
return fmt.Errorf("Invalid service kind: %v", cfgSnap.Kind)
|
|
}
|
|
}
|
|
|
|
// endpointsFromSnapshotConnectProxy returns the xDS API representation of the "endpoints"
|
|
// (upstream instances) in the snapshot.
|
|
func (s *Converter) endpointsFromSnapshotConnectProxy(cfgSnap *proxycfg.ConfigSnapshot) error {
|
|
eps := make(map[string]*pbproxystate.Endpoints)
|
|
|
|
// NOTE: Any time we skip a chain below we MUST also skip that discovery chain in clusters.go
|
|
// so that the sets of endpoints generated matches the sets of clusters.
|
|
for uid, chain := range cfgSnap.ConnectProxy.DiscoveryChain {
|
|
upstream, skip := cfgSnap.ConnectProxy.GetUpstream(uid, &cfgSnap.ProxyID.EnterpriseMeta)
|
|
if skip {
|
|
// Discovery chain is not associated with a known explicit or implicit upstream so it is skipped.
|
|
continue
|
|
}
|
|
|
|
var upstreamConfigMap map[string]interface{}
|
|
if upstream != nil {
|
|
upstreamConfigMap = upstream.Config
|
|
}
|
|
|
|
es, err := s.endpointsFromDiscoveryChain(
|
|
uid,
|
|
chain,
|
|
cfgSnap,
|
|
cfgSnap.Locality,
|
|
upstreamConfigMap,
|
|
cfgSnap.ConnectProxy.WatchedUpstreamEndpoints[uid],
|
|
cfgSnap.ConnectProxy.WatchedGatewayEndpoints[uid],
|
|
false,
|
|
)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
for clusterName, endpoints := range es {
|
|
eps[clusterName] = &pbproxystate.Endpoints{
|
|
Endpoints: endpoints,
|
|
}
|
|
|
|
}
|
|
}
|
|
|
|
// NOTE: Any time we skip an upstream below we MUST also skip that same
|
|
// upstream in clusters.go so that the sets of endpoints generated matches
|
|
// the sets of clusters.
|
|
for _, uid := range cfgSnap.ConnectProxy.PeeredUpstreamIDs() {
|
|
upstream, skip := cfgSnap.ConnectProxy.GetUpstream(uid, &cfgSnap.ProxyID.EnterpriseMeta)
|
|
if skip {
|
|
// Discovery chain is not associated with a known explicit or implicit upstream so it is skipped.
|
|
continue
|
|
}
|
|
|
|
tbs, ok := cfgSnap.ConnectProxy.UpstreamPeerTrustBundles.Get(uid.Peer)
|
|
if !ok {
|
|
// this should never happen since we loop through upstreams with
|
|
// set trust bundles
|
|
return fmt.Errorf("trust bundle not ready for peer %s", uid.Peer)
|
|
}
|
|
|
|
clusterName := generatePeeredClusterName(uid, tbs)
|
|
|
|
mgwMode := structs.MeshGatewayModeDefault
|
|
if upstream != nil {
|
|
mgwMode = upstream.MeshGateway.Mode
|
|
}
|
|
peerServiceEndpoints, err := s.makeEndpointsForPeerService(cfgSnap, uid, mgwMode)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
if peerServiceEndpoints != nil {
|
|
pbEndpoints := &pbproxystate.Endpoints{
|
|
Endpoints: peerServiceEndpoints,
|
|
}
|
|
|
|
eps[clusterName] = pbEndpoints
|
|
}
|
|
}
|
|
|
|
// Looping over explicit upstreams is only needed for prepared queries because they do not have discovery chains
|
|
for _, u := range cfgSnap.Proxy.Upstreams {
|
|
if u.DestinationType != structs.UpstreamDestTypePreparedQuery {
|
|
continue
|
|
}
|
|
uid := proxycfg.NewUpstreamID(&u)
|
|
|
|
dc := u.Datacenter
|
|
if dc == "" {
|
|
dc = cfgSnap.Datacenter
|
|
}
|
|
clusterName := connect.UpstreamSNI(&u, "", dc, cfgSnap.Roots.TrustDomain)
|
|
|
|
endpoints, ok := cfgSnap.ConnectProxy.PreparedQueryEndpoints[uid]
|
|
if ok {
|
|
epts := makeEndpointsForLoadAssignment(
|
|
cfgSnap,
|
|
nil,
|
|
[]loadAssignmentEndpointGroup{
|
|
{Endpoints: endpoints},
|
|
},
|
|
cfgSnap.Locality,
|
|
)
|
|
pbEndpoints := &pbproxystate.Endpoints{
|
|
Endpoints: epts,
|
|
}
|
|
|
|
eps[clusterName] = pbEndpoints
|
|
}
|
|
}
|
|
|
|
// Loop over potential destinations in the mesh, then grab the gateway nodes associated with each
|
|
cfgSnap.ConnectProxy.DestinationsUpstream.ForEachKey(func(uid proxycfg.UpstreamID) bool {
|
|
svcConfig, ok := cfgSnap.ConnectProxy.DestinationsUpstream.Get(uid)
|
|
if !ok || svcConfig.Destination == nil {
|
|
return true
|
|
}
|
|
|
|
for _, address := range svcConfig.Destination.Addresses {
|
|
clusterName := clusterNameForDestination(cfgSnap, uid.Name, address, uid.NamespaceOrDefault(), uid.PartitionOrDefault())
|
|
|
|
endpoints, ok := cfgSnap.ConnectProxy.DestinationGateways.Get(uid)
|
|
if ok {
|
|
epts := makeEndpointsForLoadAssignment(
|
|
cfgSnap,
|
|
nil,
|
|
[]loadAssignmentEndpointGroup{
|
|
{Endpoints: endpoints},
|
|
},
|
|
proxycfg.GatewayKey{ /*empty so it never matches*/ },
|
|
)
|
|
pbEndpoints := &pbproxystate.Endpoints{
|
|
Endpoints: epts,
|
|
}
|
|
eps[clusterName] = pbEndpoints
|
|
}
|
|
}
|
|
|
|
return true
|
|
})
|
|
|
|
s.proxyState.Endpoints = eps
|
|
return nil
|
|
}
|
|
|
|
func (s *Converter) makeEndpointsForPeerService(
|
|
cfgSnap *proxycfg.ConfigSnapshot,
|
|
uid proxycfg.UpstreamID,
|
|
upstreamGatewayMode structs.MeshGatewayMode,
|
|
) ([]*pbproxystate.Endpoint, error) {
|
|
var eps []*pbproxystate.Endpoint
|
|
|
|
upstreamsSnapshot, err := cfgSnap.ToConfigSnapshotUpstreams()
|
|
if err != nil {
|
|
return eps, err
|
|
}
|
|
|
|
if upstreamGatewayMode == structs.MeshGatewayModeNone {
|
|
s.Logger.Warn(fmt.Sprintf("invalid mesh gateway mode 'none', defaulting to 'remote' for %q", uid))
|
|
}
|
|
|
|
// If an upstream is configured with local mesh gw mode, we make a load assignment
|
|
// from the gateway endpoints instead of those of the upstreams.
|
|
if upstreamGatewayMode == structs.MeshGatewayModeLocal {
|
|
localGw, ok := cfgSnap.ConnectProxy.WatchedLocalGWEndpoints.Get(cfgSnap.Locality.String())
|
|
if !ok {
|
|
// local GW is not ready; return early
|
|
return eps, nil
|
|
}
|
|
eps = makeEndpointsForLoadAssignment(
|
|
cfgSnap,
|
|
nil,
|
|
[]loadAssignmentEndpointGroup{
|
|
{Endpoints: localGw},
|
|
},
|
|
cfgSnap.Locality,
|
|
)
|
|
return eps, nil
|
|
}
|
|
|
|
// Also skip peer instances with a hostname as their address. EDS
|
|
// cannot resolve hostnames, so we provide them through CDS instead.
|
|
if _, ok := upstreamsSnapshot.PeerUpstreamEndpointsUseHostnames[uid]; ok {
|
|
return eps, nil
|
|
}
|
|
|
|
endpoints, ok := upstreamsSnapshot.PeerUpstreamEndpoints.Get(uid)
|
|
if !ok {
|
|
return nil, nil
|
|
}
|
|
eps = makeEndpointsForLoadAssignment(
|
|
cfgSnap,
|
|
nil,
|
|
[]loadAssignmentEndpointGroup{
|
|
{Endpoints: endpoints},
|
|
},
|
|
proxycfg.GatewayKey{ /*empty so it never matches*/ },
|
|
)
|
|
return eps, nil
|
|
}
|
|
|
|
func (s *Converter) filterSubsetEndpoints(subset *structs.ServiceResolverSubset, endpoints structs.CheckServiceNodes) (structs.CheckServiceNodes, error) {
|
|
// locally execute the subsets filter
|
|
if subset.Filter != "" {
|
|
filter, err := bexpr.CreateFilter(subset.Filter, nil, endpoints)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
raw, err := filter.Execute(endpoints)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return raw.(structs.CheckServiceNodes), nil
|
|
}
|
|
return endpoints, nil
|
|
}
|
|
|
|
// TODO(proxystate): Terminating Gateway will be added in the future.
|
|
// Functions to add from agent/xds/endpoints.go:
|
|
// func endpointsFromSnapshotTerminatingGateway
|
|
|
|
// TODO(proxystate): Mesh Gateway will be added in the future.
|
|
// Functions to add from agent/xds/endpoints.go:
|
|
// func endpointsFromSnapshotMeshGateway
|
|
|
|
// TODO(proxystate): Cluster Peering will be added in the future.
|
|
// Functions to add from agent/xds/endpoints.go:
|
|
// func makeEndpointsForOutgoingPeeredServices
|
|
|
|
// TODO(proxystate): Mesh Gateway will be added in the future.
|
|
// Functions to add from agent/xds/endpoints.go:
|
|
// func endpointsFromServicesAndResolvers
|
|
|
|
// TODO(proxystate): Mesh Gateway will be added in the future.
|
|
// Functions to add from agent/xds/endpoints.go:
|
|
// func makePeerServerEndpointsForMeshGateway
|
|
|
|
// TODO(proxystate): Ingress Gateway will be added in the future.
|
|
// Functions to add from agent/xds/endpoints.go:
|
|
// func endpointsFromSnapshotIngressGateway
|
|
|
|
// TODO(proxystate): API Gateway will be added in the future.
|
|
// Functions to add from agent/xds/endpoints.go:
|
|
// func endpointsFromSnapshotAPIGateway
|
|
|
|
// used in clusters.go
|
|
func makeHostPortEndpoint(host string, port int) *pbproxystate.Endpoint {
|
|
return &pbproxystate.Endpoint{
|
|
Address: &pbproxystate.Endpoint_HostPort{
|
|
HostPort: &pbproxystate.HostPortAddress{
|
|
Host: host,
|
|
Port: uint32(port),
|
|
},
|
|
},
|
|
}
|
|
}
|
|
|
|
func makeUnixSocketEndpoint(path string) *pbproxystate.Endpoint {
|
|
return &pbproxystate.Endpoint{
|
|
Address: &pbproxystate.Endpoint_UnixSocket{
|
|
UnixSocket: &pbproxystate.UnixSocketAddress{
|
|
Path: path,
|
|
// envoy's mode is particular to a pipe address and is uint32.
|
|
// it also says "The mode for the Pipe. Not applicable for abstract sockets."
|
|
// https://www.envoyproxy.io/docs/envoy/latest/api-v3/config/core/v3/address.proto#config-core-v3-pipe
|
|
Mode: "0",
|
|
},
|
|
},
|
|
}
|
|
}
|
|
|
|
func (s *Converter) makeUpstreamLoadAssignmentEndpointForPeerService(
|
|
cfgSnap *proxycfg.ConfigSnapshot,
|
|
uid proxycfg.UpstreamID,
|
|
upstreamGatewayMode structs.MeshGatewayMode,
|
|
) ([]*pbproxystate.Endpoint, error) {
|
|
var eps []*pbproxystate.Endpoint
|
|
|
|
upstreamsSnapshot, err := cfgSnap.ToConfigSnapshotUpstreams()
|
|
if err != nil {
|
|
return eps, err
|
|
}
|
|
|
|
if upstreamGatewayMode == structs.MeshGatewayModeNone {
|
|
s.Logger.Warn(fmt.Sprintf("invalid mesh gateway mode 'none', defaulting to 'remote' for %q", uid))
|
|
}
|
|
|
|
// If an upstream is configured with local mesh gw mode, we make a load assignment
|
|
// from the gateway endpoints instead of those of the upstreams.
|
|
if upstreamGatewayMode == structs.MeshGatewayModeLocal {
|
|
localGw, ok := cfgSnap.ConnectProxy.WatchedLocalGWEndpoints.Get(cfgSnap.Locality.String())
|
|
if !ok {
|
|
// local GW is not ready; return early
|
|
return eps, nil
|
|
}
|
|
eps = makeEndpointsForLoadAssignment(
|
|
cfgSnap,
|
|
nil,
|
|
[]loadAssignmentEndpointGroup{
|
|
{Endpoints: localGw},
|
|
},
|
|
cfgSnap.Locality,
|
|
)
|
|
return eps, nil
|
|
}
|
|
|
|
// Also skip peer instances with a hostname as their address. EDS
|
|
// cannot resolve hostnames, so we provide them through CDS instead.
|
|
if _, ok := upstreamsSnapshot.PeerUpstreamEndpointsUseHostnames[uid]; ok {
|
|
return eps, nil
|
|
}
|
|
|
|
endpoints, ok := upstreamsSnapshot.PeerUpstreamEndpoints.Get(uid)
|
|
if !ok {
|
|
return nil, nil
|
|
}
|
|
eps = makeEndpointsForLoadAssignment(
|
|
cfgSnap,
|
|
nil,
|
|
[]loadAssignmentEndpointGroup{
|
|
{Endpoints: endpoints},
|
|
},
|
|
proxycfg.GatewayKey{ /*empty so it never matches*/ },
|
|
)
|
|
return eps, nil
|
|
}
|
|
|
|
func (s *Converter) endpointsFromDiscoveryChain(
|
|
uid proxycfg.UpstreamID,
|
|
chain *structs.CompiledDiscoveryChain,
|
|
cfgSnap *proxycfg.ConfigSnapshot,
|
|
gatewayKey proxycfg.GatewayKey,
|
|
upstreamConfigMap map[string]interface{},
|
|
upstreamEndpoints map[string]structs.CheckServiceNodes,
|
|
gatewayEndpoints map[string]structs.CheckServiceNodes,
|
|
forMeshGateway bool,
|
|
) (map[string][]*pbproxystate.Endpoint, error) {
|
|
if chain == nil {
|
|
if forMeshGateway {
|
|
return nil, fmt.Errorf("missing discovery chain for %s", uid)
|
|
}
|
|
return nil, nil
|
|
}
|
|
|
|
if upstreamConfigMap == nil {
|
|
upstreamConfigMap = make(map[string]interface{}) // TODO:needed?
|
|
}
|
|
|
|
clusterEndpoints := make(map[string][]*pbproxystate.Endpoint)
|
|
|
|
// TODO(proxystate): escape hatches will be implemented in the future
|
|
//var escapeHatchCluster *pbproxystate.Cluster
|
|
//if !forMeshGateway {
|
|
|
|
//cfg, err := structs.ParseUpstreamConfigNoDefaults(upstreamConfigMap)
|
|
//if err != nil {
|
|
// // Don't hard fail on a config typo, just warn. The parse func returns
|
|
// // default config if there is an error so it's safe to continue.
|
|
// s.Logger.Warn("failed to parse", "upstream", uid,
|
|
// "error", err)
|
|
//}
|
|
|
|
//if cfg.EnvoyClusterJSON != "" {
|
|
// if chain.Default {
|
|
// // If you haven't done anything to setup the discovery chain, then
|
|
// // you can use the envoy_cluster_json escape hatch.
|
|
// escapeHatchCluster, err = makeClusterFromUserConfig(cfg.EnvoyClusterJSON)
|
|
// if err != nil {
|
|
// return ce, nil
|
|
// }
|
|
// } else {
|
|
// s.Logger.Warn("ignoring escape hatch setting, because a discovery chain is configued for",
|
|
// "discovery chain", chain.ServiceName, "upstream", uid,
|
|
// "envoy_cluster_json", chain.ServiceName)
|
|
// }
|
|
//}
|
|
//}
|
|
|
|
mgwMode := structs.MeshGatewayModeDefault
|
|
if upstream, _ := cfgSnap.ConnectProxy.GetUpstream(uid, &cfgSnap.ProxyID.EnterpriseMeta); upstream != nil {
|
|
mgwMode = upstream.MeshGateway.Mode
|
|
}
|
|
|
|
// Find all resolver nodes.
|
|
for _, node := range chain.Nodes {
|
|
switch {
|
|
case node == nil:
|
|
return nil, fmt.Errorf("impossible to process a nil node")
|
|
case node.Type != structs.DiscoveryGraphNodeTypeResolver:
|
|
continue
|
|
case node.Resolver == nil:
|
|
return nil, fmt.Errorf("impossible to process a non-resolver node")
|
|
}
|
|
rawUpstreamConfig, err := structs.ParseUpstreamConfigNoDefaults(upstreamConfigMap)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
upstreamConfig := finalizeUpstreamConfig(rawUpstreamConfig, chain, node.Resolver.ConnectTimeout)
|
|
|
|
mappedTargets, err := s.mapDiscoChainTargets(cfgSnap, chain, node, upstreamConfig, forMeshGateway)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
targetGroups, err := mappedTargets.groupedTargets()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
for _, groupedTarget := range targetGroups {
|
|
clusterName := groupedTarget.ClusterName
|
|
// TODO(proxystate): escape hatches will be implemented in the future
|
|
//if escapeHatchCluster != nil {
|
|
// clusterName = escapeHatchCluster.Name
|
|
//}
|
|
switch len(groupedTarget.Targets) {
|
|
case 0:
|
|
continue
|
|
case 1:
|
|
// We expect one target so this passes through to continue setting the load assignment up.
|
|
default:
|
|
return nil, fmt.Errorf("cannot have more than one target")
|
|
}
|
|
ti := groupedTarget.Targets[0]
|
|
s.Logger.Debug("generating endpoints for", "cluster", clusterName, "targetID", ti.TargetID)
|
|
targetUID := proxycfg.NewUpstreamIDFromTargetID(ti.TargetID)
|
|
if targetUID.Peer != "" {
|
|
peerServiceEndpoints, err := s.makeEndpointsForPeerService(cfgSnap, targetUID, mgwMode)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if peerServiceEndpoints != nil {
|
|
clusterEndpoints[clusterName] = peerServiceEndpoints
|
|
}
|
|
continue
|
|
}
|
|
|
|
endpointGroup, valid := makeLoadAssignmentEndpointGroup(
|
|
chain.Targets,
|
|
upstreamEndpoints,
|
|
gatewayEndpoints,
|
|
ti.TargetID,
|
|
gatewayKey,
|
|
forMeshGateway,
|
|
)
|
|
if !valid {
|
|
continue // skip the cluster if we're still populating the snapshot
|
|
}
|
|
|
|
epts := makeEndpointsForLoadAssignment(
|
|
cfgSnap,
|
|
ti.PrioritizeByLocality,
|
|
[]loadAssignmentEndpointGroup{endpointGroup},
|
|
gatewayKey,
|
|
)
|
|
clusterEndpoints[clusterName] = epts
|
|
}
|
|
}
|
|
|
|
return clusterEndpoints, nil
|
|
}
|
|
|
|
// TODO(proxystate): Mesh Gateway will be added in the future.
|
|
// Functions to add from agent/xds/endpoints.go:
|
|
// func makeExportedUpstreamEndpointsForMeshGateway
|
|
|
|
type loadAssignmentEndpointGroup struct {
|
|
Endpoints structs.CheckServiceNodes
|
|
OnlyPassing bool
|
|
OverrideHealth pbproxystate.HealthStatus
|
|
}
|
|
|
|
func makeEndpointsForLoadAssignment(cfgSnap *proxycfg.ConfigSnapshot,
|
|
policy *structs.DiscoveryPrioritizeByLocality,
|
|
endpointGroups []loadAssignmentEndpointGroup,
|
|
localKey proxycfg.GatewayKey) []*pbproxystate.Endpoint {
|
|
pbEndpoints := make([]*pbproxystate.Endpoint, 0, len(endpointGroups))
|
|
|
|
// TODO(proxystate): this will be added with property overrides having golden files with this
|
|
//if len(endpointGroups) > 1 {
|
|
// cla.Policy = &envoy_endpoint_v3.ClusterLoadAssignment_Policy{
|
|
// // We choose such a large value here that the failover math should
|
|
// // in effect not happen until zero instances are healthy.
|
|
// OverprovisioningFactor: response.MakeUint32Value(100000),
|
|
// }
|
|
//}
|
|
|
|
var priority uint32
|
|
|
|
for _, endpointGroup := range endpointGroups {
|
|
endpointsByLocality, err := groupedEndpoints(cfgSnap.ServiceLocality, policy, endpointGroup.Endpoints)
|
|
|
|
if err != nil {
|
|
continue
|
|
}
|
|
|
|
for _, endpoints := range endpointsByLocality {
|
|
for _, ep := range endpoints {
|
|
// TODO (mesh-gateway) - should we respect the translate_wan_addrs configuration here or just always use the wan for cross-dc?
|
|
_, addr, port := ep.BestAddress(!localKey.Matches(ep.Node.Datacenter, ep.Node.PartitionOrDefault()))
|
|
healthStatus, weight := calculateEndpointHealthAndWeight(ep, endpointGroup.OnlyPassing)
|
|
|
|
if endpointGroup.OverrideHealth != pbproxystate.HealthStatus_HEALTH_STATUS_UNKNOWN {
|
|
healthStatus = endpointGroup.OverrideHealth
|
|
}
|
|
|
|
endpoint := makeHostPortEndpoint(addr, port)
|
|
endpoint.HealthStatus = healthStatus
|
|
endpoint.LoadBalancingWeight = response.MakeUint32Value(weight)
|
|
|
|
pbEndpoints = append(pbEndpoints, endpoint)
|
|
}
|
|
|
|
// TODO(proxystate): what do we do about priority downstream?
|
|
//cla.Endpoints = append(cla.Endpoints, &envoy_endpoint_v3.LocalityLbEndpoints{
|
|
// Priority: priority,
|
|
// LbEndpoints: es,
|
|
//})
|
|
|
|
priority++
|
|
}
|
|
}
|
|
|
|
return pbEndpoints
|
|
}
|
|
|
|
func makeLoadAssignmentEndpointGroup(
|
|
targets map[string]*structs.DiscoveryTarget,
|
|
targetHealth map[string]structs.CheckServiceNodes,
|
|
gatewayHealth map[string]structs.CheckServiceNodes,
|
|
targetID string,
|
|
localKey proxycfg.GatewayKey,
|
|
forMeshGateway bool,
|
|
) (loadAssignmentEndpointGroup, bool) {
|
|
realEndpoints, ok := targetHealth[targetID]
|
|
if !ok {
|
|
// skip the cluster if we're still populating the snapshot
|
|
return loadAssignmentEndpointGroup{}, false
|
|
}
|
|
target := targets[targetID]
|
|
|
|
var gatewayKey proxycfg.GatewayKey
|
|
|
|
switch target.MeshGateway.Mode {
|
|
case structs.MeshGatewayModeRemote:
|
|
gatewayKey.Datacenter = target.Datacenter
|
|
gatewayKey.Partition = target.Partition
|
|
case structs.MeshGatewayModeLocal:
|
|
gatewayKey = localKey
|
|
}
|
|
|
|
if forMeshGateway || gatewayKey.IsEmpty() || localKey.Matches(target.Datacenter, target.Partition) {
|
|
// Gateways are not needed if the request isn't for a remote DC or partition.
|
|
return loadAssignmentEndpointGroup{
|
|
Endpoints: realEndpoints,
|
|
OnlyPassing: target.Subset.OnlyPassing,
|
|
}, true
|
|
}
|
|
|
|
// If using a mesh gateway we need to pull those endpoints instead.
|
|
gatewayEndpoints, ok := gatewayHealth[gatewayKey.String()]
|
|
if !ok {
|
|
// skip the cluster if we're still populating the snapshot
|
|
return loadAssignmentEndpointGroup{}, false
|
|
}
|
|
|
|
// But we will use the health from the actual backend service.
|
|
overallHealth := pbproxystate.HealthStatus_HEALTH_STATUS_UNHEALTHY
|
|
for _, ep := range realEndpoints {
|
|
health, _ := calculateEndpointHealthAndWeight(ep, target.Subset.OnlyPassing)
|
|
if health == pbproxystate.HealthStatus_HEALTH_STATUS_HEALTHY {
|
|
overallHealth = pbproxystate.HealthStatus_HEALTH_STATUS_HEALTHY
|
|
break
|
|
}
|
|
}
|
|
|
|
return loadAssignmentEndpointGroup{
|
|
Endpoints: gatewayEndpoints,
|
|
OverrideHealth: overallHealth,
|
|
}, true
|
|
}
|
|
|
|
func calculateEndpointHealthAndWeight(
|
|
ep structs.CheckServiceNode,
|
|
onlyPassing bool,
|
|
) (pbproxystate.HealthStatus, int) {
|
|
healthStatus := pbproxystate.HealthStatus_HEALTH_STATUS_HEALTHY
|
|
weight := 1
|
|
if ep.Service.Weights != nil {
|
|
weight = ep.Service.Weights.Passing
|
|
}
|
|
|
|
for _, chk := range ep.Checks {
|
|
if chk.Status == api.HealthCritical {
|
|
healthStatus = pbproxystate.HealthStatus_HEALTH_STATUS_UNHEALTHY
|
|
}
|
|
if onlyPassing && chk.Status != api.HealthPassing {
|
|
healthStatus = pbproxystate.HealthStatus_HEALTH_STATUS_UNHEALTHY
|
|
}
|
|
if chk.Status == api.HealthWarning && ep.Service.Weights != nil {
|
|
weight = ep.Service.Weights.Warning
|
|
}
|
|
}
|
|
// Make weights fit Envoy's limits. A zero weight means that either Warning
|
|
// (likely) or Passing (weirdly) weight has been set to 0 effectively making
|
|
// this instance unhealthy and should not be sent traffic.
|
|
if weight < 1 {
|
|
healthStatus = pbproxystate.HealthStatus_HEALTH_STATUS_UNHEALTHY
|
|
weight = 1
|
|
}
|
|
if weight > 128 {
|
|
weight = 128
|
|
}
|
|
return healthStatus, weight
|
|
}
|