package agent import ( "encoding/json" "net/http" "net/http/httptest" "testing" "time" "github.com/stretchr/testify/require" "github.com/hashicorp/consul/agent/grpc-external/limiter" "github.com/hashicorp/consul/agent/proxycfg" "github.com/hashicorp/consul/agent/structs" "github.com/hashicorp/consul/api" "github.com/hashicorp/consul/testrpc" ) func TestAgent_local_proxycfg(t *testing.T) { a := NewTestAgent(t, TestACLConfig()) defer a.Shutdown() testrpc.WaitForLeader(t, a.RPC, "dc1") token := generateUUID() svc := &structs.NodeService{ ID: "db", Service: "db", Port: 5000, EnterpriseMeta: *structs.DefaultEnterpriseMetaInDefaultPartition(), } require.NoError(t, a.State.AddServiceWithChecks(svc, nil, token)) proxy := &structs.NodeService{ Kind: structs.ServiceKindConnectProxy, ID: "db-sidecar-proxy", Service: "db-sidecar-proxy", Port: 5000, // Set this internal state that we expect sidecar registrations to have. LocallyRegisteredAsSidecar: true, Proxy: structs.ConnectProxyConfig{ DestinationServiceName: "db", Upstreams: structs.TestUpstreams(t), }, EnterpriseMeta: *structs.DefaultEnterpriseMetaInDefaultPartition(), } require.NoError(t, a.State.AddServiceWithChecks(proxy, nil, token)) // This is a little gross, but this gives us the layered pair of // local/catalog sources for now. cfg := a.xdsServer.CfgSrc var ( timer = time.After(100 * time.Millisecond) timerFired = false finalTimer <-chan time.Time ) var ( firstTime = true ch <-chan *proxycfg.ConfigSnapshot stc limiter.SessionTerminatedChan cancel proxycfg.CancelFunc ) defer func() { if cancel != nil { cancel() } }() for { if ch == nil { // Sign up for a stream of config snapshots, in the same manner as the xds server. sid := proxy.CompoundServiceID() if firstTime { firstTime = false } else { t.Logf("re-creating watch") } // Prior to fixes in https://github.com/hashicorp/consul/pull/16497 // this call to Watch() would deadlock. var err error ch, stc, cancel, err = cfg.Watch(sid, a.config.NodeName, token) require.NoError(t, err) } select { case <-stc: t.Fatal("session unexpectedly terminated") case snap, ok := <-ch: if !ok { t.Logf("channel is closed") cancel() ch, stc, cancel = nil, nil, nil continue } require.NotNil(t, snap) if !timerFired { t.Fatal("should not have gotten snapshot until after we manifested the token") } return case <-timer: timerFired = true finalTimer = time.After(1 * time.Second) // This simulates the eventual consistency of a token // showing up on a server after it's creation by // pre-creating the UUID and later using that as the // initial SecretID for a real token. gotToken := testWriteToken(t, a, &api.ACLToken{ AccessorID: generateUUID(), SecretID: token, Description: "my token", ServiceIdentities: []*api.ACLServiceIdentity{{ ServiceName: "db", }}, }) require.Equal(t, token, gotToken) case <-finalTimer: t.Fatal("did not receive a snapshot after the token manifested") } } } func testWriteToken(t *testing.T, a *TestAgent, tok *api.ACLToken) string { req, _ := http.NewRequest("PUT", "/v1/acl/token", jsonReader(tok)) req.Header.Add("X-Consul-Token", "root") resp := httptest.NewRecorder() a.srv.h.ServeHTTP(resp, req) require.Equal(t, http.StatusOK, resp.Code) dec := json.NewDecoder(resp.Body) aclResp := &structs.ACLToken{} require.NoError(t, dec.Decode(aclResp)) return aclResp.SecretID }