| /* |
| * |
| * Copyright 2020 gRPC authors. |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| * |
| */ |
| |
| package client |
| |
| import ( |
| "context" |
| "testing" |
| |
| "google.golang.org/grpc/internal/testutils" |
| ) |
| |
| type clusterUpdateErr struct { |
| u ClusterUpdate |
| err error |
| } |
| |
| // TestClusterWatch covers the cases: |
| // - an update is received after a watch() |
| // - an update for another resource name |
| // - an update is received after cancel() |
| func (s) TestClusterWatch(t *testing.T) { |
| v2ClientCh, cleanup := overrideNewAPIClient() |
| defer cleanup() |
| |
| c, err := New(clientOpts(testXDSServer, false)) |
| if err != nil { |
| t.Fatalf("failed to create client: %v", err) |
| } |
| defer c.Close() |
| |
| // TODO: add a timeout to this recv. |
| // Note that this won't be necessary if we finish the TODO below to call |
| // Client directly instead of v2Client.r. |
| v2Client := <-v2ClientCh |
| |
| clusterUpdateCh := testutils.NewChannel() |
| cancelWatch := c.WatchCluster(testCDSName, func(update ClusterUpdate, err error) { |
| clusterUpdateCh.Send(clusterUpdateErr{u: update, err: err}) |
| }) |
| |
| ctx, cancel := context.WithTimeout(context.Background(), defaultTestTimeout) |
| defer cancel() |
| if _, err := v2Client.addWatches[ClusterResource].Receive(ctx); err != nil { |
| t.Fatalf("want new watch to start, got error %v", err) |
| } |
| |
| wantUpdate := ClusterUpdate{ServiceName: testEDSName} |
| // This is calling v2Client.r to send the update, but r is set to Client, so |
| // this is same as calling Client to update. The one thing this covers is |
| // that `NewXDSV2Client` is called with the right parent. |
| // |
| // TODO: in a future cleanup, this (and the same thing in other tests) can |
| // be changed call Client directly. |
| v2Client.r.NewClusters(map[string]ClusterUpdate{ |
| testCDSName: wantUpdate, |
| }) |
| |
| if u, err := clusterUpdateCh.Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { |
| t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) |
| } |
| |
| // Another update, with an extra resource for a different resource name. |
| v2Client.r.NewClusters(map[string]ClusterUpdate{ |
| testCDSName: wantUpdate, |
| "randomName": {}, |
| }) |
| |
| if u, err := clusterUpdateCh.Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { |
| t.Errorf("unexpected clusterUpdate: %+v, %v, want channel recv timeout", u, err) |
| } |
| |
| // Cancel watch, and send update again. |
| cancelWatch() |
| v2Client.r.NewClusters(map[string]ClusterUpdate{ |
| testCDSName: wantUpdate, |
| }) |
| |
| if u, err := clusterUpdateCh.Receive(ctx); err != context.DeadlineExceeded { |
| t.Errorf("unexpected clusterUpdate: %v, %v, want channel recv timeout", u, err) |
| } |
| } |
| |
| // TestClusterTwoWatchSameResourceName covers the case where an update is received |
| // after two watch() for the same resource name. |
| func (s) TestClusterTwoWatchSameResourceName(t *testing.T) { |
| v2ClientCh, cleanup := overrideNewAPIClient() |
| defer cleanup() |
| |
| c, err := New(clientOpts(testXDSServer, false)) |
| if err != nil { |
| t.Fatalf("failed to create client: %v", err) |
| } |
| defer c.Close() |
| |
| v2Client := <-v2ClientCh |
| |
| var clusterUpdateChs []*testutils.Channel |
| var cancelLastWatch func() |
| |
| ctx, cancel := context.WithTimeout(context.Background(), defaultTestTimeout) |
| defer cancel() |
| |
| const count = 2 |
| for i := 0; i < count; i++ { |
| clusterUpdateCh := testutils.NewChannel() |
| clusterUpdateChs = append(clusterUpdateChs, clusterUpdateCh) |
| cancelLastWatch = c.WatchCluster(testCDSName, func(update ClusterUpdate, err error) { |
| clusterUpdateCh.Send(clusterUpdateErr{u: update, err: err}) |
| }) |
| |
| if i == 0 { |
| // A new watch is registered on the underlying API client only for |
| // the first iteration because we are using the same resource name. |
| if _, err := v2Client.addWatches[ClusterResource].Receive(ctx); err != nil { |
| t.Fatalf("want new watch to start, got error %v", err) |
| } |
| } |
| } |
| |
| wantUpdate := ClusterUpdate{ServiceName: testEDSName} |
| v2Client.r.NewClusters(map[string]ClusterUpdate{ |
| testCDSName: wantUpdate, |
| }) |
| |
| for i := 0; i < count; i++ { |
| if u, err := clusterUpdateChs[i].Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { |
| t.Errorf("i=%v, unexpected clusterUpdate: %v, error receiving from channel: %v", i, u, err) |
| } |
| } |
| |
| // Cancel the last watch, and send update again. |
| cancelLastWatch() |
| v2Client.r.NewClusters(map[string]ClusterUpdate{ |
| testCDSName: wantUpdate, |
| }) |
| |
| for i := 0; i < count-1; i++ { |
| if u, err := clusterUpdateChs[i].Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { |
| t.Errorf("i=%v, unexpected clusterUpdate: %v, error receiving from channel: %v", i, u, err) |
| } |
| } |
| |
| if u, err := clusterUpdateChs[count-1].Receive(ctx); err != context.DeadlineExceeded { |
| t.Errorf("unexpected clusterUpdate: %v, %v, want channel recv timeout", u, err) |
| } |
| } |
| |
| // TestClusterThreeWatchDifferentResourceName covers the case where an update is |
| // received after three watch() for different resource names. |
| func (s) TestClusterThreeWatchDifferentResourceName(t *testing.T) { |
| v2ClientCh, cleanup := overrideNewAPIClient() |
| defer cleanup() |
| |
| c, err := New(clientOpts(testXDSServer, false)) |
| if err != nil { |
| t.Fatalf("failed to create client: %v", err) |
| } |
| defer c.Close() |
| |
| v2Client := <-v2ClientCh |
| |
| var clusterUpdateChs []*testutils.Channel |
| const count = 2 |
| |
| // Two watches for the same name. |
| ctx, cancel := context.WithTimeout(context.Background(), defaultTestTimeout) |
| defer cancel() |
| for i := 0; i < count; i++ { |
| clusterUpdateCh := testutils.NewChannel() |
| clusterUpdateChs = append(clusterUpdateChs, clusterUpdateCh) |
| c.WatchCluster(testCDSName+"1", func(update ClusterUpdate, err error) { |
| clusterUpdateCh.Send(clusterUpdateErr{u: update, err: err}) |
| }) |
| |
| if i == 0 { |
| // A new watch is registered on the underlying API client only for |
| // the first iteration because we are using the same resource name. |
| if _, err := v2Client.addWatches[ClusterResource].Receive(ctx); err != nil { |
| t.Fatalf("want new watch to start, got error %v", err) |
| } |
| } |
| } |
| |
| // Third watch for a different name. |
| clusterUpdateCh2 := testutils.NewChannel() |
| c.WatchCluster(testCDSName+"2", func(update ClusterUpdate, err error) { |
| clusterUpdateCh2.Send(clusterUpdateErr{u: update, err: err}) |
| }) |
| if _, err := v2Client.addWatches[ClusterResource].Receive(ctx); err != nil { |
| t.Fatalf("want new watch to start, got error %v", err) |
| } |
| |
| wantUpdate1 := ClusterUpdate{ServiceName: testEDSName + "1"} |
| wantUpdate2 := ClusterUpdate{ServiceName: testEDSName + "2"} |
| v2Client.r.NewClusters(map[string]ClusterUpdate{ |
| testCDSName + "1": wantUpdate1, |
| testCDSName + "2": wantUpdate2, |
| }) |
| |
| for i := 0; i < count; i++ { |
| if u, err := clusterUpdateChs[i].Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate1, nil}) { |
| t.Errorf("i=%v, unexpected clusterUpdate: %v, error receiving from channel: %v", i, u, err) |
| } |
| } |
| |
| if u, err := clusterUpdateCh2.Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate2, nil}) { |
| t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) |
| } |
| } |
| |
| // TestClusterWatchAfterCache covers the case where watch is called after the update |
| // is in cache. |
| func (s) TestClusterWatchAfterCache(t *testing.T) { |
| v2ClientCh, cleanup := overrideNewAPIClient() |
| defer cleanup() |
| |
| c, err := New(clientOpts(testXDSServer, false)) |
| if err != nil { |
| t.Fatalf("failed to create client: %v", err) |
| } |
| defer c.Close() |
| |
| v2Client := <-v2ClientCh |
| |
| clusterUpdateCh := testutils.NewChannel() |
| c.WatchCluster(testCDSName, func(update ClusterUpdate, err error) { |
| clusterUpdateCh.Send(clusterUpdateErr{u: update, err: err}) |
| }) |
| |
| ctx, cancel := context.WithTimeout(context.Background(), defaultTestTimeout) |
| defer cancel() |
| if _, err := v2Client.addWatches[ClusterResource].Receive(ctx); err != nil { |
| t.Fatalf("want new watch to start, got error %v", err) |
| } |
| |
| wantUpdate := ClusterUpdate{ServiceName: testEDSName} |
| v2Client.r.NewClusters(map[string]ClusterUpdate{ |
| testCDSName: wantUpdate, |
| }) |
| |
| if u, err := clusterUpdateCh.Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { |
| t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) |
| } |
| |
| // Another watch for the resource in cache. |
| clusterUpdateCh2 := testutils.NewChannel() |
| c.WatchCluster(testCDSName, func(update ClusterUpdate, err error) { |
| clusterUpdateCh2.Send(clusterUpdateErr{u: update, err: err}) |
| }) |
| if n, err := v2Client.addWatches[ClusterResource].Receive(ctx); err != context.DeadlineExceeded { |
| t.Fatalf("want no new watch to start (recv timeout), got resource name: %v error %v", n, err) |
| } |
| |
| // New watch should receives the update. |
| ctx, cancel = context.WithTimeout(context.Background(), defaultTestTimeout) |
| defer cancel() |
| if u, err := clusterUpdateCh2.Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { |
| t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) |
| } |
| |
| // Old watch should see nothing. |
| if u, err := clusterUpdateCh.Receive(ctx); err != context.DeadlineExceeded { |
| t.Errorf("unexpected clusterUpdate: %v, %v, want channel recv timeout", u, err) |
| } |
| } |
| |
| // TestClusterWatchExpiryTimer tests the case where the client does not receive |
| // an CDS response for the request that it sends out. We want the watch callback |
| // to be invoked with an error once the watchExpiryTimer fires. |
| func (s) TestClusterWatchExpiryTimer(t *testing.T) { |
| v2ClientCh, cleanup := overrideNewAPIClient() |
| defer cleanup() |
| |
| c, err := New(clientOpts(testXDSServer, true)) |
| if err != nil { |
| t.Fatalf("failed to create client: %v", err) |
| } |
| defer c.Close() |
| |
| v2Client := <-v2ClientCh |
| |
| clusterUpdateCh := testutils.NewChannel() |
| c.WatchCluster(testCDSName, func(u ClusterUpdate, err error) { |
| clusterUpdateCh.Send(clusterUpdateErr{u: u, err: err}) |
| }) |
| |
| ctx, cancel := context.WithTimeout(context.Background(), defaultTestTimeout) |
| defer cancel() |
| if _, err := v2Client.addWatches[ClusterResource].Receive(ctx); err != nil { |
| t.Fatalf("want new watch to start, got error %v", err) |
| } |
| |
| u, err := clusterUpdateCh.Receive(ctx) |
| if err != nil { |
| t.Fatalf("failed to get clusterUpdate: %v", err) |
| } |
| uu := u.(clusterUpdateErr) |
| if uu.u != (ClusterUpdate{}) { |
| t.Errorf("unexpected clusterUpdate: %v, want %v", uu.u, ClusterUpdate{}) |
| } |
| if uu.err == nil { |
| t.Errorf("unexpected clusterError: <nil>, want error watcher timeout") |
| } |
| } |
| |
| // TestClusterWatchExpiryTimerStop tests the case where the client does receive |
| // an CDS response for the request that it sends out. We want no error even |
| // after expiry timeout. |
| func (s) TestClusterWatchExpiryTimerStop(t *testing.T) { |
| v2ClientCh, cleanup := overrideNewAPIClient() |
| defer cleanup() |
| |
| c, err := New(clientOpts(testXDSServer, true)) |
| if err != nil { |
| t.Fatalf("failed to create client: %v", err) |
| } |
| defer c.Close() |
| |
| v2Client := <-v2ClientCh |
| |
| clusterUpdateCh := testutils.NewChannel() |
| c.WatchCluster(testCDSName, func(u ClusterUpdate, err error) { |
| clusterUpdateCh.Send(clusterUpdateErr{u: u, err: err}) |
| }) |
| |
| ctx, cancel := context.WithTimeout(context.Background(), defaultTestTimeout) |
| defer cancel() |
| if _, err := v2Client.addWatches[ClusterResource].Receive(ctx); err != nil { |
| t.Fatalf("want new watch to start, got error %v", err) |
| } |
| |
| wantUpdate := ClusterUpdate{ServiceName: testEDSName} |
| v2Client.r.NewClusters(map[string]ClusterUpdate{ |
| testCDSName: wantUpdate, |
| }) |
| |
| if u, err := clusterUpdateCh.Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { |
| t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) |
| } |
| |
| // Wait for an error, the error should never happen. |
| u, err := clusterUpdateCh.Receive(ctx) |
| if err != context.DeadlineExceeded { |
| t.Fatalf("got unexpected: %v, %v, want recv timeout", u.(clusterUpdateErr).u, u.(clusterUpdateErr).err) |
| } |
| } |
| |
| // TestClusterResourceRemoved covers the cases: |
| // - an update is received after a watch() |
| // - another update is received, with one resource removed |
| // - this should trigger callback with resource removed error |
| // - one more update without the removed resource |
| // - the callback (above) shouldn't receive any update |
| func (s) TestClusterResourceRemoved(t *testing.T) { |
| v2ClientCh, cleanup := overrideNewAPIClient() |
| defer cleanup() |
| |
| c, err := New(clientOpts(testXDSServer, false)) |
| if err != nil { |
| t.Fatalf("failed to create client: %v", err) |
| } |
| defer c.Close() |
| |
| v2Client := <-v2ClientCh |
| |
| clusterUpdateCh1 := testutils.NewChannel() |
| c.WatchCluster(testCDSName+"1", func(update ClusterUpdate, err error) { |
| clusterUpdateCh1.Send(clusterUpdateErr{u: update, err: err}) |
| }) |
| |
| ctx, cancel := context.WithTimeout(context.Background(), defaultTestTimeout) |
| defer cancel() |
| if _, err := v2Client.addWatches[ClusterResource].Receive(ctx); err != nil { |
| t.Fatalf("want new watch to start, got error %v", err) |
| } |
| // Another watch for a different name. |
| clusterUpdateCh2 := testutils.NewChannel() |
| c.WatchCluster(testCDSName+"2", func(update ClusterUpdate, err error) { |
| clusterUpdateCh2.Send(clusterUpdateErr{u: update, err: err}) |
| }) |
| if _, err := v2Client.addWatches[ClusterResource].Receive(ctx); err != nil { |
| t.Fatalf("want new watch to start, got error %v", err) |
| } |
| |
| wantUpdate1 := ClusterUpdate{ServiceName: testEDSName + "1"} |
| wantUpdate2 := ClusterUpdate{ServiceName: testEDSName + "2"} |
| v2Client.r.NewClusters(map[string]ClusterUpdate{ |
| testCDSName + "1": wantUpdate1, |
| testCDSName + "2": wantUpdate2, |
| }) |
| |
| if u, err := clusterUpdateCh1.Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate1, nil}) { |
| t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) |
| } |
| |
| if u, err := clusterUpdateCh2.Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate2, nil}) { |
| t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) |
| } |
| |
| // Send another update to remove resource 1. |
| v2Client.r.NewClusters(map[string]ClusterUpdate{ |
| testCDSName + "2": wantUpdate2, |
| }) |
| |
| // watcher 1 should get an error. |
| if u, err := clusterUpdateCh1.Receive(ctx); err != nil || ErrType(u.(clusterUpdateErr).err) != ErrorTypeResourceNotFound { |
| t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v, want update with error resource not found", u, err) |
| } |
| |
| // watcher 2 should get the same update again. |
| if u, err := clusterUpdateCh2.Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate2, nil}) { |
| t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) |
| } |
| |
| // Send one more update without resource 1. |
| v2Client.r.NewClusters(map[string]ClusterUpdate{ |
| testCDSName + "2": wantUpdate2, |
| }) |
| |
| // watcher 1 should get an error. |
| if u, err := clusterUpdateCh1.Receive(ctx); err != context.DeadlineExceeded { |
| t.Errorf("unexpected clusterUpdate: %v, want receiving from channel timeout", u) |
| } |
| |
| // watcher 2 should get the same update again. |
| ctx, cancel = context.WithTimeout(context.Background(), defaultTestTimeout) |
| defer cancel() |
| if u, err := clusterUpdateCh2.Receive(ctx); err != nil || u != (clusterUpdateErr{wantUpdate2, nil}) { |
| t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) |
| } |
| } |