| /* |
| * |
| * Copyright 2018 gRPC authors. |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| * |
| */ |
| |
| package test |
| |
| import ( |
| "context" |
| "errors" |
| "fmt" |
| "net" |
| "sync" |
| "testing" |
| "time" |
| |
| "google.golang.org/grpc" |
| "google.golang.org/grpc/codes" |
| "google.golang.org/grpc/connectivity" |
| _ "google.golang.org/grpc/health" |
| healthgrpc "google.golang.org/grpc/health/grpc_health_v1" |
| healthpb "google.golang.org/grpc/health/grpc_health_v1" |
| "google.golang.org/grpc/internal" |
| "google.golang.org/grpc/internal/channelz" |
| "google.golang.org/grpc/internal/grpctest" |
| "google.golang.org/grpc/resolver" |
| "google.golang.org/grpc/resolver/manual" |
| "google.golang.org/grpc/status" |
| testpb "google.golang.org/grpc/test/grpc_testing" |
| ) |
| |
| var testHealthCheckFunc = internal.HealthCheckFunc |
| |
| func newTestHealthServer() *testHealthServer { |
| return newTestHealthServerWithWatchFunc(defaultWatchFunc) |
| } |
| |
| func newTestHealthServerWithWatchFunc(f func(s *testHealthServer, in *healthpb.HealthCheckRequest, stream healthgrpc.Health_WatchServer) error) *testHealthServer { |
| return &testHealthServer{ |
| watchFunc: f, |
| update: make(chan struct{}, 1), |
| status: make(map[string]healthpb.HealthCheckResponse_ServingStatus), |
| } |
| } |
| |
| // defaultWatchFunc will send a HealthCheckResponse to the client whenever SetServingStatus is called. |
| func defaultWatchFunc(s *testHealthServer, in *healthpb.HealthCheckRequest, stream healthgrpc.Health_WatchServer) error { |
| if in.Service != "foo" { |
| return status.Error(codes.FailedPrecondition, |
| "the defaultWatchFunc only handles request with service name to be \"foo\"") |
| } |
| var done bool |
| for { |
| select { |
| case <-stream.Context().Done(): |
| done = true |
| case <-s.update: |
| } |
| if done { |
| break |
| } |
| s.mu.Lock() |
| resp := &healthpb.HealthCheckResponse{ |
| Status: s.status[in.Service], |
| } |
| s.mu.Unlock() |
| stream.SendMsg(resp) |
| } |
| return nil |
| } |
| |
| type testHealthServer struct { |
| healthpb.UnimplementedHealthServer |
| watchFunc func(s *testHealthServer, in *healthpb.HealthCheckRequest, stream healthgrpc.Health_WatchServer) error |
| mu sync.Mutex |
| status map[string]healthpb.HealthCheckResponse_ServingStatus |
| update chan struct{} |
| } |
| |
| func (s *testHealthServer) Check(ctx context.Context, in *healthpb.HealthCheckRequest) (*healthpb.HealthCheckResponse, error) { |
| return &healthpb.HealthCheckResponse{ |
| Status: healthpb.HealthCheckResponse_SERVING, |
| }, nil |
| } |
| |
| func (s *testHealthServer) Watch(in *healthpb.HealthCheckRequest, stream healthgrpc.Health_WatchServer) error { |
| return s.watchFunc(s, in, stream) |
| } |
| |
| // SetServingStatus is called when need to reset the serving status of a service |
| // or insert a new service entry into the statusMap. |
| func (s *testHealthServer) SetServingStatus(service string, status healthpb.HealthCheckResponse_ServingStatus) { |
| s.mu.Lock() |
| s.status[service] = status |
| select { |
| case <-s.update: |
| default: |
| } |
| s.update <- struct{}{} |
| s.mu.Unlock() |
| } |
| |
| func setupHealthCheckWrapper() (hcEnterChan chan struct{}, hcExitChan chan struct{}, wrapper internal.HealthChecker) { |
| hcEnterChan = make(chan struct{}) |
| hcExitChan = make(chan struct{}) |
| wrapper = func(ctx context.Context, newStream func(string) (interface{}, error), update func(connectivity.State, error), service string) error { |
| close(hcEnterChan) |
| defer close(hcExitChan) |
| return testHealthCheckFunc(ctx, newStream, update, service) |
| } |
| return |
| } |
| |
| type svrConfig struct { |
| specialWatchFunc func(s *testHealthServer, in *healthpb.HealthCheckRequest, stream healthgrpc.Health_WatchServer) error |
| } |
| |
| func setupServer(sc *svrConfig) (s *grpc.Server, lis net.Listener, ts *testHealthServer, deferFunc func(), err error) { |
| s = grpc.NewServer() |
| lis, err = net.Listen("tcp", "localhost:0") |
| if err != nil { |
| return nil, nil, nil, func() {}, fmt.Errorf("failed to listen due to err %v", err) |
| } |
| if sc.specialWatchFunc != nil { |
| ts = newTestHealthServerWithWatchFunc(sc.specialWatchFunc) |
| } else { |
| ts = newTestHealthServer() |
| } |
| healthgrpc.RegisterHealthServer(s, ts) |
| testpb.RegisterTestServiceServer(s, &testServer{}) |
| go s.Serve(lis) |
| return s, lis, ts, s.Stop, nil |
| } |
| |
| type clientConfig struct { |
| balancerName string |
| testHealthCheckFuncWrapper internal.HealthChecker |
| extraDialOption []grpc.DialOption |
| } |
| |
| func setupClient(c *clientConfig) (cc *grpc.ClientConn, r *manual.Resolver, deferFunc func(), err error) { |
| r = manual.NewBuilderWithScheme("whatever") |
| var opts []grpc.DialOption |
| opts = append(opts, grpc.WithInsecure(), grpc.WithResolvers(r), grpc.WithBalancerName(c.balancerName)) |
| if c.testHealthCheckFuncWrapper != nil { |
| opts = append(opts, internal.WithHealthCheckFunc.(func(internal.HealthChecker) grpc.DialOption)(c.testHealthCheckFuncWrapper)) |
| } |
| opts = append(opts, c.extraDialOption...) |
| cc, err = grpc.Dial(r.Scheme()+":///test.server", opts...) |
| if err != nil { |
| |
| return nil, nil, nil, fmt.Errorf("dial failed due to err: %v", err) |
| } |
| return cc, r, func() { cc.Close() }, nil |
| } |
| |
| func (s) TestHealthCheckWatchStateChange(t *testing.T) { |
| _, lis, ts, deferFunc, err := setupServer(&svrConfig{}) |
| defer deferFunc() |
| if err != nil { |
| t.Fatal(err) |
| } |
| |
| // The table below shows the expected series of addrConn connectivity transitions when server |
| // updates its health status. As there's only one addrConn corresponds with the ClientConn in this |
| // test, we use ClientConn's connectivity state as the addrConn connectivity state. |
| //+------------------------------+-------------------------------------------+ |
| //| Health Check Returned Status | Expected addrConn Connectivity Transition | |
| //+------------------------------+-------------------------------------------+ |
| //| NOT_SERVING | ->TRANSIENT FAILURE | |
| //| SERVING | ->READY | |
| //| SERVICE_UNKNOWN | ->TRANSIENT FAILURE | |
| //| SERVING | ->READY | |
| //| UNKNOWN | ->TRANSIENT FAILURE | |
| //+------------------------------+-------------------------------------------+ |
| ts.SetServingStatus("foo", healthpb.HealthCheckResponse_NOT_SERVING) |
| |
| cc, r, deferFunc, err := setupClient(&clientConfig{balancerName: "round_robin"}) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| r.UpdateState(resolver.State{ |
| Addresses: []resolver.Address{{Addr: lis.Addr().String()}}, |
| ServiceConfig: parseCfg(r, `{ |
| "healthCheckConfig": { |
| "serviceName": "foo" |
| } |
| }`)}) |
| ctx, cancel := context.WithTimeout(context.Background(), 10*time.Second) |
| defer cancel() |
| if ok := cc.WaitForStateChange(ctx, connectivity.Idle); !ok { |
| t.Fatal("ClientConn is still in IDLE state when the context times out.") |
| } |
| if ok := cc.WaitForStateChange(ctx, connectivity.Connecting); !ok { |
| t.Fatal("ClientConn is still in CONNECTING state when the context times out.") |
| } |
| if s := cc.GetState(); s != connectivity.TransientFailure { |
| t.Fatalf("ClientConn is in %v state, want TRANSIENT FAILURE", s) |
| } |
| |
| ts.SetServingStatus("foo", healthpb.HealthCheckResponse_SERVING) |
| if ok := cc.WaitForStateChange(ctx, connectivity.TransientFailure); !ok { |
| t.Fatal("ClientConn is still in TRANSIENT FAILURE state when the context times out.") |
| } |
| if s := cc.GetState(); s != connectivity.Ready { |
| t.Fatalf("ClientConn is in %v state, want READY", s) |
| } |
| |
| ts.SetServingStatus("foo", healthpb.HealthCheckResponse_SERVICE_UNKNOWN) |
| if ok := cc.WaitForStateChange(ctx, connectivity.Ready); !ok { |
| t.Fatal("ClientConn is still in READY state when the context times out.") |
| } |
| if s := cc.GetState(); s != connectivity.TransientFailure { |
| t.Fatalf("ClientConn is in %v state, want TRANSIENT FAILURE", s) |
| } |
| |
| ts.SetServingStatus("foo", healthpb.HealthCheckResponse_SERVING) |
| if ok := cc.WaitForStateChange(ctx, connectivity.TransientFailure); !ok { |
| t.Fatal("ClientConn is still in TRANSIENT FAILURE state when the context times out.") |
| } |
| if s := cc.GetState(); s != connectivity.Ready { |
| t.Fatalf("ClientConn is in %v state, want READY", s) |
| } |
| |
| ts.SetServingStatus("foo", healthpb.HealthCheckResponse_UNKNOWN) |
| if ok := cc.WaitForStateChange(ctx, connectivity.Ready); !ok { |
| t.Fatal("ClientConn is still in READY state when the context times out.") |
| } |
| if s := cc.GetState(); s != connectivity.TransientFailure { |
| t.Fatalf("ClientConn is in %v state, want TRANSIENT FAILURE", s) |
| } |
| } |
| |
| // If Watch returns Unimplemented, then the ClientConn should go into READY state. |
| func (s) TestHealthCheckHealthServerNotRegistered(t *testing.T) { |
| grpctest.TLogger.ExpectError("Subchannel health check is unimplemented at server side, thus health check is disabled") |
| s := grpc.NewServer() |
| lis, err := net.Listen("tcp", "localhost:0") |
| if err != nil { |
| t.Fatalf("failed to listen due to err: %v", err) |
| } |
| go s.Serve(lis) |
| defer s.Stop() |
| |
| cc, r, deferFunc, err := setupClient(&clientConfig{balancerName: "round_robin"}) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| r.UpdateState(resolver.State{ |
| Addresses: []resolver.Address{{Addr: lis.Addr().String()}}, |
| ServiceConfig: parseCfg(r, `{ |
| "healthCheckConfig": { |
| "serviceName": "foo" |
| } |
| }`)}) |
| ctx, cancel := context.WithTimeout(context.Background(), 10*time.Second) |
| defer cancel() |
| |
| if ok := cc.WaitForStateChange(ctx, connectivity.Idle); !ok { |
| t.Fatal("ClientConn is still in IDLE state when the context times out.") |
| } |
| if ok := cc.WaitForStateChange(ctx, connectivity.Connecting); !ok { |
| t.Fatal("ClientConn is still in CONNECTING state when the context times out.") |
| } |
| if s := cc.GetState(); s != connectivity.Ready { |
| t.Fatalf("ClientConn is in %v state, want READY", s) |
| } |
| } |
| |
| // In the case of a goaway received, the health check stream should be terminated and health check |
| // function should exit. |
| func (s) TestHealthCheckWithGoAway(t *testing.T) { |
| hcEnterChan, hcExitChan, testHealthCheckFuncWrapper := setupHealthCheckWrapper() |
| |
| s, lis, ts, deferFunc, err := setupServer(&svrConfig{}) |
| defer deferFunc() |
| if err != nil { |
| t.Fatal(err) |
| } |
| |
| ts.SetServingStatus("foo", healthpb.HealthCheckResponse_SERVING) |
| |
| cc, r, deferFunc, err := setupClient(&clientConfig{ |
| balancerName: "round_robin", |
| testHealthCheckFuncWrapper: testHealthCheckFuncWrapper, |
| }) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| tc := testpb.NewTestServiceClient(cc) |
| r.UpdateState(resolver.State{ |
| Addresses: []resolver.Address{{Addr: lis.Addr().String()}}, |
| ServiceConfig: parseCfg(r, `{ |
| "healthCheckConfig": { |
| "serviceName": "foo" |
| } |
| }`)}) |
| |
| // make some rpcs to make sure connection is working. |
| if err := verifyResultWithDelay(func() (bool, error) { |
| if _, err := tc.EmptyCall(context.Background(), &testpb.Empty{}); err != nil { |
| return false, fmt.Errorf("TestService/EmptyCall(_, _) = _, %v, want _, <nil>", err) |
| } |
| return true, nil |
| }); err != nil { |
| t.Fatal(err) |
| } |
| |
| // the stream rpc will persist through goaway event. |
| ctx, cancel := context.WithTimeout(context.Background(), 5*time.Second) |
| defer cancel() |
| stream, err := tc.FullDuplexCall(ctx, grpc.WaitForReady(true)) |
| if err != nil { |
| t.Fatalf("%v.FullDuplexCall(_) = _, %v, want <nil>", tc, err) |
| } |
| respParam := []*testpb.ResponseParameters{{Size: 1}} |
| payload, err := newPayload(testpb.PayloadType_COMPRESSABLE, int32(1)) |
| if err != nil { |
| t.Fatal(err) |
| } |
| req := &testpb.StreamingOutputCallRequest{ |
| ResponseParameters: respParam, |
| Payload: payload, |
| } |
| if err := stream.Send(req); err != nil { |
| t.Fatalf("%v.Send(_) = %v, want <nil>", stream, err) |
| } |
| if _, err := stream.Recv(); err != nil { |
| t.Fatalf("%v.Recv() = _, %v, want _, <nil>", stream, err) |
| } |
| |
| select { |
| case <-hcExitChan: |
| t.Fatal("Health check function has exited, which is not expected.") |
| default: |
| } |
| |
| // server sends GoAway |
| go s.GracefulStop() |
| |
| select { |
| case <-hcExitChan: |
| case <-time.After(5 * time.Second): |
| select { |
| case <-hcEnterChan: |
| default: |
| t.Fatal("Health check function has not entered after 5s.") |
| } |
| t.Fatal("Health check function has not exited after 5s.") |
| } |
| |
| // The existing RPC should be still good to proceed. |
| if err := stream.Send(req); err != nil { |
| t.Fatalf("%v.Send(_) = %v, want <nil>", stream, err) |
| } |
| if _, err := stream.Recv(); err != nil { |
| t.Fatalf("%v.Recv() = _, %v, want _, <nil>", stream, err) |
| } |
| } |
| |
| func (s) TestHealthCheckWithConnClose(t *testing.T) { |
| hcEnterChan, hcExitChan, testHealthCheckFuncWrapper := setupHealthCheckWrapper() |
| |
| s, lis, ts, deferFunc, err := setupServer(&svrConfig{}) |
| defer deferFunc() |
| if err != nil { |
| t.Fatal(err) |
| } |
| |
| ts.SetServingStatus("foo", healthpb.HealthCheckResponse_SERVING) |
| |
| cc, r, deferFunc, err := setupClient(&clientConfig{ |
| balancerName: "round_robin", |
| testHealthCheckFuncWrapper: testHealthCheckFuncWrapper, |
| }) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| tc := testpb.NewTestServiceClient(cc) |
| |
| r.UpdateState(resolver.State{ |
| Addresses: []resolver.Address{{Addr: lis.Addr().String()}}, |
| ServiceConfig: parseCfg(r, `{ |
| "healthCheckConfig": { |
| "serviceName": "foo" |
| } |
| }`)}) |
| |
| // make some rpcs to make sure connection is working. |
| if err := verifyResultWithDelay(func() (bool, error) { |
| if _, err := tc.EmptyCall(context.Background(), &testpb.Empty{}); err != nil { |
| return false, fmt.Errorf("TestService/EmptyCall(_, _) = _, %v, want _, <nil>", err) |
| } |
| return true, nil |
| }); err != nil { |
| t.Fatal(err) |
| } |
| |
| select { |
| case <-hcExitChan: |
| t.Fatal("Health check function has exited, which is not expected.") |
| default: |
| } |
| // server closes the connection |
| s.Stop() |
| |
| select { |
| case <-hcExitChan: |
| case <-time.After(5 * time.Second): |
| select { |
| case <-hcEnterChan: |
| default: |
| t.Fatal("Health check function has not entered after 5s.") |
| } |
| t.Fatal("Health check function has not exited after 5s.") |
| } |
| } |
| |
| // addrConn drain happens when addrConn gets torn down due to its address being no longer in the |
| // address list returned by the resolver. |
| func (s) TestHealthCheckWithAddrConnDrain(t *testing.T) { |
| hcEnterChan, hcExitChan, testHealthCheckFuncWrapper := setupHealthCheckWrapper() |
| |
| _, lis, ts, deferFunc, err := setupServer(&svrConfig{}) |
| defer deferFunc() |
| if err != nil { |
| t.Fatal(err) |
| } |
| |
| ts.SetServingStatus("foo", healthpb.HealthCheckResponse_SERVING) |
| |
| cc, r, deferFunc, err := setupClient(&clientConfig{ |
| balancerName: "round_robin", |
| testHealthCheckFuncWrapper: testHealthCheckFuncWrapper, |
| }) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| tc := testpb.NewTestServiceClient(cc) |
| sc := parseCfg(r, `{ |
| "healthCheckConfig": { |
| "serviceName": "foo" |
| } |
| }`) |
| r.UpdateState(resolver.State{ |
| Addresses: []resolver.Address{{Addr: lis.Addr().String()}}, |
| ServiceConfig: sc, |
| }) |
| |
| // make some rpcs to make sure connection is working. |
| if err := verifyResultWithDelay(func() (bool, error) { |
| if _, err := tc.EmptyCall(context.Background(), &testpb.Empty{}); err != nil { |
| return false, fmt.Errorf("TestService/EmptyCall(_, _) = _, %v, want _, <nil>", err) |
| } |
| return true, nil |
| }); err != nil { |
| t.Fatal(err) |
| } |
| |
| // the stream rpc will persist through goaway event. |
| ctx, cancel := context.WithTimeout(context.Background(), 5*time.Second) |
| defer cancel() |
| stream, err := tc.FullDuplexCall(ctx, grpc.WaitForReady(true)) |
| if err != nil { |
| t.Fatalf("%v.FullDuplexCall(_) = _, %v, want <nil>", tc, err) |
| } |
| respParam := []*testpb.ResponseParameters{{Size: 1}} |
| payload, err := newPayload(testpb.PayloadType_COMPRESSABLE, int32(1)) |
| if err != nil { |
| t.Fatal(err) |
| } |
| req := &testpb.StreamingOutputCallRequest{ |
| ResponseParameters: respParam, |
| Payload: payload, |
| } |
| if err := stream.Send(req); err != nil { |
| t.Fatalf("%v.Send(_) = %v, want <nil>", stream, err) |
| } |
| if _, err := stream.Recv(); err != nil { |
| t.Fatalf("%v.Recv() = _, %v, want _, <nil>", stream, err) |
| } |
| |
| select { |
| case <-hcExitChan: |
| t.Fatal("Health check function has exited, which is not expected.") |
| default: |
| } |
| // trigger teardown of the ac |
| r.UpdateState(resolver.State{Addresses: []resolver.Address{{Addr: "fake address"}}, ServiceConfig: sc}) |
| |
| select { |
| case <-hcExitChan: |
| case <-time.After(5 * time.Second): |
| select { |
| case <-hcEnterChan: |
| default: |
| t.Fatal("Health check function has not entered after 5s.") |
| } |
| t.Fatal("Health check function has not exited after 5s.") |
| } |
| |
| // The existing RPC should be still good to proceed. |
| if err := stream.Send(req); err != nil { |
| t.Fatalf("%v.Send(_) = %v, want <nil>", stream, err) |
| } |
| if _, err := stream.Recv(); err != nil { |
| t.Fatalf("%v.Recv() = _, %v, want _, <nil>", stream, err) |
| } |
| } |
| |
| // ClientConn close will lead to its addrConns being torn down. |
| func (s) TestHealthCheckWithClientConnClose(t *testing.T) { |
| hcEnterChan, hcExitChan, testHealthCheckFuncWrapper := setupHealthCheckWrapper() |
| |
| _, lis, ts, deferFunc, err := setupServer(&svrConfig{}) |
| defer deferFunc() |
| if err != nil { |
| t.Fatal(err) |
| } |
| |
| ts.SetServingStatus("foo", healthpb.HealthCheckResponse_SERVING) |
| |
| cc, r, deferFunc, err := setupClient(&clientConfig{ |
| balancerName: "round_robin", |
| testHealthCheckFuncWrapper: testHealthCheckFuncWrapper, |
| }) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| tc := testpb.NewTestServiceClient(cc) |
| r.UpdateState(resolver.State{ |
| Addresses: []resolver.Address{{Addr: lis.Addr().String()}}, |
| ServiceConfig: parseCfg(r, `{ |
| "healthCheckConfig": { |
| "serviceName": "foo" |
| } |
| }`)}) |
| |
| // make some rpcs to make sure connection is working. |
| if err := verifyResultWithDelay(func() (bool, error) { |
| if _, err := tc.EmptyCall(context.Background(), &testpb.Empty{}); err != nil { |
| return false, fmt.Errorf("TestService/EmptyCall(_, _) = _, %v, want _, <nil>", err) |
| } |
| return true, nil |
| }); err != nil { |
| t.Fatal(err) |
| } |
| |
| select { |
| case <-hcExitChan: |
| t.Fatal("Health check function has exited, which is not expected.") |
| default: |
| } |
| |
| // trigger addrConn teardown |
| cc.Close() |
| |
| select { |
| case <-hcExitChan: |
| case <-time.After(5 * time.Second): |
| select { |
| case <-hcEnterChan: |
| default: |
| t.Fatal("Health check function has not entered after 5s.") |
| } |
| t.Fatal("Health check function has not exited after 5s.") |
| } |
| } |
| |
| // This test is to test the logic in the createTransport after the health check function returns which |
| // closes the skipReset channel(since it has not been closed inside health check func) to unblock |
| // onGoAway/onClose goroutine. |
| func (s) TestHealthCheckWithoutSetConnectivityStateCalledAddrConnShutDown(t *testing.T) { |
| hcEnterChan, hcExitChan, testHealthCheckFuncWrapper := setupHealthCheckWrapper() |
| |
| _, lis, ts, deferFunc, err := setupServer(&svrConfig{ |
| specialWatchFunc: func(s *testHealthServer, in *healthpb.HealthCheckRequest, stream healthgrpc.Health_WatchServer) error { |
| if in.Service != "delay" { |
| return status.Error(codes.FailedPrecondition, |
| "this special Watch function only handles request with service name to be \"delay\"") |
| } |
| // Do nothing to mock a delay of health check response from server side. |
| // This case is to help with the test that covers the condition that setConnectivityState is not |
| // called inside HealthCheckFunc before the func returns. |
| select { |
| case <-stream.Context().Done(): |
| case <-time.After(5 * time.Second): |
| } |
| return nil |
| }, |
| }) |
| defer deferFunc() |
| if err != nil { |
| t.Fatal(err) |
| } |
| |
| ts.SetServingStatus("delay", healthpb.HealthCheckResponse_SERVING) |
| |
| _, r, deferFunc, err := setupClient(&clientConfig{ |
| balancerName: "round_robin", |
| testHealthCheckFuncWrapper: testHealthCheckFuncWrapper, |
| }) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| // The serviceName "delay" is specially handled at server side, where response will not be sent |
| // back to client immediately upon receiving the request (client should receive no response until |
| // test ends). |
| sc := parseCfg(r, `{ |
| "healthCheckConfig": { |
| "serviceName": "delay" |
| } |
| }`) |
| r.UpdateState(resolver.State{ |
| Addresses: []resolver.Address{{Addr: lis.Addr().String()}}, |
| ServiceConfig: sc, |
| }) |
| |
| select { |
| case <-hcExitChan: |
| t.Fatal("Health check function has exited, which is not expected.") |
| default: |
| } |
| |
| select { |
| case <-hcEnterChan: |
| case <-time.After(5 * time.Second): |
| t.Fatal("Health check function has not been invoked after 5s.") |
| } |
| // trigger teardown of the ac, ac in SHUTDOWN state |
| r.UpdateState(resolver.State{Addresses: []resolver.Address{{Addr: "fake address"}}, ServiceConfig: sc}) |
| |
| // The health check func should exit without calling the setConnectivityState func, as server hasn't sent |
| // any response. |
| select { |
| case <-hcExitChan: |
| case <-time.After(5 * time.Second): |
| t.Fatal("Health check function has not exited after 5s.") |
| } |
| // The deferred leakcheck will check whether there's leaked goroutine, which is an indication |
| // whether we closes the skipReset channel to unblock onGoAway/onClose goroutine. |
| } |
| |
| // This test is to test the logic in the createTransport after the health check function returns which |
| // closes the allowedToReset channel(since it has not been closed inside health check func) to unblock |
| // onGoAway/onClose goroutine. |
| func (s) TestHealthCheckWithoutSetConnectivityStateCalled(t *testing.T) { |
| hcEnterChan, hcExitChan, testHealthCheckFuncWrapper := setupHealthCheckWrapper() |
| |
| s, lis, ts, deferFunc, err := setupServer(&svrConfig{ |
| specialWatchFunc: func(s *testHealthServer, in *healthpb.HealthCheckRequest, stream healthgrpc.Health_WatchServer) error { |
| if in.Service != "delay" { |
| return status.Error(codes.FailedPrecondition, |
| "this special Watch function only handles request with service name to be \"delay\"") |
| } |
| // Do nothing to mock a delay of health check response from server side. |
| // This case is to help with the test that covers the condition that setConnectivityState is not |
| // called inside HealthCheckFunc before the func returns. |
| select { |
| case <-stream.Context().Done(): |
| case <-time.After(5 * time.Second): |
| } |
| return nil |
| }, |
| }) |
| defer deferFunc() |
| if err != nil { |
| t.Fatal(err) |
| } |
| |
| ts.SetServingStatus("delay", healthpb.HealthCheckResponse_SERVING) |
| |
| _, r, deferFunc, err := setupClient(&clientConfig{ |
| balancerName: "round_robin", |
| testHealthCheckFuncWrapper: testHealthCheckFuncWrapper, |
| }) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| // The serviceName "delay" is specially handled at server side, where response will not be sent |
| // back to client immediately upon receiving the request (client should receive no response until |
| // test ends). |
| r.UpdateState(resolver.State{ |
| Addresses: []resolver.Address{{Addr: lis.Addr().String()}}, |
| ServiceConfig: parseCfg(r, `{ |
| "healthCheckConfig": { |
| "serviceName": "delay" |
| } |
| }`)}) |
| |
| select { |
| case <-hcExitChan: |
| t.Fatal("Health check function has exited, which is not expected.") |
| default: |
| } |
| |
| select { |
| case <-hcEnterChan: |
| case <-time.After(5 * time.Second): |
| t.Fatal("Health check function has not been invoked after 5s.") |
| } |
| // trigger transport being closed |
| s.Stop() |
| |
| // The health check func should exit without calling the setConnectivityState func, as server hasn't sent |
| // any response. |
| select { |
| case <-hcExitChan: |
| case <-time.After(5 * time.Second): |
| t.Fatal("Health check function has not exited after 5s.") |
| } |
| // The deferred leakcheck will check whether there's leaked goroutine, which is an indication |
| // whether we closes the allowedToReset channel to unblock onGoAway/onClose goroutine. |
| } |
| |
| func testHealthCheckDisableWithDialOption(t *testing.T, addr string) { |
| hcEnterChan, _, testHealthCheckFuncWrapper := setupHealthCheckWrapper() |
| |
| cc, r, deferFunc, err := setupClient(&clientConfig{ |
| balancerName: "round_robin", |
| testHealthCheckFuncWrapper: testHealthCheckFuncWrapper, |
| extraDialOption: []grpc.DialOption{grpc.WithDisableHealthCheck()}, |
| }) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| tc := testpb.NewTestServiceClient(cc) |
| |
| r.UpdateState(resolver.State{ |
| Addresses: []resolver.Address{{Addr: addr}}, |
| ServiceConfig: parseCfg(r, `{ |
| "healthCheckConfig": { |
| "serviceName": "foo" |
| } |
| }`)}) |
| |
| // send some rpcs to make sure transport has been created and is ready for use. |
| if err := verifyResultWithDelay(func() (bool, error) { |
| if _, err := tc.EmptyCall(context.Background(), &testpb.Empty{}); err != nil { |
| return false, fmt.Errorf("TestService/EmptyCall(_, _) = _, %v, want _, <nil>", err) |
| } |
| return true, nil |
| }); err != nil { |
| t.Fatal(err) |
| } |
| |
| select { |
| case <-hcEnterChan: |
| t.Fatal("Health check function has exited, which is not expected.") |
| default: |
| } |
| } |
| |
| func testHealthCheckDisableWithBalancer(t *testing.T, addr string) { |
| hcEnterChan, _, testHealthCheckFuncWrapper := setupHealthCheckWrapper() |
| |
| cc, r, deferFunc, err := setupClient(&clientConfig{ |
| balancerName: "pick_first", |
| testHealthCheckFuncWrapper: testHealthCheckFuncWrapper, |
| }) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| tc := testpb.NewTestServiceClient(cc) |
| |
| r.UpdateState(resolver.State{ |
| Addresses: []resolver.Address{{Addr: addr}}, |
| ServiceConfig: parseCfg(r, `{ |
| "healthCheckConfig": { |
| "serviceName": "foo" |
| } |
| }`)}) |
| |
| // send some rpcs to make sure transport has been created and is ready for use. |
| if err := verifyResultWithDelay(func() (bool, error) { |
| if _, err := tc.EmptyCall(context.Background(), &testpb.Empty{}); err != nil { |
| return false, fmt.Errorf("TestService/EmptyCall(_, _) = _, %v, want _, <nil>", err) |
| } |
| return true, nil |
| }); err != nil { |
| t.Fatal(err) |
| } |
| |
| select { |
| case <-hcEnterChan: |
| t.Fatal("Health check function has started, which is not expected.") |
| default: |
| } |
| } |
| |
| func testHealthCheckDisableWithServiceConfig(t *testing.T, addr string) { |
| hcEnterChan, _, testHealthCheckFuncWrapper := setupHealthCheckWrapper() |
| |
| cc, r, deferFunc, err := setupClient(&clientConfig{ |
| balancerName: "round_robin", |
| testHealthCheckFuncWrapper: testHealthCheckFuncWrapper, |
| }) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| tc := testpb.NewTestServiceClient(cc) |
| |
| r.UpdateState(resolver.State{Addresses: []resolver.Address{{Addr: addr}}}) |
| |
| // send some rpcs to make sure transport has been created and is ready for use. |
| if err := verifyResultWithDelay(func() (bool, error) { |
| if _, err := tc.EmptyCall(context.Background(), &testpb.Empty{}); err != nil { |
| return false, fmt.Errorf("TestService/EmptyCall(_, _) = _, %v, want _, <nil>", err) |
| } |
| return true, nil |
| }); err != nil { |
| t.Fatal(err) |
| } |
| |
| select { |
| case <-hcEnterChan: |
| t.Fatal("Health check function has started, which is not expected.") |
| default: |
| } |
| } |
| |
| func (s) TestHealthCheckDisable(t *testing.T) { |
| _, lis, ts, deferFunc, err := setupServer(&svrConfig{}) |
| defer deferFunc() |
| if err != nil { |
| t.Fatal(err) |
| } |
| ts.SetServingStatus("foo", healthpb.HealthCheckResponse_SERVING) |
| |
| // test client side disabling configuration. |
| testHealthCheckDisableWithDialOption(t, lis.Addr().String()) |
| testHealthCheckDisableWithBalancer(t, lis.Addr().String()) |
| testHealthCheckDisableWithServiceConfig(t, lis.Addr().String()) |
| } |
| |
| func (s) TestHealthCheckChannelzCountingCallSuccess(t *testing.T) { |
| _, lis, _, deferFunc, err := setupServer(&svrConfig{ |
| specialWatchFunc: func(s *testHealthServer, in *healthpb.HealthCheckRequest, stream healthgrpc.Health_WatchServer) error { |
| if in.Service != "channelzSuccess" { |
| return status.Error(codes.FailedPrecondition, |
| "this special Watch function only handles request with service name to be \"channelzSuccess\"") |
| } |
| return status.Error(codes.OK, "fake success") |
| }, |
| }) |
| defer deferFunc() |
| if err != nil { |
| t.Fatal(err) |
| } |
| |
| _, r, deferFunc, err := setupClient(&clientConfig{balancerName: "round_robin"}) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| r.UpdateState(resolver.State{ |
| Addresses: []resolver.Address{{Addr: lis.Addr().String()}}, |
| ServiceConfig: parseCfg(r, `{ |
| "healthCheckConfig": { |
| "serviceName": "channelzSuccess" |
| } |
| }`)}) |
| |
| if err := verifyResultWithDelay(func() (bool, error) { |
| cm, _ := channelz.GetTopChannels(0, 0) |
| if len(cm) == 0 { |
| return false, errors.New("channelz.GetTopChannels return 0 top channel") |
| } |
| if len(cm[0].SubChans) == 0 { |
| return false, errors.New("there is 0 subchannel") |
| } |
| var id int64 |
| for k := range cm[0].SubChans { |
| id = k |
| break |
| } |
| scm := channelz.GetSubChannel(id) |
| if scm == nil || scm.ChannelData == nil { |
| return false, errors.New("nil subchannel metric or nil subchannel metric ChannelData returned") |
| } |
| // exponential backoff retry may result in more than one health check call. |
| if scm.ChannelData.CallsStarted > 0 && scm.ChannelData.CallsSucceeded > 0 && scm.ChannelData.CallsFailed == 0 { |
| return true, nil |
| } |
| return false, fmt.Errorf("got %d CallsStarted, %d CallsSucceeded, want >0 >0", scm.ChannelData.CallsStarted, scm.ChannelData.CallsSucceeded) |
| }); err != nil { |
| t.Fatal(err) |
| } |
| } |
| |
| func (s) TestHealthCheckChannelzCountingCallFailure(t *testing.T) { |
| _, lis, _, deferFunc, err := setupServer(&svrConfig{ |
| specialWatchFunc: func(s *testHealthServer, in *healthpb.HealthCheckRequest, stream healthgrpc.Health_WatchServer) error { |
| if in.Service != "channelzFailure" { |
| return status.Error(codes.FailedPrecondition, |
| "this special Watch function only handles request with service name to be \"channelzFailure\"") |
| } |
| return status.Error(codes.Internal, "fake failure") |
| }, |
| }) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| _, r, deferFunc, err := setupClient(&clientConfig{balancerName: "round_robin"}) |
| if err != nil { |
| t.Fatal(err) |
| } |
| defer deferFunc() |
| |
| r.UpdateState(resolver.State{ |
| Addresses: []resolver.Address{{Addr: lis.Addr().String()}}, |
| ServiceConfig: parseCfg(r, `{ |
| "healthCheckConfig": { |
| "serviceName": "channelzFailure" |
| } |
| }`)}) |
| |
| if err := verifyResultWithDelay(func() (bool, error) { |
| cm, _ := channelz.GetTopChannels(0, 0) |
| if len(cm) == 0 { |
| return false, errors.New("channelz.GetTopChannels return 0 top channel") |
| } |
| if len(cm[0].SubChans) == 0 { |
| return false, errors.New("there is 0 subchannel") |
| } |
| var id int64 |
| for k := range cm[0].SubChans { |
| id = k |
| break |
| } |
| scm := channelz.GetSubChannel(id) |
| if scm == nil || scm.ChannelData == nil { |
| return false, errors.New("nil subchannel metric or nil subchannel metric ChannelData returned") |
| } |
| // exponential backoff retry may result in more than one health check call. |
| if scm.ChannelData.CallsStarted > 0 && scm.ChannelData.CallsFailed > 0 && scm.ChannelData.CallsSucceeded == 0 { |
| return true, nil |
| } |
| return false, fmt.Errorf("got %d CallsStarted, %d CallsFailed, want >0, >0", scm.ChannelData.CallsStarted, scm.ChannelData.CallsFailed) |
| }); err != nil { |
| t.Fatal(err) |
| } |
| } |