You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I've launched a temporal cluster with the linkerd option and have been running into grpc errors. I don't have issues with regards to non-linkerd deployments.
TEMPORAL_ADDRESS is not set, setting it to 10.8.21.62:7233
2023/10/17 22:29:19 Loading config; env=docker,zone=,configDir=config
2023/10/17 22:29:19 Loading config files=[config/docker.yaml]
"level":"info","ts":"2023-10-17T22:29:19.304Z","msg":"Build info.","git-time":"2023-08-14T17:17:11.000Z","git-revision":"c79c00ac8f96c94dffae6e59eba9a279f7ebc656","git-modified":true,"go-arch":"amd64","go-os":"linux","go-version":"go1.20.6","cgo-enabled":false,"server-version":"1.21.5","debug-mode":false,"logging-call-at":"main.go:148"}
{"level":"info","ts":"2023-10-17T22:29:19.304Z","msg":"Dynamic config client is not configured. Using noop client.","logging-call-at":"main.go:168"}
{"level":"warn","ts":"2023-10-17T22:29:19.304Z","msg":"Not using any authorizer and flag `--allow-no-auth` not detected. Future versions will require using the flag `--allow-no-auth` if you do not want to set an authorizer.","logging-call-at":"main.go:178"}
{"level":"info","ts":"2023-10-17T22:29:19.330Z","msg":"Use rpc address 127.0.0.1:7233 for cluster prod.","component":"metadata-initializer","logging-call-at":"fx.go:840"}
{"level":"info","ts":"2023-10-17T22:29:19.330Z","msg":"Service is not requested, skipping initialization.","service":"history","logging-call-at":"fx.go:371"}
{"level":"info","ts":"2023-10-17T22:29:19.330Z","msg":"Service is not requested, skipping initialization.","service":"matching","logging-call-at":"fx.go:421"}
{"level":"info","ts":"2023-10-17T22:29:19.330Z","msg":"Service is not requested, skipping initialization.","service":"frontend","logging-call-at":"fx.go:479"}
{"level":"info","ts":"2023-10-17T22:29:19.330Z","msg":"Service is not requested, skipping initialization.","service":"internal-frontend","logging-call-at":"fx.go:479"}
{"level":"info","ts":"2023-10-17T22:29:19.347Z","msg":"historyClient: ownership caching disabled","service":"worker","logging-call-at":"client.go:82"}
{"level":"info","ts":"2023-10-17T22:29:19.348Z","msg":"PProf not started due to port not set","logging-call-at":"pprof.go:67"}
{"level":"info","ts":"2023-10-17T22:29:19.348Z","msg":"Starting server for services","value":{"worker":{}},"logging-call-at":"server_impl.go:88"}
{"level":"info","ts":"2023-10-17T22:29:19.363Z","msg":"RuntimeMetricsReporter started","service":"worker","logging-call-at":"runtime.go:138"}
{"level":"info","ts":"2023-10-17T22:29:19.363Z","msg":"worker starting","service":"worker","component":"worker","logging-call-at":"service.go:391"}
{"level":"info","ts":"2023-10-17T22:29:19.369Z","msg":"Membership heartbeat upserted successfully","address":"10.8.21.62","port":6939,"hostId":"9cef8afb-6d3c-11ee-93a5-560035998413","logging-call-at":"monitor.go:256"}
{"level":"info","ts":"2023-10-17T22:29:19.371Z","msg":"bootstrap hosts fetched","bootstrap-hostports":"10.8.11.88:6934,10.8.6.148:6933,10.8.21.62:6939,10.8.20.37:6935","logging-call-at":"monitor.go:298"}
{"level":"info","ts":"2023-10-17T22:29:19.377Z","msg":"Current reachable members","component":"service-resolver","service":"matching","addresses":["10.8.20.37:7235"],"logging-call-at":"service_resolver.go:279"}
{"level":"info","ts":"2023-10-17T22:29:19.377Z","msg":"Current reachable members","component":"service-resolver","service":"worker","addresses":["10.8.21.62:7239"],"logging-call-at":"service_resolver.go:279"}
{"level":"info","ts":"2023-10-17T22:29:19.377Z","msg":"Current reachable members","component":"service-resolver","service":"frontend","addresses":["10.8.6.148:7233"],"logging-call-at":"service_resolver.go:279"}
{"level":"info","ts":"2023-10-17T22:29:19.377Z","msg":"Current reachable members","component":"service-resolver","service":"history","addresses":["10.8.11.88:7234"],"logging-call-at":"service_resolver.go:279"}
{"level":"warn","ts":"2023-10-17T22:29:19.380Z","msg":"error creating sdk client","service":"worker","error":"failed reaching server: stream terminated by RST_STREAM with error code: PROTOCOL_ERROR","logging-call-at":"factory.go:114"}
{"level":"fatal","ts":"2023-10-17T22:29:19.380Z","msg":"error creating sdk client","service":"worker","error":"failed reaching server: stream terminated by RST_STREAM with error code: PROTOCOL_ERROR","logging-call-at":"factory.go:121","stacktrace":"go.temporal.io/server/common/log.(*zapLogger).Fatal\n\t/home/builder/temporal/common/log/zap_logger.go:180\ngo.temporal.io/server/common/sdk.(*clientFactory).GetSystemClient.func1\n\t/home/builder/temporal/common/sdk/factory.go:121\nsync.(*Once).doSlow\n\t/usr/local/go/src/sync/once.go:74\nsync.(*Once).Do\n\t/usr/local/go/src/sync/once.go:65\ngo.temporal.io/server/common/sdk.(*clientFactory).GetSystemClient\n\t/home/builder/temporal/common/sdk/factory.go:108\ngo.temporal.io/server/service/worker/scanner.(*Scanner).Start\n\t/home/builder/temporal/service/worker/scanner/scanner.go:229\ngo.temporal.io/server/service/worker.(*Service).startScanner\n\t/home/builder/temporal/service/worker/service.go:523\ngo.temporal.io/server/service/worker.(*Service).Start\n\t/home/builder/temporal/service/worker/service.go:408\ngo.temporal.io/server/service/worker.ServiceLifetimeHooks.func1.1\n\t/home/builder/temporal/service/worker/fx.go:136"}
I've launched a temporal cluster with the linkerd option and have been running into grpc errors. I don't have issues with regards to non-linkerd deployments.
This seems like a GRPC issue.
My temporal cluster:
The text was updated successfully, but these errors were encountered: