504 errors on worker

Hi guys, I have a temporal cluster (version 1.21.6) running on GKE and the workers are flooded with these errors

[2m2024-10-29T09:46:58.066017Ze[0m e[31mERRORe[0m e[2mtemporal_client::retrye[0me[2m:e[0m gRPC call poll_workflow_task_queue retried 361 times e[3merrore[0me[2m=e[0mStatus { code: Internal, message: "protocol error: received message with invalid compression flag: 60 (valid flags are 0 and 1) while receiving response with status: 504 Gateway Timeout", metadata: MetadataMap { headers: {"date": "Tue, 29 Oct 2024 09:46:58 GMT", "content-type": "text/html", "content-length": "160", "strict-transport-security": "max-age=15724800; includeSubDomains"} }, source: None }

I bump the frontend ingress controller timeouts to 3600 seconds and also bump the backend service on gcp to 3600 seconds, but I’m still getting the error.

I don’t have any errors on the workflows and the frontend pod doesn’t have any errors on the logs either, just the amount of logs on the worker it’s huge.

I know this has been asked couple years back but couldn’t really find if there’s a solution to this.
Thanks.

Would check your proxy as in using http1instead of http2. That was reason in cases where have seen this error in past.

Hi @tihomir , thanks for answering, I’ve checked all our GCP backends and k8s ingress and we do have configured protocol HTTP2, we are still seeing the error.

I’m having the same issue. Were you able to resolve it?