I'm trying to improve network performance between gRPC client and server. The client network speed is 1Gps, the
Assuming my server takes 200ms to respond, and I measure the latency in the client.
Now, if the server processing time goes up, say to 700ms for a response. Where the requests will accumulate? Will they stay in the client network queue, or they will still be sent to the server and will wait in the server queue?
In other words, does a grpc client holds a queue for requests, or every request is always sent - which means that the latency does not depends on the server processing time.
And is there a setting for it in grpc-python?