yep! it's `grpc_server_handling_seconds_
# spicedb
c
yep! it's
grpc_server_handling_seconds_bucket
m
So I generated this histogram and it's doing a simple test - querying for the same permission for the same subject and resource over and over again. My question is - is the latency of the permission service includes the latency for dispatch. For example if pod1 receives a request but dispatches to pod2, is the permission latency from the perspective of pod1 includes the time to dispatch as well?
the green is the permission latency and yellow is the dispatch latency
j
if cache is used, dispatches won't be made
so it can show a lower average amount for the handling for calls
if it is the same resource and subject over and over, its likely reusing the cache until the cache expires
m
I'm a bit confused here, - isn't the cache distributed itself, i.e it's only cached on the pod to which the query hashes to ? In my simple setup I have 2 pods. So if the requests are load balanced - half of the requests goto the pod 1 and half to pod2 and if say pod1 only caches it because it's the only home of the query it will answer the requests directly from the cache, but pod2 will need to do a dispatch to get to pod 1. Is that understanding correct?
j
if you're making the same call
they'll all dispatch to the same pod, ultimately
and if it is cached, the turn around time will be lower if you hit that pod on the first call