Slackbot
05/20/2023, 11:07 AMMikel Menta
05/20/2023, 11:08 AMMikel Menta
05/22/2023, 8:47 AMlarme (shenyang)
05/23/2023, 3:19 AMsauyon
05/23/2023, 3:21 AMmax_latency_ms
is specifically only for each individual runner, so the API server can still take arbitrarily long to handle requests.sauyon
05/23/2023, 3:30 AMsauyon
05/23/2023, 3:33 AMMikel Menta
05/24/2023, 10:05 AMcompute
part (which should consider the computation in the Runner + the time in the queue of the Runner).
We have found that:
• In the asyncio.Semaphore(K)
if K is set to 30 or more we get the weird behaviour in the compute
(items waiting as much as the whole benchmark). But if it is below 30 (we tried 10 and 20) it works fine (and still serves around 30RPS). Also note that ~30 is the RPS that the machine can handle.
• When running some tests with locust.io we saw that if we spawn 100 users to perform requests and we spawn them alltogether suddenly this weird situation happens. However if the users are spawned 1/second it still handles them quite fine and the compute
numbers are fine (with just a few oultliers) .
• Not important: Performing a benchmark with a similar script to the python one but written in JavaScript gives the same weird results with K=100.
So maybe it is just a matter of the benchmark because I was doing too many requests (K=100) all of a sudden? Not sure if this could happen in a real situation.sauyon
05/24/2023, 9:09 PMMikel Menta
05/25/2023, 6:46 AM