When running load tests for an application server, the average results are as follows for 1.25 concurrent users.
I calculated 1.25 concurrent users by taking a peak hour of visitors (5 visitors) for a 15 minute duration (900 seconds), then divided by 3600.
Average Result:
Completed Requests: 2683
Requests Per Second: 134
Mean Latency: 14.8ms
This is on a server with 1GB of RAM and 1 vCPU. The CPU usage didn't go above 30%. Memory seemed unaffected.
So what exactly does it mean to have a result of 134 Requests Per Second? I'm trying to determine the server requirements that I'll require to handle X users and how this will scale, but I'm having trouble finding information on the correlation between concurrent users and request per seconds in terms of scalability.
Anything to help me further understand how to analyse these results would be very helpful, thank you.