树莓派 redis 压力测试 QPS 和 TPS

树莓派 redis 压力测试 QPS 和 TPS
文章图片

使用一个树莓派 4B 4GB 版本
使用 redis-benchmark 作为压力测试工具
用法:直接在中终端输入该命令

redis-benchmark

部分结果:
╭─ubuntu@RaspberryPi ~ ╰─?redis-benchmark ====== PING_INLINE ====== 100000 requests completed in 6.72 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 93.56% <= 2 milliseconds 99.98% <= 3 milliseconds 100.00% <= 3 milliseconds 14872.10 requests per second====== PING_BULK ====== 100000 requests completed in 6.82 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 93.04% <= 2 milliseconds 99.99% <= 3 milliseconds 100.00% <= 4 milliseconds 14652.01 requests per second====== SET ====== 100000 requests completed in 6.99 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 88.83% <= 2 milliseconds 99.75% <= 3 milliseconds 99.85% <= 4 milliseconds 99.95% <= 5 milliseconds 99.98% <= 6 milliseconds 99.99% <= 7 milliseconds 100.00% <= 7 milliseconds 14300.01 requests per second====== GET ====== 100000 requests completed in 6.75 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 93.66% <= 2 milliseconds 100.00% <= 2 milliseconds 14806.04 requests per second====== INCR ====== 100000 requests completed in 6.73 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 94.82% <= 2 milliseconds 99.98% <= 3 milliseconds 100.00% <= 3 milliseconds 14858.84 requests per second====== LPUSH ====== 100000 requests completed in 6.70 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 94.51% <= 2 milliseconds 100.00% <= 3 milliseconds 100.00% <= 3 milliseconds 14923.15 requests per second====== RPUSH ====== 100000 requests completed in 6.71 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 94.91% <= 2 milliseconds 99.97% <= 3 milliseconds 100.00% <= 3 milliseconds 14909.80 requests per second====== LPOP ====== 100000 requests completed in 6.69 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 95.04% <= 2 milliseconds 100.00% <= 2 milliseconds 14938.75 requests per second====== RPOP ====== 100000 requests completed in 6.71 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 94.07% <= 2 milliseconds 99.94% <= 3 milliseconds 99.97% <= 4 milliseconds 99.98% <= 5 milliseconds 99.99% <= 6 milliseconds 100.00% <= 7 milliseconds 14903.13 requests per second====== SADD ====== 100000 requests completed in 6.70 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 95.02% <= 2 milliseconds 100.00% <= 3 milliseconds 100.00% <= 3 milliseconds 14914.24 requests per second====== HSET ====== 100000 requests completed in 6.73 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 92.32% <= 2 milliseconds 99.97% <= 5 milliseconds 100.00% <= 6 milliseconds 14865.47 requests per second====== SPOP ====== 100000 requests completed in 6.73 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 95.41% <= 2 milliseconds 99.98% <= 3 milliseconds 100.00% <= 3 milliseconds 14861.05 requests per second====== LPUSH (needed to benchmark LRANGE) ====== 100000 requests completed in 6.66 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 96.09% <= 2 milliseconds 99.98% <= 3 milliseconds 100.00% <= 3 milliseconds 15008.25 requests per second====== LRANGE_100 (first 100 elements) ====== 100000 requests completed in 8.77 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 0.01% <= 2 milliseconds 99.96% <= 3 milliseconds 99.99% <= 4 milliseconds 100.00% <= 4 milliseconds 11401.21 requests per second====== LRANGE_300 (first 300 elements) ====== 100000 requests completed in 15.59 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 0.00% <= 2 milliseconds 0.01% <= 3 milliseconds 76.04% <= 4 milliseconds 99.65% <= 5 milliseconds 99.92% <= 6 milliseconds 99.96% <= 7 milliseconds 99.99% <= 8 milliseconds 100.00% <= 9 milliseconds 6413.55 requests per second====== LRANGE_500 (first 450 elements) ====== 100000 requests completed in 19.95 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 2 milliseconds 0.01% <= 3 milliseconds 0.01% <= 4 milliseconds 62.01% <= 5 milliseconds 99.31% <= 6 milliseconds 99.80% <= 7 milliseconds 99.98% <= 8 milliseconds 99.99% <= 9 milliseconds 100.00% <= 10 milliseconds 100.00% <= 12 milliseconds 100.00% <= 12 milliseconds 5011.53 requests per second====== LRANGE_600 (first 600 elements) ====== 100000 requests completed in 24.36 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 2 milliseconds 0.00% <= 3 milliseconds 0.01% <= 4 milliseconds 0.02% <= 5 milliseconds 38.54% <= 6 milliseconds 99.88% <= 7 milliseconds 99.96% <= 8 milliseconds 99.97% <= 9 milliseconds 99.98% <= 10 milliseconds 99.99% <= 11 milliseconds 100.00% <= 12 milliseconds 100.00% <= 13 milliseconds 4104.92 requests per second====== MSET (10 keys) ====== 100000 requests completed in 6.67 seconds 50 parallel clients 3 bytes payload keep alive: 10.00% <= 1 milliseconds 95.90% <= 2 milliseconds 99.97% <= 3 milliseconds 100.00% <= 3 milliseconds 14994.75 requests per second╭─ubuntu@RaspberryPi ~ ╰─?

【树莓派 redis 压力测试 QPS 和 TPS】可以看到 QPSTPS 在 4000-16000之间
相比之下,使用 5700g 测试,QPS 和 TPS 提高了 10-20 倍,轻松上 10w
╭─bot@amd-5700G ~/Desktop/python ╰─?redis-benchmark ====== PING_INLINE ====== 100000 requests completed in 0.57 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.039 milliseconds (cumulative count 2) 50.000% <= 0.135 milliseconds (cumulative count 59347) 75.000% <= 0.151 milliseconds (cumulative count 76345) 87.500% <= 0.175 milliseconds (cumulative count 88058) 93.750% <= 0.199 milliseconds (cumulative count 94218) 96.875% <= 0.239 milliseconds (cumulative count 97089) 98.438% <= 0.279 milliseconds (cumulative count 98615) 99.219% <= 0.359 milliseconds (cumulative count 99223) 99.609% <= 0.575 milliseconds (cumulative count 99614) 99.805% <= 0.735 milliseconds (cumulative count 99821) 99.902% <= 0.791 milliseconds (cumulative count 99908) 99.951% <= 0.903 milliseconds (cumulative count 99953) 99.976% <= 0.943 milliseconds (cumulative count 99978) 99.988% <= 0.959 milliseconds (cumulative count 99990) 99.994% <= 0.967 milliseconds (cumulative count 100000) 100.000% <= 0.967 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 0.081% <= 0.103 milliseconds (cumulative count 81) 95.131% <= 0.207 milliseconds (cumulative count 95131) 98.976% <= 0.303 milliseconds (cumulative count 98976) 99.331% <= 0.407 milliseconds (cumulative count 99331) 99.446% <= 0.503 milliseconds (cumulative count 99446) 99.684% <= 0.607 milliseconds (cumulative count 99684) 99.767% <= 0.703 milliseconds (cumulative count 99767) 99.913% <= 0.807 milliseconds (cumulative count 99913) 99.953% <= 0.903 milliseconds (cumulative count 99953) 100.000% <= 1.007 milliseconds (cumulative count 100000)Summary: throughput summary: 176366.86 requests per second latency summary (msec): avgminp50p95p99max 0.1450.0320.1350.2070.3110.967 ====== PING_MBULK ====== 100000 requests completed in 0.58 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.063 milliseconds (cumulative count 1) 50.000% <= 0.143 milliseconds (cumulative count 70232) 75.000% <= 0.151 milliseconds (cumulative count 75071) 87.500% <= 0.183 milliseconds (cumulative count 90255) 93.750% <= 0.199 milliseconds (cumulative count 94609) 96.875% <= 0.231 milliseconds (cumulative count 97069) 98.438% <= 0.287 milliseconds (cumulative count 98528) 99.219% <= 0.383 milliseconds (cumulative count 99220) 99.609% <= 0.535 milliseconds (cumulative count 99629) 99.805% <= 0.687 milliseconds (cumulative count 99806) 99.902% <= 0.767 milliseconds (cumulative count 99908) 99.951% <= 0.879 milliseconds (cumulative count 99960) 99.976% <= 0.895 milliseconds (cumulative count 99977) 99.988% <= 1.143 milliseconds (cumulative count 99989) 99.994% <= 1.183 milliseconds (cumulative count 99994) 99.997% <= 1.199 milliseconds (cumulative count 99997) 99.998% <= 1.215 milliseconds (cumulative count 100000) 100.000% <= 1.215 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 0.030% <= 0.103 milliseconds (cumulative count 30) 95.513% <= 0.207 milliseconds (cumulative count 95513) 98.747% <= 0.303 milliseconds (cumulative count 98747) 99.303% <= 0.407 milliseconds (cumulative count 99303) 99.534% <= 0.503 milliseconds (cumulative count 99534) 99.749% <= 0.607 milliseconds (cumulative count 99749) 99.820% <= 0.703 milliseconds (cumulative count 99820) 99.924% <= 0.807 milliseconds (cumulative count 99924) 99.980% <= 0.903 milliseconds (cumulative count 99980) 99.983% <= 1.103 milliseconds (cumulative count 99983) 99.998% <= 1.207 milliseconds (cumulative count 99998) 100.000% <= 1.303 milliseconds (cumulative count 100000)Summary: throughput summary: 172117.05 requests per second latency summary (msec): avgminp50p95p99max 0.1480.0560.1430.2070.3351.215 ====== SET ====== 100000 requests completed in 0.52 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.063 milliseconds (cumulative count 1) 50.000% <= 0.127 milliseconds (cumulative count 52318) 75.000% <= 0.143 milliseconds (cumulative count 76574) 87.500% <= 0.167 milliseconds (cumulative count 87660) 93.750% <= 0.191 milliseconds (cumulative count 94698) 96.875% <= 0.215 milliseconds (cumulative count 96877) 98.438% <= 0.303 milliseconds (cumulative count 98453) 99.219% <= 0.471 milliseconds (cumulative count 99233) 99.609% <= 0.591 milliseconds (cumulative count 99621) 99.805% <= 0.687 milliseconds (cumulative count 99807) 99.902% <= 0.847 milliseconds (cumulative count 99903) 99.951% <= 0.935 milliseconds (cumulative count 99957) 99.976% <= 1.071 milliseconds (cumulative count 99978) 99.988% <= 1.095 milliseconds (cumulative count 99991) 99.994% <= 1.119 milliseconds (cumulative count 99994) 99.997% <= 1.127 milliseconds (cumulative count 99997) 99.998% <= 1.135 milliseconds (cumulative count 100000) 100.000% <= 1.135 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 8.896% <= 0.103 milliseconds (cumulative count 8896) 96.387% <= 0.207 milliseconds (cumulative count 96387) 98.453% <= 0.303 milliseconds (cumulative count 98453) 98.953% <= 0.407 milliseconds (cumulative count 98953) 99.331% <= 0.503 milliseconds (cumulative count 99331) 99.674% <= 0.607 milliseconds (cumulative count 99674) 99.825% <= 0.703 milliseconds (cumulative count 99825) 99.899% <= 0.807 milliseconds (cumulative count 99899) 99.923% <= 0.903 milliseconds (cumulative count 99923) 99.971% <= 1.007 milliseconds (cumulative count 99971) 99.991% <= 1.103 milliseconds (cumulative count 99991) 100.000% <= 1.207 milliseconds (cumulative count 100000)Summary: throughput summary: 191204.59 requests per second latency summary (msec): avgminp50p95p99max 0.1380.0560.1270.1990.4231.135 ====== GET ====== 100000 requests completed in 0.55 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.063 milliseconds (cumulative count 2) 50.000% <= 0.135 milliseconds (cumulative count 67996) 75.000% <= 0.143 milliseconds (cumulative count 75770) 87.500% <= 0.175 milliseconds (cumulative count 89447) 93.750% <= 0.199 milliseconds (cumulative count 94491) 96.875% <= 0.239 milliseconds (cumulative count 96943) 98.438% <= 0.295 milliseconds (cumulative count 98514) 99.219% <= 0.415 milliseconds (cumulative count 99248) 99.609% <= 0.615 milliseconds (cumulative count 99611) 99.805% <= 0.703 milliseconds (cumulative count 99812) 99.902% <= 0.807 milliseconds (cumulative count 99903) 99.951% <= 0.887 milliseconds (cumulative count 99956) 99.976% <= 0.951 milliseconds (cumulative count 99976) 99.988% <= 1.039 milliseconds (cumulative count 99988) 99.994% <= 1.071 milliseconds (cumulative count 99995) 99.997% <= 1.079 milliseconds (cumulative count 99998) 99.998% <= 1.087 milliseconds (cumulative count 100000) 100.000% <= 1.087 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 1.716% <= 0.103 milliseconds (cumulative count 1716) 95.269% <= 0.207 milliseconds (cumulative count 95269) 98.623% <= 0.303 milliseconds (cumulative count 98623) 99.215% <= 0.407 milliseconds (cumulative count 99215) 99.422% <= 0.503 milliseconds (cumulative count 99422) 99.596% <= 0.607 milliseconds (cumulative count 99596) 99.812% <= 0.703 milliseconds (cumulative count 99812) 99.903% <= 0.807 milliseconds (cumulative count 99903) 99.960% <= 0.903 milliseconds (cumulative count 99960) 99.982% <= 1.007 milliseconds (cumulative count 99982) 100.000% <= 1.103 milliseconds (cumulative count 100000)Summary: throughput summary: 183150.19 requests per second latency summary (msec): avgminp50p95p99max 0.1410.0560.1350.2070.3591.087 ====== INCR ====== 100000 requests completed in 0.58 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.047 milliseconds (cumulative count 2) 50.000% <= 0.143 milliseconds (cumulative count 68051) 75.000% <= 0.159 milliseconds (cumulative count 77162) 87.500% <= 0.183 milliseconds (cumulative count 88919) 93.750% <= 0.207 milliseconds (cumulative count 94423) 96.875% <= 0.247 milliseconds (cumulative count 97073) 98.438% <= 0.287 milliseconds (cumulative count 98547) 99.219% <= 0.335 milliseconds (cumulative count 99230) 99.609% <= 0.559 milliseconds (cumulative count 99621) 99.805% <= 0.655 milliseconds (cumulative count 99814) 99.902% <= 0.759 milliseconds (cumulative count 99903) 99.951% <= 0.823 milliseconds (cumulative count 99960) 99.976% <= 0.879 milliseconds (cumulative count 99976) 99.988% <= 0.927 milliseconds (cumulative count 99992) 99.994% <= 0.935 milliseconds (cumulative count 100000) 100.000% <= 0.935 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 0.166% <= 0.103 milliseconds (cumulative count 166) 94.423% <= 0.207 milliseconds (cumulative count 94423) 98.936% <= 0.303 milliseconds (cumulative count 98936) 99.387% <= 0.407 milliseconds (cumulative count 99387) 99.519% <= 0.503 milliseconds (cumulative count 99519) 99.699% <= 0.607 milliseconds (cumulative count 99699) 99.879% <= 0.703 milliseconds (cumulative count 99879) 99.942% <= 0.807 milliseconds (cumulative count 99942) 99.985% <= 0.903 milliseconds (cumulative count 99985) 100.000% <= 1.007 milliseconds (cumulative count 100000)Summary: throughput summary: 172413.80 requests per second latency summary (msec): avgminp50p95p99max 0.1490.0400.1430.2150.3110.935 ====== LPUSH ====== 100000 requests completed in 0.55 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.039 milliseconds (cumulative count 2) 50.000% <= 0.135 milliseconds (cumulative count 60644) 75.000% <= 0.151 milliseconds (cumulative count 77368) 87.500% <= 0.175 milliseconds (cumulative count 87986) 93.750% <= 0.199 milliseconds (cumulative count 94060) 96.875% <= 0.231 milliseconds (cumulative count 97053) 98.438% <= 0.271 milliseconds (cumulative count 98444) 99.219% <= 0.343 milliseconds (cumulative count 99234) 99.609% <= 0.567 milliseconds (cumulative count 99619) 99.805% <= 0.759 milliseconds (cumulative count 99808) 99.902% <= 0.855 milliseconds (cumulative count 99913) 99.951% <= 0.903 milliseconds (cumulative count 99963) 99.976% <= 0.975 milliseconds (cumulative count 99977) 99.988% <= 0.999 milliseconds (cumulative count 99988) 99.994% <= 1.047 milliseconds (cumulative count 99997) 99.998% <= 1.055 milliseconds (cumulative count 100000) 100.000% <= 1.055 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 2.639% <= 0.103 milliseconds (cumulative count 2639) 95.028% <= 0.207 milliseconds (cumulative count 95028) 99.059% <= 0.303 milliseconds (cumulative count 99059) 99.365% <= 0.407 milliseconds (cumulative count 99365) 99.512% <= 0.503 milliseconds (cumulative count 99512) 99.653% <= 0.607 milliseconds (cumulative count 99653) 99.762% <= 0.703 milliseconds (cumulative count 99762) 99.857% <= 0.807 milliseconds (cumulative count 99857) 99.963% <= 0.903 milliseconds (cumulative count 99963) 99.988% <= 1.007 milliseconds (cumulative count 99988) 100.000% <= 1.103 milliseconds (cumulative count 100000)Summary: throughput summary: 182149.36 requests per second latency summary (msec): avgminp50p95p99max 0.1420.0320.1350.2070.3031.055 ====== RPUSH ====== 100000 requests completed in 0.57 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.047 milliseconds (cumulative count 2) 50.000% <= 0.135 milliseconds (cumulative count 66248) 75.000% <= 0.151 milliseconds (cumulative count 75181) 87.500% <= 0.191 milliseconds (cumulative count 87591) 93.750% <= 0.247 milliseconds (cumulative count 94165) 96.875% <= 0.287 milliseconds (cumulative count 96953) 98.438% <= 0.335 milliseconds (cumulative count 98541) 99.219% <= 0.455 milliseconds (cumulative count 99225) 99.609% <= 0.631 milliseconds (cumulative count 99619) 99.805% <= 0.815 milliseconds (cumulative count 99807) 99.902% <= 0.919 milliseconds (cumulative count 99907) 99.951% <= 1.015 milliseconds (cumulative count 99955) 99.976% <= 1.255 milliseconds (cumulative count 99978) 99.988% <= 1.295 milliseconds (cumulative count 99990) 99.994% <= 1.303 milliseconds (cumulative count 99994) 99.997% <= 1.311 milliseconds (cumulative count 100000) 100.000% <= 1.311 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 3.805% <= 0.103 milliseconds (cumulative count 3805) 90.162% <= 0.207 milliseconds (cumulative count 90162) 97.587% <= 0.303 milliseconds (cumulative count 97587) 99.154% <= 0.407 milliseconds (cumulative count 99154) 99.302% <= 0.503 milliseconds (cumulative count 99302) 99.572% <= 0.607 milliseconds (cumulative count 99572) 99.741% <= 0.703 milliseconds (cumulative count 99741) 99.802% <= 0.807 milliseconds (cumulative count 99802) 99.897% <= 0.903 milliseconds (cumulative count 99897) 99.950% <= 1.007 milliseconds (cumulative count 99950) 99.974% <= 1.103 milliseconds (cumulative count 99974) 99.994% <= 1.303 milliseconds (cumulative count 99994) 100.000% <= 1.407 milliseconds (cumulative count 100000)Summary: throughput summary: 175131.36 requests per second latency summary (msec): avgminp50p95p99max 0.1480.0400.1350.2550.3751.311 ====== LPOP ====== 100000 requests completed in 0.57 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.039 milliseconds (cumulative count 1) 50.000% <= 0.135 milliseconds (cumulative count 57478) 75.000% <= 0.159 milliseconds (cumulative count 76183) 87.500% <= 0.183 milliseconds (cumulative count 88761) 93.750% <= 0.215 milliseconds (cumulative count 93885) 96.875% <= 0.255 milliseconds (cumulative count 97277) 98.438% <= 0.279 milliseconds (cumulative count 98447) 99.219% <= 0.327 milliseconds (cumulative count 99246) 99.609% <= 0.479 milliseconds (cumulative count 99615) 99.805% <= 0.623 milliseconds (cumulative count 99805) 99.902% <= 0.735 milliseconds (cumulative count 99904) 99.951% <= 0.791 milliseconds (cumulative count 99953) 99.976% <= 0.847 milliseconds (cumulative count 99977) 99.988% <= 0.879 milliseconds (cumulative count 99990) 99.994% <= 0.887 milliseconds (cumulative count 99995) 99.997% <= 0.895 milliseconds (cumulative count 99997) 99.998% <= 0.903 milliseconds (cumulative count 100000) 100.000% <= 0.903 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 0.412% <= 0.103 milliseconds (cumulative count 412) 93.070% <= 0.207 milliseconds (cumulative count 93070) 99.022% <= 0.303 milliseconds (cumulative count 99022) 99.465% <= 0.407 milliseconds (cumulative count 99465) 99.654% <= 0.503 milliseconds (cumulative count 99654) 99.777% <= 0.607 milliseconds (cumulative count 99777) 99.882% <= 0.703 milliseconds (cumulative count 99882) 99.958% <= 0.807 milliseconds (cumulative count 99958) 100.000% <= 0.903 milliseconds (cumulative count 100000)Summary: throughput summary: 175131.36 requests per second latency summary (msec): avgminp50p95p99max 0.1470.0320.1350.2310.3030.903 ====== RPOP ====== 100000 requests completed in 0.57 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.063 milliseconds (cumulative count 2) 50.000% <= 0.135 milliseconds (cumulative count 57341) 75.000% <= 0.159 milliseconds (cumulative count 78943) 87.500% <= 0.183 milliseconds (cumulative count 89854) 93.750% <= 0.207 milliseconds (cumulative count 94610) 96.875% <= 0.247 milliseconds (cumulative count 97218) 98.438% <= 0.287 milliseconds (cumulative count 98562) 99.219% <= 0.399 milliseconds (cumulative count 99219) 99.609% <= 0.591 milliseconds (cumulative count 99617) 99.805% <= 0.703 milliseconds (cumulative count 99809) 99.902% <= 0.775 milliseconds (cumulative count 99915) 99.951% <= 0.815 milliseconds (cumulative count 99958) 99.976% <= 0.847 milliseconds (cumulative count 99979) 99.988% <= 0.879 milliseconds (cumulative count 99990) 99.994% <= 0.903 milliseconds (cumulative count 99995) 99.997% <= 0.911 milliseconds (cumulative count 99999) 99.999% <= 0.919 milliseconds (cumulative count 100000) 100.000% <= 0.919 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 0.047% <= 0.103 milliseconds (cumulative count 47) 94.610% <= 0.207 milliseconds (cumulative count 94610) 98.819% <= 0.303 milliseconds (cumulative count 98819) 99.229% <= 0.407 milliseconds (cumulative count 99229) 99.332% <= 0.503 milliseconds (cumulative count 99332) 99.672% <= 0.607 milliseconds (cumulative count 99672) 99.809% <= 0.703 milliseconds (cumulative count 99809) 99.948% <= 0.807 milliseconds (cumulative count 99948) 99.995% <= 0.903 milliseconds (cumulative count 99995) 100.000% <= 1.007 milliseconds (cumulative count 100000)Summary: throughput summary: 174520.06 requests per second latency summary (msec): avgminp50p95p99max 0.1470.0560.1350.2150.3270.919 ====== SADD ====== 100000 requests completed in 0.57 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.047 milliseconds (cumulative count 2) 50.000% <= 0.135 milliseconds (cumulative count 57363) 75.000% <= 0.151 milliseconds (cumulative count 75129) 87.500% <= 0.175 milliseconds (cumulative count 88183) 93.750% <= 0.199 milliseconds (cumulative count 94968) 96.875% <= 0.231 milliseconds (cumulative count 97161) 98.438% <= 0.279 milliseconds (cumulative count 98589) 99.219% <= 0.391 milliseconds (cumulative count 99225) 99.609% <= 0.583 milliseconds (cumulative count 99623) 99.805% <= 0.727 milliseconds (cumulative count 99809) 99.902% <= 0.823 milliseconds (cumulative count 99907) 99.951% <= 0.879 milliseconds (cumulative count 99952) 99.976% <= 0.927 milliseconds (cumulative count 99977) 99.988% <= 0.951 milliseconds (cumulative count 99989) 99.994% <= 0.967 milliseconds (cumulative count 99999) 99.999% <= 0.975 milliseconds (cumulative count 100000) 100.000% <= 0.975 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 0.562% <= 0.103 milliseconds (cumulative count 562) 95.818% <= 0.207 milliseconds (cumulative count 95818) 98.939% <= 0.303 milliseconds (cumulative count 98939) 99.249% <= 0.407 milliseconds (cumulative count 99249) 99.387% <= 0.503 milliseconds (cumulative count 99387) 99.674% <= 0.607 milliseconds (cumulative count 99674) 99.780% <= 0.703 milliseconds (cumulative count 99780) 99.889% <= 0.807 milliseconds (cumulative count 99889) 99.964% <= 0.903 milliseconds (cumulative count 99964) 100.000% <= 1.007 milliseconds (cumulative count 100000)Summary: throughput summary: 176056.33 requests per second latency summary (msec): avgminp50p95p99max 0.1460.0400.1350.2070.3110.975 ====== HSET ====== 100000 requests completed in 0.59 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.063 milliseconds (cumulative count 1) 50.000% <= 0.135 milliseconds (cumulative count 52022) 75.000% <= 0.159 milliseconds (cumulative count 76364) 87.500% <= 0.191 milliseconds (cumulative count 89232) 93.750% <= 0.223 milliseconds (cumulative count 94892) 96.875% <= 0.255 milliseconds (cumulative count 97171) 98.438% <= 0.319 milliseconds (cumulative count 98457) 99.219% <= 0.503 milliseconds (cumulative count 99236) 99.609% <= 0.615 milliseconds (cumulative count 99614) 99.805% <= 0.719 milliseconds (cumulative count 99817) 99.902% <= 0.775 milliseconds (cumulative count 99904) 99.951% <= 0.823 milliseconds (cumulative count 99952) 99.976% <= 0.871 milliseconds (cumulative count 99976) 99.988% <= 1.039 milliseconds (cumulative count 99988) 99.994% <= 1.079 milliseconds (cumulative count 99995) 99.997% <= 1.087 milliseconds (cumulative count 99998) 99.998% <= 1.095 milliseconds (cumulative count 100000) 100.000% <= 1.095 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 0.043% <= 0.103 milliseconds (cumulative count 43) 91.823% <= 0.207 milliseconds (cumulative count 91823) 98.290% <= 0.303 milliseconds (cumulative count 98290) 99.037% <= 0.407 milliseconds (cumulative count 99037) 99.236% <= 0.503 milliseconds (cumulative count 99236) 99.604% <= 0.607 milliseconds (cumulative count 99604) 99.786% <= 0.703 milliseconds (cumulative count 99786) 99.937% <= 0.807 milliseconds (cumulative count 99937) 99.979% <= 0.903 milliseconds (cumulative count 99979) 99.980% <= 1.007 milliseconds (cumulative count 99980) 100.000% <= 1.103 milliseconds (cumulative count 100000)Summary: throughput summary: 168918.92 requests per second latency summary (msec): avgminp50p95p99max 0.1520.0560.1350.2310.4071.095 ====== SPOP ====== 100000 requests completed in 0.59 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.055 milliseconds (cumulative count 1) 50.000% <= 0.135 milliseconds (cumulative count 51371) 75.000% <= 0.167 milliseconds (cumulative count 80840) 87.500% <= 0.183 milliseconds (cumulative count 89375) 93.750% <= 0.207 milliseconds (cumulative count 94285) 96.875% <= 0.247 milliseconds (cumulative count 97177) 98.438% <= 0.287 milliseconds (cumulative count 98631) 99.219% <= 0.327 milliseconds (cumulative count 99258) 99.609% <= 0.447 milliseconds (cumulative count 99614) 99.805% <= 0.623 milliseconds (cumulative count 99805) 99.902% <= 0.799 milliseconds (cumulative count 99907) 99.951% <= 0.887 milliseconds (cumulative count 99952) 99.976% <= 0.919 milliseconds (cumulative count 99980) 99.988% <= 0.943 milliseconds (cumulative count 99988) 99.994% <= 0.959 milliseconds (cumulative count 99997) 99.998% <= 0.967 milliseconds (cumulative count 100000) 100.000% <= 0.967 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 0.041% <= 0.103 milliseconds (cumulative count 41) 94.285% <= 0.207 milliseconds (cumulative count 94285) 98.940% <= 0.303 milliseconds (cumulative count 98940) 99.528% <= 0.407 milliseconds (cumulative count 99528) 99.665% <= 0.503 milliseconds (cumulative count 99665) 99.799% <= 0.607 milliseconds (cumulative count 99799) 99.852% <= 0.703 milliseconds (cumulative count 99852) 99.908% <= 0.807 milliseconds (cumulative count 99908) 99.969% <= 0.903 milliseconds (cumulative count 99969) 100.000% <= 1.007 milliseconds (cumulative count 100000)Summary: throughput summary: 169204.73 requests per second latency summary (msec): avgminp50p95p99max 0.1500.0480.1350.2150.3110.967 ====== ZADD ====== 100000 requests completed in 0.56 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.047 milliseconds (cumulative count 4) 50.000% <= 0.135 milliseconds (cumulative count 58400) 75.000% <= 0.151 milliseconds (cumulative count 76612) 87.500% <= 0.183 milliseconds (cumulative count 89978) 93.750% <= 0.207 milliseconds (cumulative count 94675) 96.875% <= 0.239 milliseconds (cumulative count 97046) 98.438% <= 0.279 milliseconds (cumulative count 98695) 99.219% <= 0.311 milliseconds (cumulative count 99264) 99.609% <= 0.535 milliseconds (cumulative count 99612) 99.805% <= 0.719 milliseconds (cumulative count 99812) 99.902% <= 0.791 milliseconds (cumulative count 99916) 99.951% <= 0.807 milliseconds (cumulative count 99953) 99.976% <= 0.887 milliseconds (cumulative count 99977) 99.988% <= 0.943 milliseconds (cumulative count 99989) 99.994% <= 0.967 milliseconds (cumulative count 99997) 99.998% <= 0.975 milliseconds (cumulative count 100000) 100.000% <= 0.975 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 0.118% <= 0.103 milliseconds (cumulative count 118) 94.675% <= 0.207 milliseconds (cumulative count 94675) 99.183% <= 0.303 milliseconds (cumulative count 99183) 99.517% <= 0.407 milliseconds (cumulative count 99517) 99.576% <= 0.503 milliseconds (cumulative count 99576) 99.705% <= 0.607 milliseconds (cumulative count 99705) 99.792% <= 0.703 milliseconds (cumulative count 99792) 99.953% <= 0.807 milliseconds (cumulative count 99953) 99.979% <= 0.903 milliseconds (cumulative count 99979) 100.000% <= 1.007 milliseconds (cumulative count 100000)Summary: throughput summary: 178253.12 requests per second latency summary (msec): avgminp50p95p99max 0.1450.0400.1350.2150.2950.975 ====== ZPOPMIN ====== 100000 requests completed in 0.60 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.079 milliseconds (cumulative count 1) 50.000% <= 0.135 milliseconds (cumulative count 52822) 75.000% <= 0.167 milliseconds (cumulative count 78683) 87.500% <= 0.199 milliseconds (cumulative count 88653) 93.750% <= 0.247 milliseconds (cumulative count 94200) 96.875% <= 0.287 milliseconds (cumulative count 96951) 98.438% <= 0.319 milliseconds (cumulative count 98457) 99.219% <= 0.359 milliseconds (cumulative count 99228) 99.609% <= 0.511 milliseconds (cumulative count 99617) 99.805% <= 0.647 milliseconds (cumulative count 99808) 99.902% <= 0.767 milliseconds (cumulative count 99907) 99.951% <= 0.847 milliseconds (cumulative count 99962) 99.976% <= 0.871 milliseconds (cumulative count 99980) 99.988% <= 0.895 milliseconds (cumulative count 99993) 99.994% <= 0.903 milliseconds (cumulative count 99996) 99.997% <= 0.911 milliseconds (cumulative count 99998) 99.998% <= 0.927 milliseconds (cumulative count 100000) 100.000% <= 0.927 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 0.030% <= 0.103 milliseconds (cumulative count 30) 89.709% <= 0.207 milliseconds (cumulative count 89709) 97.815% <= 0.303 milliseconds (cumulative count 97815) 99.368% <= 0.407 milliseconds (cumulative count 99368) 99.608% <= 0.503 milliseconds (cumulative count 99608) 99.778% <= 0.607 milliseconds (cumulative count 99778) 99.855% <= 0.703 milliseconds (cumulative count 99855) 99.927% <= 0.807 milliseconds (cumulative count 99927) 99.996% <= 0.903 milliseconds (cumulative count 99996) 100.000% <= 1.007 milliseconds (cumulative count 100000)Summary: throughput summary: 165562.92 requests per second latency summary (msec): avgminp50p95p99max 0.1540.0720.1350.2630.3430.927 ====== LPUSH (needed to benchmark LRANGE) ====== 100000 requests completed in 0.52 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.039 milliseconds (cumulative count 2) 50.000% <= 0.127 milliseconds (cumulative count 57330) 75.000% <= 0.143 milliseconds (cumulative count 78341) 87.500% <= 0.167 milliseconds (cumulative count 88472) 93.750% <= 0.191 milliseconds (cumulative count 94001) 96.875% <= 0.239 milliseconds (cumulative count 97076) 98.438% <= 0.279 milliseconds (cumulative count 98472) 99.219% <= 0.407 milliseconds (cumulative count 99223) 99.609% <= 0.591 milliseconds (cumulative count 99630) 99.805% <= 0.687 milliseconds (cumulative count 99809) 99.902% <= 0.831 milliseconds (cumulative count 99908) 99.951% <= 0.887 milliseconds (cumulative count 99953) 99.976% <= 0.927 milliseconds (cumulative count 99978) 99.988% <= 0.967 milliseconds (cumulative count 99988) 99.994% <= 1.007 milliseconds (cumulative count 99996) 99.997% <= 1.015 milliseconds (cumulative count 100000) 100.000% <= 1.015 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 4.325% <= 0.103 milliseconds (cumulative count 4325) 95.503% <= 0.207 milliseconds (cumulative count 95503) 98.807% <= 0.303 milliseconds (cumulative count 98807) 99.223% <= 0.407 milliseconds (cumulative count 99223) 99.422% <= 0.503 milliseconds (cumulative count 99422) 99.662% <= 0.607 milliseconds (cumulative count 99662) 99.815% <= 0.703 milliseconds (cumulative count 99815) 99.892% <= 0.807 milliseconds (cumulative count 99892) 99.962% <= 0.903 milliseconds (cumulative count 99962) 99.996% <= 1.007 milliseconds (cumulative count 99996) 100.000% <= 1.103 milliseconds (cumulative count 100000)Summary: throughput summary: 191938.56 requests per second latency summary (msec): avgminp50p95p99max 0.1360.0320.1270.2070.3431.015 ====== LRANGE_100 (first 100 elements) ====== 100000 requests completed in 0.92 seconds 50 parallel clients 3 bytes payload keep alive: 1 host configuration "save": 3600 1 300 100 60 10000 host configuration "appendonly": no multi-thread: noLatency by percentile distribution: 0.000% <= 0.143 milliseconds (cumulative count 1) 50.000% <= 0.223 milliseconds (cumulative count 51105) 75.000% <= 0.263 milliseconds (cumulative count 77241) 87.500% <= 0.295 milliseconds (cumulative count 87596) 93.750% <= 0.375 milliseconds (cumulative count 94099) 96.875% <= 0.447 milliseconds (cumulative count 96895) 98.438% <= 0.527 milliseconds (cumulative count 98539) 99.219% <= 0.591 milliseconds (cumulative count 99249) 99.609% <= 0.647 milliseconds (cumulative count 99640) 99.805% <= 0.687 milliseconds (cumulative count 99806) 99.902% <= 0.735 milliseconds (cumulative count 99909) 99.951% <= 0.791 milliseconds (cumulative count 99954) 99.976% <= 0.839 milliseconds (cumulative count 99978) 99.988% <= 0.879 milliseconds (cumulative count 99989) 99.994% <= 0.903 milliseconds (cumulative count 99994) 99.997% <= 0.943 milliseconds (cumulative count 99998) 99.998% <= 0.967 milliseconds (cumulative count 99999) 99.999% <= 0.975 milliseconds (cumulative count 100000) 100.000% <= 0.975 milliseconds (cumulative count 100000)Cumulative distribution of latencies: 0.000% <= 0.103 milliseconds (cumulative count 0) 38.497% <= 0.207 milliseconds (cumulative count 38497) 88.558% <= 0.303 milliseconds (cumulative count 88558) 95.506% <= 0.407 milliseconds (cumulative count 95506) 98.180% <= 0.503 milliseconds (cumulative count 98180) 99.394% <= 0.607 milliseconds (cumulative count 99394) 99.848% <= 0.703 milliseconds (cumulative count 99848) 99.962% <= 0.807 milliseconds (cumulative count 99962) 99.994% <= 0.903 milliseconds (cumulative count 99994) 100.000% <= 1.007 milliseconds (cumulative count 100000)Summary: throughput summary: 109051.26 requests per second latency summary (msec): avgminp50p95p99max 0.2470.1360.2230.3990.5670.975 ^CANGE_300 (first 300 elements): rps=40920.0 (overall: 43920.2) avg_msec=0.611 (overall: 0.574) ╭─bot@amd-5700G ~/Desktop/python ╰─?

树莓派 redis 压力测试 QPS 和 TPS
文章图片

    推荐阅读