While searching for a lightweight tool to stress test my simple Nginx installation hosting a Hugo-generated website, I found Taurus. So far I’ve been using mostly custom in-house developed tools for stress testing, and those tools are not something that I can use at home.
Taurus looks good at first glance and has a simple installation manual, but unfortunately, it did not work correctly on Debian 12 when I tested it. So I wrote a short manual on how to set it up from scratch.
The problem I had with our tools was that for well optimized systems the testing tool tends to use more resources to conduct the test than the target site uses to handle that artificial traffic. It’s mostly because those tools were created to test complex logic, and were not designed with static pages in mind.
So I need a well optimized tool and I need to learn how to create a vast set of organic-looking tests. I could just send a lot simple requests and probably have a good result, but that would be misleading. After successfully testing the system with artificial traffic, it’s easy to believe that the production system will handle this kind of traffic in a real-life scenario. Then life happens and sometimes a traffic level an order of magnitude lower can kill the system, as organic traffic differs from the artificial one. One of the simplest scenarios I have seen is when a mobile device with poor internet connectivity tries to load a page and blocks a port for longer than usual. If multiple clients experience this issue, it can lead to port exhaustion. And a static site like a blog is going to be accesed from smartphone. With wrong Nginx configuration the host’s CPU may idle and at the same time the page is not available. I think it’s good to be able to test for that.
So, I need two things:
- Ability to test for traffic an order of magnitude higher than expected.
- Ability to design a better test scenario that can emulate various properties of organic traffic.
I’m running my preliminary tests in my home lab, equipped with mostly miniPCs, connected through 1Gbps or 2.5Gbps wired LAN.
Right now, my Nginx instance is hosted as a one-core Debian 12 VM on a ProxmoxVE host with Intel i5-1135G7 CPU - mobile CPU released in 2020. Taurus was set up on 4-core Debian 12 VM hosted on ProxmoxVE with Intel i5-13420H - more modern mobile CPU released in 2023. Both hosts are connected via a 2.5Gbps network and use only SSDs.
The basic scenario with 100 concurrent clients led to 80% CPU utilization of my one-core Nginx VM, with close to 450 requests handled per second. Meanwhile the Taurus instance was using 100% of 4 assigned cores, with nothing else running on that PVE host. I was hoping Taurus could do better.
Also, this kind of traffic is extremely artificial, as
- Taurus was not honoring any cache settings and was always asking for everything,
- All tests were performed on the main page,
- Connectivity is perfect,
- Every request is made from the same IP, browser etc.
So two things to improve before I start any useful testing:
- I need to take a closer look at Taurus’s speed. I also want to replicate Taurus to my PVE cluster, to have a little bit more resources.
- I’d like to gain a deeper understanding of Taurus to better emulate organic traffic and create more realistic test scenarios.
Taurus: https://gettaurus.org/