Sending test HTTP traffic with Hyperfoil

This is a brief overview of Hyperfoil, the “microservice-oriented distributed benchmark framework”

hyperfoil logo

I needed to send some requests to my microservice and a colleague suggested to use a tool that I did not know: Hyperfoil. Basically, you can configure a load that will be sent to a set of HTTP endpoints. Something nice about Hyperfoil is that the core of it is concurrent. The tool will not be a bottleneck when sending lots of requests, as could happen with others. Below I describe a short tutorial about how to use it.

The first step is to download the binaries and save then in a known directory. Once there, you can access the CLI:

cd hyperfoil-0.27.1/
./bin/cli.sh
help

The latest command shows all options, but I will just focus on a simple execution of the tool. Next execute

start-local

hyperfoil logo

to start a controller within the CLI. The next step is to upload the configuration file. This is an example that I will explain:

name: my-benchmark
http:
  host: http://localhost:8080
  sharedConnections: 60
phases:
- main:
    constantRate:
      startAfter: rampup
      usersPerSec: 750
      duration: 300s
      forks:
      - createJournal: &createJournal
          weight: 3
          scenario:
          - putJournal:
            - randomInt:
                min: 1
                max: 10
                toVar: mscId
            - httpRequest:
                PUT: /v1/subscribers/${id}/journal
                headers:
                  Content-Type: application/json
                body:
                  fromFile: journal.json
      - createSubs: &createSubs
          weight: 2
          scenario:
          - putSubs:
            - randomInt:
                min: 1
                max: 10
                toVar: mscId
            - httpRequest:
                PUT: /v1/subscribers/${id}
                headers:
                  Content-Type: application/json
                body: |
                  {
                  "site": 0
                  }
      - getSubs: &getSubs
          weight: 1
          scenario:
          - getSubs:
            - randomInt:
                min: 1
                max: 10
                toVar: id
            - httpRequest:
                GET: /v1/subscribers/${id}
- rampup:
    increasingRate:
      initialUsersPerSec: 3
      targetUsersPerSec: 30
      duration: 10s
      forks:
      - createSubs: *createSubs
      - getSubs: *getSubs

At the first level, you have http section where you can configure HTTP options like the protocol type, TLS, timeouts… in my case I set the host where the traffic will be sent and the sharedConnections, the number of connections to open against the host.

http:
  host: http://localhost:8080
  sharedConnections: 60

Then I defined the phases, in my example I have the rampup and main phases.

In the rampup I want to start sending traffic but not abruptly. To achieve it the phase property “increasingRate” can be used to define the number of requests used initially and the number to reach: initialUsersPerSec and targetUsersPerSec, for a duration of 10s. Then you have to define what is called the forks, the sub-phases of traffic. In my rampup configuration I do not define but point to then because the definition happens in the other phase. Note the * to reference the definition.

- rampup:
    increasingRate:
      initialUsersPerSec: 3
      targetUsersPerSec: 30
      duration: 10s
      forks:
      - createSubs: *createSubs
      - getSubs: *getSubs

There are different types of traffic, being: constantRate, increasingRate, decreasingRate, atOnce, always, noop. The names are moreless descriptive of the purpose.

I have used constantRate in my main phase definition:

constantRate:
  startAfter: rampup
  usersPerSec: 750
  duration: 300s
  forks:

This phase starts after the rampup phase. You can link different type of phases using startAfter. The number of request per second is defined by usersPerSec and also the duration of the traffic being sent with duration. Again, we have to set the forks, but this time they will be defined and not referenced.

Let’s take the first fork as example:

- createJournal: &createJournal
    weight: 3
    scenario:
    - putJournal:
      - randomInt:
          min: 1
          max: 10
          toVar: id
      - httpRequest:
          PUT: /v1/subscribers/${id}/journal
          headers:
            Content-Type: application/json
          body:
            fromFile: journal.json

You define a fork with a name and an address (used for reference). Then you can set a weight. The bigger the weight the more requests will be sent. For instance, if you send 100 request per second, and you have two forks with weight 1, each one will receive 1/(1+1) * 100 requests/s. Then you define a scenario, the HTTP request. Interesting here is the use of variables to be included in the request, in my example you have id variable taking values from 1 to 10. This is included in a PUT URI with some headers and a body. The body can be defined directly in the yaml (consider using | to define multiple lines) or be included by a file. This is what I have done since my body is quite large.

Once you have your traffic scenario defined, you have to upload it to the local hyperfoil server. In the CLI:

upload path/myfile.yaml

The next step is to run traffic:

run my-benchmark

where my-benchmark is the name defined in the yaml file.

You will see live information about the traffic being sent execution

When the traffic execution is finished, you can check some quick stats

stats

stats

and also generate a report with graphs

report --destination <dir>

report1 report2

Conclusion

In my opinion it is a performant tool easy to use with enough configuration option to define flexible burst of traffic to test agains some endpoints

Written on March 18, 2025