Drnbench

Description

Drnbench is a benchmark tool for Droonga.

It may be used for other HTTP servers.

Drnbench provides features to send multiple random requests with different settings periodically. Number of clients (requests) in each period will be automatically increased progressively. So you'll be able to guess the limit performance of the throughput of a server, via the report like following:

n_clients,total_n_requests,queries_per_second,min_elapsed_time,max_elapsed_time,average_elapsed_time,200
1,33,3.3,0.164632187,0.164632187,0.19133309036363635,0
2,70,7.0,0.161510877,0.161510877,0.1846983412285715,0
3,87,8.7,0.1658357,0.1658357,0.24303329366666668,0
...

Drnbench provides two commands, for different styles of benchmarked servers.

  • drnbench-request-response: benchmarks simple request-response style APIs.
  • drnbench-publish-subscribe: benchmarks publish-subscribe style (including HTTP streaming) APIs. This command is not implemented yet.

How to run benchmark?

Benchmarking with an HTTP server

Drnbench can benchmark throughput performance of an HTTP server with random requests.

In this scenario, you have to do:

  • setup an HTTP server.
  • prepare patterns of requests.

Drnbench will start multiple clients and send many requests based on the patterns file.

  1. Create a patterns file "patterns.json", like:

    {
      "user search": {
        "frequency": 0.61,
        "method":    "GET",
        "patterns":  [
          { "path": "/users?q=foo" },
          { "path": "/users?q=bar" },
          ...
        ]
      },
      "item search": {
        "frequency": 0.32,
        "method":    "GET",
        "patterns":  [
          { "path": "/items?q=foo" },
          { "path": "/items?q=bar" },
          ...
        ]
      },
      ...
    }
    
  2. Setup an HTTP server. For example, localhost:80.

  3. Run drnbench with the pattern file.

    # cd ~/drnbench
    # bundle exec bin/drnbench-request-response      --start-n-clients=1      --end-n-clients=32      --step=1      --duration=10      --wait=0.01      --mode=http      --request-patterns-file=/tmp/patterns.json      --host=localhost      --port=80
    
  4. You'll get a report.

Detailed format of the patterns file

Abstract

Abstract type A, grouped patterns with names:

{
  "group name 1": (patterns group),
  "group name 2": (patterns group),
  ...
}

Abstract type B, grouped patterns without names:

[
  (patterns group),
  (patterns group),
  ...
]

Abstract type C, non-grouped patterns:

(patterns group)
Patterns group

Patterns group type A, without options:

[
  (pattern1),
  (pattern2),
  ...
]

Patterns group type B, with options:

{
  "frequency": (appearance ratio in all requests, optional),
  "path":      "(path to the endpoint, optional)",
  "method":    "(HTTP method, optional)",
  "patterns":  [
    (pattern1),
    (pattern2),
    ...
  ]
}
Pattern

Pattern type A, simple path string:

"/path/to/the/API/or/page"

Pattern B, detailed object:

{
  "path":   "/path/to/the/API/or/page",
  "method": "(HTTP method)",
  "body":   (request body sent by POST method)
}

Benchmarking of request-responsne style commands, with a Droonga-based search system

Drnbench can benchmark throughput performance of a Droonga-based search system with random requests.

In this scenario, you have to do:

  • setup a Droonga-based search system.
  • prepare patterns of requests for commands.

Drnbench will start multiple clients and send many requests based on the patterns file.

  1. Create a patterns file in the format:

    {
      "(pattern type 1 name)": {
        "frequency": (appearance ratio in all requests),
        "command":   "(command name)",
        "patterns":  [
          { command parameters 1 },
          { command parameters 2 },
          { command parameters 3 },
          ...
        ]
      }
      "(patterns type 2 name)": {
        ...
      },
      ...
    }
    

    For example, a file "patterns.json" like:

    {
      "user search": {
        "frequency": 0.61,
        "command":   "search",
        "patterns":  [
          {
            "queries": {
              "users": {
                "source":    "User",
                "condition": "age >= 10",
                "sortBy":    { "keys": ["-birthday"], "offset": 0, "limit": 100" },
                "output":    {
                  "elements": [
                    "count",
                    "records"
                  ],
                  "attributes": ["_key", "name", "age", "birhtday"],
                  "offset": 0,
                  "limit":  100
                }
              }
            }
          },
          ...
        ]
      },
      "item search": {
        "frequency": 0.32,
        "command":   "search",
        "patterns":  [
          {
            "queries": {
              "users": {
                "source":    "Item",
                "condition": "visible == true",
                "sortBy":    { "keys": ["title"], "offset": 0, "limit": 100" },
                "output":    {
                  "elements": [
                    "count",
                    "records"
                  ],
                  "attributes": ["title", "price"],
                  "offset": 0,
                  "limit":  100
                }
              }
            }
          },
          ...
        ]
      },
      ...
    }
    
  2. Setup a Droonga Engine server. For example, localhost:23003.

  3. Setup a Protocol Adapter server. For example, localhost:3003.

  4. Run drnbench with the pattern file.

    # cd ~/drnbench
    # bundle exec bin/drnbench-request-response      --start-n-clients=1      --end-n-clients=32      --step=1      --duration=10      --wait=0.01      --mode=http-droonga      --request-patterns-file=/tmp/patterns.json      --host=localhost      --port=3003
    
  5. You'll get a report.

Benchmarking of HTTP streaming APIs, with a Droonga-based search system

Drnbench can benchmark an HTTP streaming API based on a publish-subscribe command.

In this scenario, you have to do:

  • prepare configuration files "fluentd.conf" and "catalog.json" for a Droonga Engine.
  • prepare an express application works as a Droonga Protocol Adapter.
  • prepare pattern files for subscribe request and feeded data.

Drnbench will run benchmark like:

  1. Prepare subscribers.
  2. Send "feed" messages to the Droonga Engine. All subscribers will receive all published messages.
  3. Increase the number of subscribers.
  4. Repeat 2, 3, and 4.

Steps to run:

  1. Create a patterns file for a subscribe request:

    {
      "path":    "/path/to/endpoint",
      "method":  "HTTP method",
      "headers": {
        "X-xxxx": "xxxxx",
      },
      "body":    (sent as the request body)
    }
    

    For example, a file "watch-subscribe.json" like:

    {
      "path": "/droonga-streaming/watch?condition=keyword"
    }
    
  2. Create a patterns file for messages to be feeded:

    {
      "type":    "(message type)",
      "dataset": "(dataset name)",
      "body":    (message body)
    }
    

    For example, a file "watch-feed.json" like:

    {
      "type":    "watch.feed",
      "dataset": "Watch",
      "body":    {
        "targets": {
          "body": "a content including the keyword"
        }
      }
    }
    
  3. Run drnbench with the pattern file.

    # cd ~/drnbench
    # bundle exec bin/drnbench-publish-subscribe      --start-n-subscribers=1000      --n-publishings=1000      --n-steps=10      --timeout=5      --subscribe-request-file=/tmp/watch-subscribe.json      --feed-file=/tmp/watch-feed.json      --protocol-adapter-port=80      --engine-config-path=/tmp/engine/
    
  4. You'll get a report.

License

GPLv3 or later.

Copyright (c) 2013-2014 Droonga Project

See LICENSE.txt for details.