rpk cluster self-test status

Queries the status of the currently running or last completed self-test run. Returns the status of the currently running or last completed self-test run.

Use this command after invoking rpk cluster self-test start to determine the status of the jobs launched. Possible results are:

  • One or more jobs still running

    • Returns the IDs of Redpanda nodes still running self-tests.

  • No jobs running

    • Returns cached results for all nodes of the last completed test.

Usage

rpk cluster self-test status [flags]

Flags

Value Type Description

--format

string

Output format. Possible values: json, yaml, text, wide, help. Default: text.

-h, --help

-

Help for status.

--config

string

Redpanda or rpk config file; default search paths are /var/lib/redpanda/.config/rpk/rpk.yaml, $PWD/redpanda.yaml, and /etc/redpanda/redpanda.yaml.

-X, --config-opt

stringArray

Override rpk configuration settings. See rpk -X or execute rpk -X help for inline detail or rpk -X list for terser detail.

--profile

string

Profile to use. See rpk profile for more details.

-v, --verbose

-

Enable verbose logging.

Example output

Test results are grouped by node ID. Each test returns the following:

  • NAME: Description of the test.

  • INFO: Detail about the test run attached by Redpanda itself.

  • TYPE: Either disk or network test.

  • TEST ID: Unique identifier given to jobs of a run. All IDs in a test should match. If they don’t match, then newer and/or older test results have been included erroneously.

  • TIMEOUTS: Number of timeouts incurred during the test.

  • DURATION: Duration of the test.

  • IOPS: Number of operations per second. For disk, it’s seastar::dma_read and seastar::dma_write. For network, it’s rpc.send()

  • THROUGHPUT: For disk, it’s throughput rate in bytes per second. For network, it’s throughput rate in bits per second in. (Note: GiB vs. Gib is the correct notation displayed by the UI.)

  • LATENCY: 50th, 90th, etc. percentiles of operation latency, reported in microseconds.

$ rpk cluster self-test status
NODE ID: 1 | STATUS: IDLE
=========================
NAME        512K sequential r/w throughput disk test
INFO        write run
TYPE        disk
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5001ms
IOPS        1590 req/sec
THROUGHPUT  795.2MiB/sec
LATENCY     P50    P90     P99      P999     MAX
            831us  5887us  11263us  24575us  507903us

NAME        512K sequential r/w throughput disk test
INFO        read run
TYPE        disk
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5001ms
IOPS        4504 req/sec
THROUGHPUT  2.2GiB/sec
LATENCY     P50    P90     P99     P999    MAX
            703us  1599us  4351us  6399us  10239us

NAME        4k sequential r/w latency/iops disk test
INFO        write run
TYPE        disk
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5031ms
IOPS        289 req/sec
THROUGHPUT  144.7MiB/sec
LATENCY     P50    P90      P99      P999     MAX
            543us  34815us  69631us  77823us  77823us

NAME        4k sequential r/w latency/iops disk test
INFO        read run
TYPE        disk
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5000ms
IOPS        8275 req/sec
THROUGHPUT  4.041GiB/sec
LATENCY     P50    P90    P99    P999    MAX
            191us  447us  831us  2175us  278527us

NAME        8K Network Throughput Test
INFO        Test performed against node: 0
TYPE        network
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5000ms
IOPS        61254 req/sec
THROUGHPUT  3.74Gib/sec
LATENCY     P50    P90    P99    P999   MAX
            159us  207us  303us  415us  1087us

NAME        8K Network Throughput Test
INFO        Test performed against node: 2
TYPE        network
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5000ms
IOPS        54814 req/sec
THROUGHPUT  3.35Gib/sec
LATENCY     P50    P90    P99    P999   MAX
            167us  255us  367us  511us  25599us

NODE ID: 0 | STATUS: IDLE
=========================
NAME        512K sequential r/w throughput disk test
INFO        write run
TYPE        disk
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5002ms
IOPS        1593 req/sec
THROUGHPUT  796.8MiB/sec
LATENCY     P50    P90     P99      P999     MAX
            735us  5887us  11263us  69631us  507903us

NAME        512K sequential r/w throughput disk test
INFO        read run
TYPE        disk
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5000ms
IOPS        4372 req/sec
THROUGHPUT  2.135GiB/sec
LATENCY     P50    P90     P99     P999    MAX
            735us  1599us  4351us  7423us  9215us

NAME        4k sequential r/w latency/iops disk test
INFO        write run
TYPE        disk
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5026ms
IOPS        286 req/sec
THROUGHPUT  143.1MiB/sec
LATENCY     P50    P90      P99      P999     MAX
            543us  34815us  69631us  77823us  77823us

NAME        4k sequential r/w latency/iops disk test
INFO        read run
TYPE        disk
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5000ms
IOPS        8269 req/sec
THROUGHPUT  4.038GiB/sec
LATENCY     P50    P90    P99    P999    MAX
            191us  447us  831us  2175us  278527us

NAME        8K Network Throughput Test
INFO        Test performed against node: 1
TYPE        network
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5000ms
IOPS        61612 req/sec
THROUGHPUT  3.76Gib/sec
LATENCY     P50    P90    P99    P999   MAX
            159us  207us  303us  431us  1151us

NAME        8K Network Throughput Test
INFO        Test performed against node: 2
TYPE        network
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5000ms
IOPS        60306 req/sec
THROUGHPUT  3.68Gib/sec
LATENCY     P50    P90    P99    P999   MAX
            159us  215us  351us  495us  11263us

NODE ID: 2 | STATUS: IDLE
=========================
NAME        512K sequential r/w throughput disk test
INFO        write run
TYPE        disk
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5001ms
IOPS        1580 req/sec
THROUGHPUT  790MiB/sec
LATENCY     P50    P90     P99      P999     MAX
            671us  5887us  12287us  47103us  507903us

NAME        512K sequential r/w throughput disk test
INFO        read run
TYPE        disk
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5000ms
IOPS        3932 req/sec
THROUGHPUT  1.92GiB/sec
LATENCY     P50    P90     P99     P999    MAX
            831us  1791us  4351us  7167us  9215us

NAME        4k sequential r/w latency/iops disk test
INFO        write run
TYPE        disk
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5027ms
IOPS        280 req/sec
THROUGHPUT  140.1MiB/sec
LATENCY     P50    P90      P99      P999     MAX
            575us  34815us  73727us  86015us  86015us

NAME        4k sequential r/w latency/iops disk test
INFO        read run
TYPE        disk
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5000ms
IOPS        8699 req/sec
THROUGHPUT  4.248GiB/sec
LATENCY     P50    P90    P99    P999    MAX
            183us  367us  831us  2175us  278527us

NAME        8K Network Throughput Test
INFO        Test performed against node: 0
TYPE        network
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5000ms
IOPS        60027 req/sec
THROUGHPUT  3.66Gib/sec
LATENCY     P50    P90    P99    P999   MAX
            159us  223us  351us  511us  11775us

NAME        8K Network Throughput Test
INFO        Test performed against node: 1
TYPE        network
TEST ID     5e4052f3-b828-4c0d-8fd0-b34ff0b6c35d
TIMEOUTS    0
DURATION    5000ms
IOPS        63090 req/sec
THROUGHPUT  3.85Gib/sec
LATENCY     P50    P90    P99    P999   MAX
            151us  207us  319us  463us  17407us
If self-test returns write results that are unexpectedly and significantly lower than read results, it may be because the Redpanda rpk client hardcodes the DSync option to true. When DSync is enabled, files are opened with the O_DSYNC flag set, and this represents the actual setting that Redpanda uses when it writes to disk.