Skip to content

Commit

Permalink
Expose bulk_size as track parameter (elastic#63)
Browse files Browse the repository at this point in the history
With this commit we expose the track parameter `bulk_size` for the challenge
`index-logs-fixed-daily-volume`. This also eliminates the need for the
challenges `bulk-size-evaluation`, `bulk-size-evaluation-mini` and
`bulk-size-concurrency-evaluation` as they can be simulated using the track
parameters `bulk_size` and `bulk_indexing_clients`.
  • Loading branch information
danielmitterdorfer authored Dec 16, 2019
1 parent 2eb36b7 commit f974d3d
Show file tree
Hide file tree
Showing 5 changed files with 4 additions and 209 deletions.
13 changes: 1 addition & 12 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -47,18 +47,6 @@ You can specify what challenge you want to run with the `--challenge=YOUR_CHALLE

## Available Challenges

### bulk-size-evaluation

This challenge performs bulk-indexing against a single index with varying bulk request sizes, ranging from 125 events/request to 50000 events/request.

The table below shows the track parameters that can be adjusted along with default values:

| Parameter | Explanation | Type | Default Value |
| --------- | ----------- | ---- | ------------- |
| `number_of_replicas` | Number of index replicas | `int` | `0` |
| `shard_count` | Number of primary shards | `int` | `2` |
| `bulk_indexing_clients` | Number of bulk indexing clients/connections | `int` | `16` |

### shard-sizing

This challenge indexes 2 million events into an index consisting of a single shard 25 times. After each group of 2 million events has been inserted, 4 different Kibana dashboard configurations are benchmarked against the index. At this time no indexing takes place. There are two different dashboards being simulated, aggregating across 50% and 90% of the data in the shard.
Expand Down Expand Up @@ -242,6 +230,7 @@ This challenge indexes a fixed (raw) logging volume of logs per day into daily i
| Parameter | Explanation | Type | Default Value |
| ----------------------- | -------------------------------------------------------------------------------------------------------------------------------------- | ----- | ------------- |
| `bulk_indexing_clients` | Number of bulk indexing clients/connections | `int` | `8` |
| `bulk_size` | Number of documents to send per bulk | `int` | `1000` |
| `daily_logging_volume` | The raw logging volume. Supported units are bytes (without any unit), `kb`, `MB` and `GB`). For the value, only integers are allowed. | `str` | `100GB` |
| `number_of_days` | The number of simulated days for which data should be generated. | `int` | `24` |
| `shard_count` | Number of primary shards | `int` | `3` |
Expand Down
182 changes: 0 additions & 182 deletions eventdata/challenges/bulk-size-evaluation.json

This file was deleted.

3 changes: 2 additions & 1 deletion eventdata/challenges/daily-log-volume-index.json
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
{% set p_bulk_indexing_clients = (bulk_indexing_clients | default(8)) %}
{% set p_bulk_size = (bulk_size | default(1000)) %}
{% set p_number_of_days = (number_of_days | default(24)) %}
{% set p_daily_logging_volume = (daily_logging_volume | default("100GB")) %}

Expand Down Expand Up @@ -35,7 +36,7 @@
"param-source": "elasticlogs_bulk",
"index": "elasticlogs-<yyyy>-<mm>-<dd>",
"starting_point": "2018-05-01:00:00:00",
"bulk-size": 1000,
"bulk-size": {{p_bulk_size}},
"daily_logging_volume": "{{p_daily_logging_volume}}",
"number_of_days": {{p_number_of_days}},
"record_raw_event_size": {{p_record_raw_event_size}}
Expand Down
13 changes: 0 additions & 13 deletions eventdata/operations/indexing.json
Original file line number Diff line number Diff line change
Expand Up @@ -5,19 +5,6 @@ max_rolledover_indices: used by the `delete_rolledover_index_pattern` operation.
rolledover_indices_suffix_separator: used by the `delete_rolledover_index_pattern` operation. Defaults to `-`.
#}

{% set comma = joiner() %}
{% for bulk_size in [50000, 20000, 10000, 5000, 2000, 1000, 500, 250, 125] %}
{{comma()}}
{
"name": "index-append-{{bulk_size}}",
"operation-type": "bulk",
"param-source": "elasticlogs_bulk",
"index": "elasticlogs",
"bulk-size": {{bulk_size}},
"record_raw_event_size": {{p_record_raw_event_size}}
}
{% endfor%}
,
{
"name": "index-append-1000-shard-sizing",
"operation-type": "bulk",
Expand Down
2 changes: 1 addition & 1 deletion smoke-test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@ readonly ES_VERSION=${ES_VERSION:-7.3.0}
# * combined-indexing-and-querying (depends on any challenge that has already created elasticlogs-q* indices)
# * elasticlogs-querying (depends on any challenge that has already created elasticlogs-q* indices)

readonly CHALLENGES=(elasticlogs-continuous-index-and-query document_id_evaluation bulk-update shard-sizing frozen-data-generation index-logs-fixed-daily-volume refresh-interval max-indexing-querying index-and-query-logs-fixed-daily-volume shard-size-on-disk bulk-size-evaluation bulk-size-evaluation-mini bulk-size-concurrency-evaluation generate-historic-data large-shard-sizing large-shard-id-type-evaluation elasticlogs-1bn-load)
readonly CHALLENGES=(elasticlogs-continuous-index-and-query document_id_evaluation bulk-update shard-sizing frozen-data-generation index-logs-fixed-daily-volume refresh-interval max-indexing-querying index-and-query-logs-fixed-daily-volume shard-size-on-disk generate-historic-data large-shard-sizing large-shard-id-type-evaluation elasticlogs-1bn-load)

esrally list tracks --track-repository=eventdata

Expand Down

0 comments on commit f974d3d

Please sign in to comment.