Skip to content

Update elastic/logs challenges#433

Merged
ebadyano merged 6 commits intoelastic:masterfrom
ebadyano:logs-docs
Jul 18, 2023
Merged

Update elastic/logs challenges#433
ebadyano merged 6 commits intoelastic:masterfrom
ebadyano:logs-docs

Conversation

@ebadyano
Copy link
Copy Markdown
Contributor

A few older many shards challenges got replaced with a new one. Updated README to reflect this. Also adding a few missing challenges and fixing a few typos.

Relates to: #303
Relates to: #294

A few older many shards challenges got replaced with a new one. Updated README to reflect this.
Also adding a few missing challenges and fixing a few typos.

Relates to: elastic#303
Relates to: elastic#294
@ebadyano ebadyano requested review from b-deam and dliappis July 11, 2023 15:08
@ebadyano ebadyano self-assigned this Jul 11, 2023
@ebadyano ebadyano added the docs Anything related to rally-tracks documentation label Jul 11, 2023

This challenge aims to get more specific numbers of what we can support in terms of indices count. It creates initial
set of indices as before and then index to small set of data streams. These data streams will almost never
rollover (rollover based on size with 150gb as `max_size`). This is supposed to be run with multiple values of
Copy link
Copy Markdown
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

both many-shards use 100gb as a rollover value


Was the intent 150gb, should we update the docs or the policy?

Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think we are happy with the 100gb we have now. @original-brownbear WDYT?

Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yea I don't think it makes a difference, we almost index nothing per index :)

Copy link
Copy Markdown
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

sounds good! i will update the docs then just to be consistent :)

Co-authored-by: Brad Deam <54515790+b-deam@users.noreply.github.com>
Copy link
Copy Markdown
Contributor

@dliappis dliappis left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is basically LGTM from my side. Left a suggestion for a more up to date description of many-shards-snapshots.

### Many Shards Snapshots (many-shards-snapshots)

### Many Shards Full (many-shards-full)
This benchmarks aims to evaluate the performance of the Log Monitoring part of Elastic's Observability solution with a large amount of shards. It sets up initial set of indices (count controlled by `data.initial.indices` param) with a large amount of shards, with `auditbeat` template and ILM policy (hot tier only) and then sequentially takes a configurable via `snapshot_counts` number of snapshots. These data streams will almost never rollover (rollover based on size with 100gb as max_size). Used for benchmarks to help identify regressions related to snapshots with high index counts. The performance can be evaluated by the `service_time` of the `wait-for-snapshots` task.
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

When I read "with a large amount of shards" I wrongly thought that the index settings specify a high shard count.

I'd suggest a few clarifications based on the description in https://elasticsearch-benchmarks.elastic.co/#tracks/many-shards-snapshots/nightly/default/90d and https://www.elastic.co/blog/benchmark-driven-optimizations-scalability-elasticsearch-8

This benchmarks aims to evaluate the performance of the Log Monitoring part of Elastic's Observability solution with a high shard count. It sets up initial set of indices (count controlled by data.initial.indices param), using an auditbeat template and ILM policy (hot tier only) and then sequentially takes a configurable (via snapshot_counts) number of snapshots. These data streams will almost never rollover (rollover based on size with 100gb as max_size). This challenge is used by benchmarks to help identify regressions and improvements related to snapshots in use cases with a high shard count. The performance can be evaluated by the service_time of the wait-for-snapshots task.


This challenge aims to get more specific numbers of what we can support in terms of indices count. It creates initial
set of indices as before and then index to small set of data streams. These data streams will almost never
rollover (rollover based on size with 150gb as `max_size`). This is supposed to be run with multiple values of
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think we are happy with the 100gb we have now. @original-brownbear WDYT?

@ebadyano ebadyano merged commit 4c94cc7 into elastic:master Jul 18, 2023
@ebadyano ebadyano deleted the logs-docs branch August 25, 2023 17:08
inqueue pushed a commit to inqueue/rally-tracks that referenced this pull request Dec 6, 2023
A few older many shards challenges got replaced with a new one. Updated README to reflect this.
Also adding a few missing challenges and fixing a few typos.

Relates to: elastic#303
Relates to: elastic#294
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

cleanup docs Anything related to rally-tracks documentation

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants