Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

full_storage_utilization_test: Storage utilization at 90% cluster size #9018

Draft
wants to merge 1 commit into
base: master
Choose a base branch
from

Conversation

Lakshmipathi
Copy link

Populate data until it reaches over 90% disk storage then perform db and topology change cluster options.

Description

This PR covers basic part of 90% storage utilization, populate data when the cluster reaches over 90% disk usage
performs scaleout operation.Goal of this task is to run the cluster at 90% disk utilization. This will help the users to utilize their instances fully. To help achieve this goal, we need to utilize features like concurrent topology changes, tablets and migrations.

Keeping this as draft version so that it acts as common base between dev and qa team so it helps to improve this PR further.

Testing

  • [ ]

PR pre-checks (self review)

  • I added the relevant backport labels
  • I didn't leave commented-out/debugging code

Reminders

  • Add New configuration option and document them (in sdcm/sct_config.py)
  • Add unit tests to cover my changes (under unit-test/ folder)
  • Update the Readme/doc folder relevant to this change (if needed)

Populate data until it reaches over 90% disk storage and perform
db and cluster options.

Signed-off-by: Lakshmipathi <[email protected]>
@Lakshmipathi Lakshmipathi requested review from lukepio and paszkow and removed request for lukepio October 21, 2024 11:15
self.run_stress(93)

self.log.info("Adding a new node")
self.add_new_node()
Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

One major comment here. Adding a node, will always succeed because you no longer write data. Your implementation of run_stress_until_target to prepare a cluster is fine. Once this is achieved, you want to rather react based on the storage utilization.

I think, this part of adding a node shall be written differently. Namely, you want to monitor the disk usage while writing bunch of data and if a certain threshold is reached, you extend a cluster by adding a new node.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants