-
-
Notifications
You must be signed in to change notification settings - Fork 130
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
update batch size in session config #1250
Conversation
WalkthroughThe pull request updates the batch size configuration within the Changes
Possibly related PRs
Suggested reviewers
Poem
✨ Finishing Touches
🪧 TipsChatThere are 3 ways to chat with CodeRabbit:
Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments. CodeRabbit Commands (Invoked using PR comments)
Other keywords and placeholders
CodeRabbit Configuration File (
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 0
🧹 Nitpick comments (1)
src/query/mod.rs (1)
114-114
: Approve batch size reduction for improved performance.The change from 1,000,000 to 8,192 aligns with the PR objective to optimize memory consumption and improve query performance. According to the PR, this has yielded improved clickbench metrics [181, 68, 68].
Consider adding a comment explaining why 8,192 was chosen as the optimal batch size and what performance benefits it provides compared to the previous value:
- .with_batch_size(8192); + // Use 8192 as the batch size (instead of 1000000) to reduce memory consumption + // and improve query performance. Benchmark results: [181, 68, 68] + .with_batch_size(8192);
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (1)
src/query/mod.rs
(1 hunks)
⏰ Context from checks skipped due to timeout of 90000ms (10)
- GitHub Check: Build Default x86_64-unknown-linux-gnu
- GitHub Check: Build Default x86_64-pc-windows-msvc
- GitHub Check: coverage
- GitHub Check: Build Default aarch64-unknown-linux-gnu
- GitHub Check: Build Default aarch64-apple-darwin
- GitHub Check: Build Default x86_64-apple-darwin
- GitHub Check: Build Kafka aarch64-apple-darwin
- GitHub Check: Build Kafka x86_64-unknown-linux-gnu
- GitHub Check: Quest Smoke and Load Tests for Distributed deployments
- GitHub Check: Quest Smoke and Load Tests for Standalone deployments
current: 1000000 update to 8192 (default value) this helps in query performance as larger batch size consumes more memory and slows down performance with this change, parseable's clickbench numbers are [181,68,68]
cb370bb
to
e28a445
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 0
🧹 Nitpick comments (1)
src/cli.rs (1)
273-279
: Looks good, consider enhancing the help textThe new
execution_batch_size
parameter is correctly implemented, following the established patterns in the codebase. Reducing the default batch size from 1,000,000 to 20,000 should improve query performance as indicated in the PR objectives.However, the help text could be more descriptive to guide users on the purpose and impact of this setting. Consider expanding it to explain how this parameter affects memory usage and query performance.
- help = "batch size for query execution" + help = "Batch size for query execution - lower values reduce memory usage but may increase processing overhead, higher values consume more memory but may improve throughput"
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (2)
src/cli.rs
(1 hunks)src/query/mod.rs
(1 hunks)
🚧 Files skipped from review as they are similar to previous changes (1)
- src/query/mod.rs
⏰ Context from checks skipped due to timeout of 90000ms (10)
- GitHub Check: Quest Smoke and Load Tests for Standalone deployments
- GitHub Check: Quest Smoke and Load Tests for Distributed deployments
- GitHub Check: Build Default x86_64-apple-darwin
- GitHub Check: Build Default aarch64-apple-darwin
- GitHub Check: Build Default x86_64-pc-windows-msvc
- GitHub Check: Build Kafka aarch64-apple-darwin
- GitHub Check: Build Default x86_64-unknown-linux-gnu
- GitHub Check: Build Kafka x86_64-unknown-linux-gnu
- GitHub Check: Build Default aarch64-unknown-linux-gnu
- GitHub Check: coverage
current: 1000000
update to 20000
this helps in query performance as larger batch size consumes more memory and slows down performance
with this change, parseable's clickbench numbers are [182,75,75]
Summary by CodeRabbit
20000
.