Skip to content

Conversation

JakeSCahill
Copy link
Contributor

Description

Resolves https://redpandadata.atlassian.net/browse/DOC-330
Resolves https://redpandadata.atlassian.net/browse/DOC-1147

Page previews

Checks

  • New feature
  • Content gap
  • Support Follow-up
  • Small fix (typos, links, copyedits, etc)

@JakeSCahill JakeSCahill requested a review from a team as a code owner August 26, 2025 14:50
Copy link
Contributor

coderabbitai bot commented Aug 26, 2025

📝 Walkthrough

Walkthrough

Documentation updates in modules/deploy/partials/requirements.adoc and modules/deploy/partials/self-test.adoc add explicit production hardware requirements: NVMe drives and minimum 16,000 IOPS. Requirements reiterate storage constraints (including NFS unsupported for data/cache), add Kubernetes storage guidance, and adjust Kubernetes memory guidance to 2.22 GiB per core with a new explanation of partition-replica memory allocation parameters. The self-test doc adds a production IOPS threshold note. Some formatting updates and references to disk/network benchmarks and Kubernetes storage pages are included. No code or API changes.

Estimated code review effort

🎯 2 (Simple) | ⏱️ ~10 minutes

Assessment against linked issues

Objective Addressed Explanation
Make NVMe a requirement in hardware requirements [DOC-330, DOC-1147]
Specify minimum 16,000 IOPS for production DOC-1147

Assessment against linked issues: Out-of-scope changes

Code Change Explanation
Increased Kubernetes pod memory per core to 2.22 GiB (modules/deploy/partials/requirements.adoc, lines n/a) Not requested by DOC-330 or DOC-1147; scope limited to NVMe and IOPS.
New explanation of topic_partitions_memory_allocation_percent and topic_memory_per_partition (modules/deploy/partials/requirements.adoc, lines n/a) Configuration/memory model details are outside the NVMe/IOPS objectives.
Added NFS unsupported caution for data directory and Tiered Storage cache (modules/deploy/partials/requirements.adoc, lines n/a) Storage backend support note not specified in the linked issues.
Kubernetes storage guidance and PV usage notes (modules/deploy/partials/requirements.adoc, lines n/a) Additional guidance not requested by the linked issues.

Possibly related PRs

Suggested reviewers

  • david-yu
  • mattschumpert
  • travisdowns

Tip

🔌 Remote MCP (Model Context Protocol) integration is now available!

Pro plan users can now connect to remote MCP servers from the Integrations page. Connect with popular remote MCPs such as Notion and Linear to add more context to your reviews and chats.

✨ Finishing Touches
🧪 Generate unit tests
  • Create PR with unit tests
  • Post copyable unit tests in a comment
  • Commit unit tests in branch requirements

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share
🪧 Tips

Chat

There are 3 ways to chat with CodeRabbit:

  • Review comments: Directly reply to a review comment made by CodeRabbit. Example:
    • I pushed a fix in commit <commit_id>, please review it.
    • Open a follow-up GitHub issue for this discussion.
  • Files and specific lines of code (under the "Files changed" tab): Tag @coderabbitai in a new review comment at the desired location with your query.
  • PR comments: Tag @coderabbitai in a new PR comment to ask questions about the PR branch. For the best results, please provide a very specific query, as very limited context is provided in this mode. Examples:
    • @coderabbitai gather interesting stats about this repository and render them as a table. Additionally, render a pie chart showing the language distribution in the codebase.
    • @coderabbitai read the files in the src/scheduler package and generate a class diagram using mermaid and a README in the markdown format.

Support

Need help? Create a ticket on our support page for assistance with any issues or questions.

CodeRabbit Commands (Invoked using PR/Issue comments)

Type @coderabbitai help to get the list of available commands.

Other keywords and placeholders

  • Add @coderabbitai ignore anywhere in the PR description to prevent this PR from being reviewed.
  • Add @coderabbitai summary to generate the high-level summary at a specific location in the PR description.
  • Add @coderabbitai anywhere in the PR title to generate the title automatically.

CodeRabbit Configuration File (.coderabbit.yaml)

  • You can programmatically configure CodeRabbit by adding a .coderabbit.yaml file to the root of your repository.
  • Please see the configuration documentation for more information.
  • If your editor has YAML language server enabled, you can add the path at the top of this file to enable auto-completion and validation: # yaml-language-server: $schema=https://coderabbit.ai/integrations/schema.v2.json

Status, Documentation and Community

  • Visit our Status Page to check the current availability of CodeRabbit.
  • Visit our Documentation for detailed information on how to use CodeRabbit.
  • Join our Discord Community to get help, request features, and share feedback.
  • Follow us on X/Twitter for updates and announcements.

Copy link

netlify bot commented Aug 26, 2025

Deploy Preview for redpanda-docs-preview ready!

Name Link
🔨 Latest commit de1eb5e
🔍 Latest deploy log https://app.netlify.com/projects/redpanda-docs-preview/deploys/68adc9b105891000080dfdf2
😎 Deploy Preview https://deploy-preview-1343--redpanda-docs-preview.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify project configuration.

Copy link

netlify bot commented Aug 26, 2025

Deploy Preview for redpanda-docs-preview ready!

Name Link
🔨 Latest commit 80ab4dd
🔍 Latest deploy log https://app.netlify.com/projects/redpanda-docs-preview/deploys/68adc9e07dd67b00087cf3e3
😎 Deploy Preview https://deploy-preview-1343--redpanda-docs-preview.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify project configuration.

Copy link
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 1

🧹 Nitpick comments (2)
modules/deploy/partials/requirements.adoc (1)

137-138: Clarify scope: “NVMe required” is per broker node; tighten wording for SSDs.

Explicitly state that the requirement is per broker node and prefer “NVMe SSDs” for precision. This avoids readers interpreting it as a cluster‑wide aggregate requirement or as optional NVMe interfaces on non-SSD media.

-- NVMe (Non-Volatile Memory Express) drives are required for production deployments. NVMe drives provide the high throughput and low latency needed for optimal Redpanda performance.
+- NVMe (Non-Volatile Memory Express) SSDs are required for production deployments, per broker node. NVMe provides the high throughput and low latency needed for optimal Redpanda performance.
modules/deploy/partials/self-test.adoc (1)

5-7: Avoid “bandwidth test” wording and add “per broker node” qualifier for IOPS.

The self-test section currently says “storage bandwidth test” while asking readers to validate IOPS. That can confuse readers because “bandwidth” and “IOPS” are distinct metrics. Recommend neutralizing to “storage test” (or the exact test name used by the tool) and explicitly stating “per broker node”.

-For more information, see xref:manage:cluster-maintenance/cluster-diagnostics.adoc#self-test[Disk and network self-test benchmarks].
+For more information, see xref:manage:cluster-maintenance/cluster-diagnostics.adoc#self-test[Disk and network self-test benchmarks].

-When using the storage bandwidth test, ensure that your results show at least 16,000 IOPS (Input/Output Operations Per Second) for production environments. If your test results are below this threshold, your storage may not be suitable for production Redpanda workloads.
+When running the storage test, ensure that your results show at least 16,000 IOPS (Input/Output Operations Per Second) per broker node for production environments. If your test results are below this threshold, your storage may not be suitable for production Redpanda workloads. See xref:deploy:partials/requirements.adoc#_storage[Storage requirements] for details.

Please also verify that the self-test output surfaces an IOPS value directly. If it reports only throughput (MB/s), add a brief note on how the tool exposes IOPS (or adjust the wording to the exact metric the tool reports).

📜 Review details

Configuration used: CodeRabbit UI

Review profile: CHILL

Plan: Pro

💡 Knowledge Base configuration:

  • MCP integration is disabled by default for public repositories
  • Jira integration is disabled by default for public repositories
  • Linear integration is disabled by default for public repositories

You can enable these sources in your CodeRabbit configuration.

📥 Commits

Reviewing files that changed from the base of the PR and between 1126e17 and 80ab4dd.

📒 Files selected for processing (2)
  • modules/deploy/partials/requirements.adoc (1 hunks)
  • modules/deploy/partials/self-test.adoc (1 hunks)
🔇 Additional comments (1)
modules/deploy/partials/requirements.adoc (1)

141-142: Nice: Cross-reference to the self-test benchmarks is added.

This addresses the earlier reviewer suggestion to add the “See also” link, creating a clear validation path for users.

Comment on lines +139 to +142
- At least 16,000 IOPS (Input/Output Operations Per Second).
+
See also: xref:manage:cluster-maintenance/cluster-diagnostics.adoc#self-test[Disk and network self-test benchmarks].

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Verification agent

🧩 Analysis chain

Make the IOPS threshold explicitly “per broker node” and align phrasing with self-test.

Calling out the unit of enforcement (“per broker node”) prevents misconfiguration when users size storage for multi-broker clusters. Also, reference the self-test here to drive a single source of truth for how to validate.

-- At least 16,000 IOPS (Input/Output Operations Per Second).
-+
-See also: xref:manage:cluster-maintenance/cluster-diagnostics.adoc#self-test[Disk and network self-test benchmarks].
+- At least 16,000 IOPS (Input/Output Operations Per Second) per broker node.
++
+Validate your storage with the xref:manage:cluster-maintenance/cluster-diagnostics.adoc#self-test[Disk and network self-test benchmarks].

Verification script to catch lingering inconsistencies elsewhere (for example, pages that still call NVMe “recommended” or omit the 16,000 IOPS figure):


🏁 Script executed:

#!/bin/bash
# Find potentially outdated NVMe phrasing
rg -n -C2 -iP 'nvme.*recommend|recommend.*nvme|nvme recommended'

# Find IOPS thresholds and confirm consistency with 16,000
rg -n -C2 -iP '\b(?:16,?000)\b.*iops|iops.*\b(?:16,?000)\b'

# Scan for any allowance of NFS on data or Tiered Storage cache
rg -n -C2 -iP 'nfs.*(supported|support)'

Length of output: 5152


Clarify IOPS and NVMe requirements across documentation

The storage requirements should explicitly reference “per broker node” for IOPS and align all NVMe phrasing with the production requirement. Please update the following:

• In modules/deploy/partials/requirements.adoc (lines 139–142):

- At least 16,000 IOPS (Input/Output Operations Per Second).
- See also: xref:manage:cluster-maintenance/cluster-diagnostics.adoc#self-test[Disk and network self-test benchmarks].
+ At least 16,000 IOPS (Input/Output Operations Per Second) per broker node.
+
+ Validate your storage with the xref:manage:cluster-maintenance/cluster-diagnostics.adoc#self-test[Disk and network self-test benchmarks].

• In Kubernetes guides, change “Local NVMe disks, which is recommended for best performance” to reflect that NVMe is required for production:
– modules/deploy/pages/redpanda/kubernetes/gke-guide.adoc (line 37)
– modules/deploy/pages/redpanda/kubernetes/aks-guide.adoc (line 38)
– modules/deploy/pages/redpanda/kubernetes/eks-guide.adoc (line 254)

These edits ensure a single source of truth for storage sizing and drive home the requirement for NVMe and per-node IOPS thresholds.

📝 Committable suggestion

‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.

Suggested change
- At least 16,000 IOPS (Input/Output Operations Per Second).
+
See also: xref:manage:cluster-maintenance/cluster-diagnostics.adoc#self-test[Disk and network self-test benchmarks].
At least 16,000 IOPS (Input/Output Operations Per Second) per broker node.
Validate your storage with the xref:manage:cluster-maintenance/cluster-diagnostics.adoc#self-test[Disk and network self-test benchmarks].
🤖 Prompt for AI Agents
In modules/deploy/partials/requirements.adoc around lines 139–142, clarify that
the IOPS requirement is "at least 16,000 IOPS per broker node" and update the
sentence to reference per-node IOPS and link to the existing cluster diagnostics
self-test; additionally, in the Kubernetes guides update the phrasing to state
that Local NVMe disks are "required for production" (not just recommended) in
these files and lines: modules/deploy/pages/redpanda/kubernetes/gke-guide.adoc
(line 37), modules/deploy/pages/redpanda/kubernetes/aks-guide.adoc (line 38),
and modules/deploy/pages/redpanda/kubernetes/eks-guide.adoc (line 254), ensuring
wording is consistent across docs and aligns with production requirements.

Copy link
Contributor

@Feediver1 Feediver1 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

lgtm

@JakeSCahill JakeSCahill merged commit 7c1a60b into main Aug 27, 2025
7 checks passed
@JakeSCahill JakeSCahill deleted the requirements branch August 27, 2025 13:50
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants