Review Checklist

When reviewing tickets in Apache JIRA, the following items should be covered as part of the review process:


  • Does it conform to the Code Style guidelines?
  • Is there any redundant or duplicate code?
  • Is the code as modular as possible?
  • Can any singletons be avoided?
  • Can any of the code be replaced with library functions?
  • Are units of measurement used in the code consistent, both internally and with the rest of the ecosystem?


  • Are all data inputs and outputs checked (for the correct type, length, format, and range) and encoded?
  • Where third-party utilities are used, are returning errors being caught?
  • Are invalid parameter values handled?
  • Are any Throwable/Exceptions passed to the JVMStabilityInspector?
  • Are errors well-documented? Does the error message tell the user how to proceed?
  • Do exceptions propagate to the appropriate level in the code?


  • Do comments exist and describe the intent of the code (the “why”, not the “how”)?
  • Are javadocs added where appropriate?
  • Is any unusual behavior or edge-case handling described?
  • Are data structures and units of measurement explained?
  • Is there any incomplete code? If so, should it be removed or flagged with a suitable marker like ‘TODO’?
  • Does the code self-document via clear naming, abstractions, and flow control?
  • Have NEWS.txt, the cql3 docs, and the native protocol spec been updated if needed?
  • Is the ticket tagged with “client-impacting” and “doc-impacting”, where appropriate?
  • Has lib/licences been updated for third-party libs? Are they Apache License compatible?
  • Is the Component on the JIRA ticket set appropriately?


  • Is the code testable? i.e. don’t add too many or hide dependencies, unable to initialize objects, test frameworks can use methods etc.
  • Do tests exist and are they comprehensive?
  • Do unit tests actually test that the code is performing the intended functionality?
  • Could any test code use common functionality (e.g. ccm, dtest, or CqlTester methods) or abstract it there for reuse?
  • If the code may be affected by multi-node clusters, are there dtests?
  • If the code may take a long time to test properly, are there CVH tests?
  • Is the test passing on CI for all affected branches (up to trunk, if applicable)? Are there any regressions?
  • If patch affects read/write path, did we test for performance regressions w/multiple workloads?
  • If adding a new feature, were tests added and performed confirming it meets the expected SLA/use-case requirements for the feature?


  • Are logging statements logged at the correct level?
  • Are there logs in the critical path that could affect performance?
  • Is there any log that could be added to communicate status or troubleshoot potential problems in this feature?
  • Can any unnecessary logging statement be removed?