Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

chore: add release docs + general docs cleanup and polish #491

Merged
merged 1 commit into from
Feb 12, 2025

Conversation

tobz
Copy link
Member

@tobz tobz commented Feb 11, 2025

Summary

Started out intending to add documentation around doing ADP releases, but spiraled out in some general documentation cleanup as well.

Change Type

  • Bug fix
  • New feature
  • Non-functional (chore, refactoring, docs)
  • Performance

How did you test this PR?

N/A

References

N/A

@tobz tobz added the type/chore Updates to dependencies or general "administrative" tasks necessary to maintain the codebase/repo. label Feb 11, 2025
@tobz tobz requested a review from a team as a code owner February 11, 2025 14:24
@pr-commenter
Copy link

pr-commenter bot commented Feb 11, 2025

Regression Detector (DogStatsD)

Regression Detector Results

Run ID: cfaa05c1-3c57-4e08-b264-654be8f888e3

Baseline: 7.63.0-rc.2
Comparison: 7.63.0-rc.2

Optimization Goals: ✅ No significant changes detected

Fine details of change detection per experiment

perf experiment goal Δ mean % Δ mean % CI trials links
dsd_uds_500mb_3k_contexts ingress throughput +3.57 [+3.39, +3.74] 1
dsd_uds_100mb_3k_contexts ingress throughput +0.01 [-0.03, +0.05] 1
dsd_uds_1mb_50k_contexts_memlimit ingress throughput +0.00 [-0.00, +0.00] 1
dsd_uds_100mb_250k_contexts ingress throughput +0.00 [-0.00, +0.00] 1
dsd_uds_1mb_50k_contexts ingress throughput +0.00 [-0.00, +0.00] 1
dsd_uds_1mb_3k_contexts ingress throughput +0.00 [-0.00, +0.00] 1
dsd_uds_512kb_3k_contexts ingress throughput -0.00 [-0.01, +0.01] 1
dsd_uds_1mb_3k_contexts_dualship ingress throughput -0.00 [-0.00, +0.00] 1
dsd_uds_40mb_12k_contexts_40_senders ingress throughput -0.00 [-0.00, +0.00] 1
dsd_uds_10mb_3k_contexts ingress throughput -0.01 [-0.02, +0.00] 1
quality_gates_idle_rss memory utilization -0.07 [-0.17, +0.03] 1
dsd_uds_100mb_3k_contexts_distributions_only memory utilization -1.10 [-1.25, -0.95] 1

Bounds Checks: ❌ Failed

perf experiment bounds_check_name replicates_passed links
quality_gates_idle_rss memory_usage 0/10

Explanation

Confidence level: 90.00%
Effect size tolerance: |Δ mean %| ≥ 5.00%

Performance changes are noted in the perf column of each table:

  • ✅ = significantly better comparison variant performance
  • ❌ = significantly worse comparison variant performance
  • ➖ = no significant change in performance

A regression test is an A/B test of target performance in a repeatable rig, where "performance" is measured as "comparison variant minus baseline variant" for an optimization goal (e.g., ingress throughput). Due to intrinsic variability in measuring that goal, we can only estimate its mean value for each experiment; we report uncertainty in that value as a 90.00% confidence interval denoted "Δ mean % CI".

For each experiment, we decide whether a change in performance is a "regression" -- a change worth investigating further -- if all of the following criteria are true:

  1. Its estimated |Δ mean %| ≥ 5.00%, indicating the change is big enough to merit a closer look.

  2. Its 90.00% confidence interval "Δ mean % CI" does not contain zero, indicating that if our statistical model is accurate, there is at least a 90.00% chance there is a difference in performance between baseline and comparison variants.

  3. Its configuration does not mark it "erratic".

@pr-commenter
Copy link

pr-commenter bot commented Feb 11, 2025

Regression Detector (Saluki)

Regression Detector Results

Run ID: 56def7aa-a47a-4e58-9f85-5d27f93425f0

Baseline: ba9c3a5
Comparison: f814080
Diff

Optimization Goals: ✅ No significant changes detected

Fine details of change detection per experiment

perf experiment goal Δ mean % Δ mean % CI trials links
dsd_uds_100mb_3k_contexts_distributions_only memory utilization +1.17 [+1.06, +1.29] 1
dsd_uds_500mb_3k_contexts ingress throughput +0.92 [+0.79, +1.04] 1
dsd_uds_1mb_50k_contexts_memlimit ingress throughput +0.14 [-0.10, +0.39] 1
dsd_uds_40mb_12k_contexts_40_senders ingress throughput +0.01 [-0.02, +0.04] 1
dsd_uds_50mb_10k_contexts_no_inlining_no_allocs ingress throughput +0.01 [-0.04, +0.07] 1
dsd_uds_50mb_10k_contexts_no_inlining ingress throughput +0.00 [-0.08, +0.08] 1
dsd_uds_100mb_250k_contexts ingress throughput -0.00 [-0.04, +0.04] 1
dsd_uds_1mb_3k_contexts_dualship ingress throughput -0.00 [-0.01, +0.00] 1
dsd_uds_100mb_3k_contexts ingress throughput -0.00 [-0.05, +0.05] 1
dsd_uds_1mb_3k_contexts ingress throughput -0.00 [-0.01, +0.00] 1
dsd_uds_1mb_50k_contexts ingress throughput -0.01 [-0.02, +0.01] 1
dsd_uds_512kb_3k_contexts ingress throughput -0.01 [-0.03, +0.00] 1
dsd_uds_10mb_3k_contexts ingress throughput -0.02 [-0.05, +0.02] 1
quality_gates_idle_rss memory utilization -0.15 [-0.18, -0.11] 1

Bounds Checks: ✅ Passed

perf experiment bounds_check_name replicates_passed links
quality_gates_idle_rss memory_usage 10/10

Explanation

Confidence level: 90.00%
Effect size tolerance: |Δ mean %| ≥ 5.00%

Performance changes are noted in the perf column of each table:

  • ✅ = significantly better comparison variant performance
  • ❌ = significantly worse comparison variant performance
  • ➖ = no significant change in performance

A regression test is an A/B test of target performance in a repeatable rig, where "performance" is measured as "comparison variant minus baseline variant" for an optimization goal (e.g., ingress throughput). Due to intrinsic variability in measuring that goal, we can only estimate its mean value for each experiment; we report uncertainty in that value as a 90.00% confidence interval denoted "Δ mean % CI".

For each experiment, we decide whether a change in performance is a "regression" -- a change worth investigating further -- if all of the following criteria are true:

  1. Its estimated |Δ mean %| ≥ 5.00%, indicating the change is big enough to merit a closer look.

  2. Its 90.00% confidence interval "Δ mean % CI" does not contain zero, indicating that if our statistical model is accurate, there is at least a 90.00% chance there is a difference in performance between baseline and comparison variants.

  3. Its configuration does not mark it "erratic".

@pr-commenter
Copy link

pr-commenter bot commented Feb 11, 2025

Regression Detector Links

Experiment Result Links

experiment link(s)
dsd_uds_100mb_250k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_100mb_3k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_100mb_3k_contexts_distributions_only [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_10mb_3k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_1mb_3k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_1mb_3k_contexts_dualship [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_1mb_50k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_1mb_50k_contexts_memlimit [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_40mb_12k_contexts_40_senders [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_500mb_3k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_512kb_3k_contexts [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
quality_gates_idle_rss [Profiling (ADP)] [Profiling (DSD)] [SMP Dashboard]
dsd_uds_50mb_10k_contexts_no_inlining (ADP only) [Profiling (ADP)] [SMP Dashboard]
dsd_uds_50mb_10k_contexts_no_inlining_no_allocs (ADP only) [Profiling (ADP)] [SMP Dashboard]

Copy link
Collaborator

@jszwedko jszwedko left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Happy to see these docs!

Comment on lines +55 to +65
When deciding what version to use, we follow the [Semantic Versioning](https://semver.org/) specification. This means,
in a nutshell:

- prior to v1.0.0:
* breaking changes are indicated by incrementing the minor version
* new features and bug fixes are generally indicated by incrementing the patch version
* if there are enough new features/bug fixes, we may instead opt to increment the minor version just for simplicity
- after v1.0.0:
* breaking changes are indicated by incrementing the major version (this should generally never happen after v1)
* new features are indicated by incrementing the minor version
* bug fixes are indicated by incrementing the patch version
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Some projects put this in VERSIONING.md so people may expect to find it there. You could create that doc and just point to here or vice versa.

Example: https://github.com/open-telemetry/opentelemetry-go/blob/main/VERSIONING.md

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'll definitely consider that as a follow-up improvement.

@tobz tobz merged commit 0204847 into main Feb 12, 2025
21 checks passed
@tobz tobz deleted the tobz/release-docs branch February 12, 2025 16:18
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
type/chore Updates to dependencies or general "administrative" tasks necessary to maintain the codebase/repo.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants