Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
148 changes: 0 additions & 148 deletions docs/self-hosted/observability/alerts.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -1350,154 +1350,6 @@ Generated query for critical alert: `max((sum(increase(src_email_send{success="f

<br />

## frontend: mean_successful_sentinel_duration_over_2h

<p class="subtitle">mean successful sentinel search duration over 2h</p>

**Descriptions**

- <span class="badge badge-warning">warning</span> frontend: 5s+ mean successful sentinel search duration over 2h for 15m0s
- <span class="badge badge-critical">critical</span> frontend: 8s+ mean successful sentinel search duration over 2h for 30m0s

**Next steps**

- Look at the breakdown by query to determine if a specific query type is being affected
- Check for high CPU usage on zoekt-webserver
- Check Honeycomb for unusual activity
- More help interpreting this metric is available in the [dashboards reference](dashboards#frontend-mean-successful-sentinel-duration-over-2h).
- **Silence this alert:** If you are aware of this alert and want to silence notifications for it, add the following to your site configuration and set a reminder to re-evaluate the alert:

```json
"observability.silenceAlerts": [
"warning_frontend_mean_successful_sentinel_duration_over_2h",
"critical_frontend_mean_successful_sentinel_duration_over_2h"
]
```

<sub>*Managed by the Sourcegraph Code Understanding team.*</sub>

<details>
<summary>Technical details</summary>

Generated query for warning alert: `max((sum(rate(src_search_response_latency_seconds_sum{source=~"searchblitz.*",status="success"\}[2h])) / sum(rate(src_search_response_latency_seconds_count\{source=~"searchblitz.*",status="success"}[2h]))) &gt;= 5)`

Generated query for critical alert: `max((sum(rate(src_search_response_latency_seconds_sum{source=~"searchblitz.*",status="success"\}[2h])) / sum(rate(src_search_response_latency_seconds_count\{source=~"searchblitz.*",status="success"}[2h]))) &gt;= 8)`

</details>

<br />

## frontend: mean_sentinel_stream_latency_over_2h

<p class="subtitle">mean successful sentinel stream latency over 2h</p>

**Descriptions**

- <span class="badge badge-warning">warning</span> frontend: 2s+ mean successful sentinel stream latency over 2h for 15m0s
- <span class="badge badge-critical">critical</span> frontend: 3s+ mean successful sentinel stream latency over 2h for 30m0s

**Next steps**

- Look at the breakdown by query to determine if a specific query type is being affected
- Check for high CPU usage on zoekt-webserver
- Check Honeycomb for unusual activity
- More help interpreting this metric is available in the [dashboards reference](dashboards#frontend-mean-sentinel-stream-latency-over-2h).
- **Silence this alert:** If you are aware of this alert and want to silence notifications for it, add the following to your site configuration and set a reminder to re-evaluate the alert:

```json
"observability.silenceAlerts": [
"warning_frontend_mean_sentinel_stream_latency_over_2h",
"critical_frontend_mean_sentinel_stream_latency_over_2h"
]
```

<sub>*Managed by the Sourcegraph Code Understanding team.*</sub>

<details>
<summary>Technical details</summary>

Generated query for warning alert: `max((sum(rate(src_search_streaming_latency_seconds_sum{source=~"searchblitz.*"\}[2h])) / sum(rate(src_search_streaming_latency_seconds_count\{source=~"searchblitz.*"}[2h]))) &gt;= 2)`

Generated query for critical alert: `max((sum(rate(src_search_streaming_latency_seconds_sum{source=~"searchblitz.*"\}[2h])) / sum(rate(src_search_streaming_latency_seconds_count\{source=~"searchblitz.*"}[2h]))) &gt;= 3)`

</details>

<br />

## frontend: 90th_percentile_successful_sentinel_duration_over_2h

<p class="subtitle">90th percentile successful sentinel search duration over 2h</p>

**Descriptions**

- <span class="badge badge-warning">warning</span> frontend: 5s+ 90th percentile successful sentinel search duration over 2h for 15m0s
- <span class="badge badge-critical">critical</span> frontend: 10s+ 90th percentile successful sentinel search duration over 2h for 3h30m0s

**Next steps**

- Look at the breakdown by query to determine if a specific query type is being affected
- Check for high CPU usage on zoekt-webserver
- Check Honeycomb for unusual activity
- More help interpreting this metric is available in the [dashboards reference](dashboards#frontend-90th-percentile-successful-sentinel-duration-over-2h).
- **Silence this alert:** If you are aware of this alert and want to silence notifications for it, add the following to your site configuration and set a reminder to re-evaluate the alert:

```json
"observability.silenceAlerts": [
"warning_frontend_90th_percentile_successful_sentinel_duration_over_2h",
"critical_frontend_90th_percentile_successful_sentinel_duration_over_2h"
]
```

<sub>*Managed by the Sourcegraph Code Understanding team.*</sub>

<details>
<summary>Technical details</summary>

Generated query for warning alert: `max((histogram_quantile(0.9, sum by (le) (label_replace(rate(src_search_response_latency_seconds_bucket{source=~"searchblitz.*",status="success"\}[2h]), "source", "$1", "source", "searchblitz_(.*)")))) &gt;= 5)`

Generated query for critical alert: `max((histogram_quantile(0.9, sum by (le) (label_replace(rate(src_search_response_latency_seconds_bucket{source=~"searchblitz.*",status="success"\}[2h]), "source", "$1", "source", "searchblitz_(.*)")))) &gt;= 10)`

</details>

<br />

## frontend: 90th_percentile_sentinel_stream_latency_over_2h

<p class="subtitle">90th percentile successful sentinel stream latency over 2h</p>

**Descriptions**

- <span class="badge badge-warning">warning</span> frontend: 4s+ 90th percentile successful sentinel stream latency over 2h for 15m0s
- <span class="badge badge-critical">critical</span> frontend: 6s+ 90th percentile successful sentinel stream latency over 2h for 3h30m0s

**Next steps**

- Look at the breakdown by query to determine if a specific query type is being affected
- Check for high CPU usage on zoekt-webserver
- Check Honeycomb for unusual activity
- More help interpreting this metric is available in the [dashboards reference](dashboards#frontend-90th-percentile-sentinel-stream-latency-over-2h).
- **Silence this alert:** If you are aware of this alert and want to silence notifications for it, add the following to your site configuration and set a reminder to re-evaluate the alert:

```json
"observability.silenceAlerts": [
"warning_frontend_90th_percentile_sentinel_stream_latency_over_2h",
"critical_frontend_90th_percentile_sentinel_stream_latency_over_2h"
]
```

<sub>*Managed by the Sourcegraph Code Understanding team.*</sub>

<details>
<summary>Technical details</summary>

Generated query for warning alert: `max((histogram_quantile(0.9, sum by (le) (label_replace(rate(src_search_streaming_latency_seconds_bucket{source=~"searchblitz.*"\}[2h]), "source", "$1", "source", "searchblitz_(.*)")))) &gt;= 4)`

Generated query for critical alert: `max((histogram_quantile(0.9, sum by (le) (label_replace(rate(src_search_streaming_latency_seconds_bucket{source=~"searchblitz.*"\}[2h]), "source", "$1", "source", "searchblitz_(.*)")))) &gt;= 6)`

</details>

<br />

## gitserver: disk_space_remaining

<p class="subtitle">disk space remaining</p>
Expand Down
114 changes: 9 additions & 105 deletions docs/self-hosted/observability/dashboards.mdx

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.