Dashboards reference
This document contains a complete reference on Sourcegraph's available dashboards, as well as details on how to interpret the panels and metrics.
To learn more about Sourcegraph's metrics and how to view these dashboards, see our metrics guide.
Frontend
Serves all end-user browser and API requests.
To see this dashboard, visit /-/debug/grafana/d/frontend/frontend on your Sourcegraph instance.
Frontend: Search at a glance
frontend: 99th_percentile_search_request_duration
99th percentile successful search request duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le)(rate(src_search_streaming_latency_seconds_bucket{source="browser"}[5m])))
frontend: 90th_percentile_search_request_duration
90th percentile successful search request duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le)(rate(src_search_streaming_latency_seconds_bucket{source="browser"}[5m])))
frontend: timeout_search_responses
Timeout search responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_search_streaming_response{status=~"timeout|partial_timeout",source="browser"}[5m])) / sum(increase(src_search_streaming_response{source="browser"}[5m])) * 100
frontend: hard_error_search_responses
Hard error search responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_search_streaming_response{status="error",source="browser"}[5m])) / sum(increase(src_search_streaming_response{source="browser"}[5m])) * 100
frontend: search_no_results
Searches with no results every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100012 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_search_streaming_response{status="no_results",source="browser"}[5m])) / sum(increase(src_search_streaming_response{source="browser"}[5m])) * 100
frontend: search_alert_user_suggestions
Search alert user suggestions shown every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100013 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (alert_type)(increase(src_search_streaming_response{status="alert",alert_type!~"timed_out",source="browser"}[5m])) / ignoring(alert_type) group_left sum(increase(src_search_streaming_response{source="browser"}[5m])) * 100
frontend: page_load_latency
90th percentile page load latency over all routes over 10m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100020 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.9, sum by(le) (rate(src_http_request_duration_seconds_bucket{route!="raw",route!="blob",route!~"graphql.*"}[10m])))
Frontend: Search-based code intelligence at a glance
frontend: 99th_percentile_search_codeintel_request_duration
99th percentile code-intel successful search request duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le)(rate(src_graphql_field_seconds_bucket{type="Search",field="results",error="false",source="browser",request_name="CodeIntelSearch"}[5m])))
frontend: 90th_percentile_search_codeintel_request_duration
90th percentile code-intel successful search request duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le)(rate(src_graphql_field_seconds_bucket{type="Search",field="results",error="false",source="browser",request_name="CodeIntelSearch"}[5m])))
frontend: hard_timeout_search_codeintel_responses
Hard timeout search code-intel responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(increase(src_graphql_search_response{status="timeout",source="browser",request_name="CodeIntelSearch"}[5m])) + sum(increase(src_graphql_search_response{status="alert",alert_type="timed_out",source="browser",request_name="CodeIntelSearch"}[5m]))) / sum(increase(src_graphql_search_response{source="browser",request_name="CodeIntelSearch"}[5m])) * 100
frontend: hard_error_search_codeintel_responses
Hard error search code-intel responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (status)(increase(src_graphql_search_response{status=~"error",source="browser",request_name="CodeIntelSearch"}[5m])) / ignoring(status) group_left sum(increase(src_graphql_search_response{source="browser",request_name="CodeIntelSearch"}[5m])) * 100
frontend: partial_timeout_search_codeintel_responses
Partial timeout search code-intel responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100112 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (status)(increase(src_graphql_search_response{status="partial_timeout",source="browser",request_name="CodeIntelSearch"}[5m])) / ignoring(status) group_left sum(increase(src_graphql_search_response{status="partial_timeout",source="browser",request_name="CodeIntelSearch"}[5m])) * 100
frontend: search_codeintel_alert_user_suggestions
Search code-intel alert user suggestions shown every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100113 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (alert_type)(increase(src_graphql_search_response{status="alert",alert_type!~"timed_out",source="browser",request_name="CodeIntelSearch"}[5m])) / ignoring(alert_type) group_left sum(increase(src_graphql_search_response{source="browser",request_name="CodeIntelSearch"}[5m])) * 100
Frontend: Search GraphQL API usage at a glance
frontend: 99th_percentile_search_api_request_duration
99th percentile successful search API request duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le)(rate(src_graphql_field_seconds_bucket{type="Search",field="results",error="false",source="other"}[5m])))
frontend: 90th_percentile_search_api_request_duration
90th percentile successful search API request duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le)(rate(src_graphql_field_seconds_bucket{type="Search",field="results",error="false",source="other"}[5m])))
frontend: hard_error_search_api_responses
Hard error search API responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (status)(increase(src_graphql_search_response{status=~"error",source="other"}[5m])) / ignoring(status) group_left sum(increase(src_graphql_search_response{source="other"}[5m]))
frontend: partial_timeout_search_api_responses
Partial timeout search API responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_graphql_search_response{status="partial_timeout",source="other"}[5m])) / sum(increase(src_graphql_search_response{source="other"}[5m]))
frontend: search_api_alert_user_suggestions
Search API alert user suggestions shown every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (alert_type)(increase(src_graphql_search_response{status="alert",alert_type!~"timed_out",source="other"}[5m])) / ignoring(alert_type) group_left sum(increase(src_graphql_search_response{status="alert",source="other"}[5m]))
Frontend: Site configuration client update latency
frontend: frontend_site_configuration_duration_since_last_successful_update_by_instance
Duration since last successful site configuration update (by instance)
The duration since the configuration client used by the "frontend" service last successfully updated its site configuration. Long durations could indicate issues updating the site configuration.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLsrc_conf_client_time_since_last_successful_update_seconds{job=~`(sourcegraph-)?frontend`,instance=~`${internalInstance:regex}`}
frontend: frontend_site_configuration_duration_since_last_successful_update_by_instance
Maximum duration since last successful site configuration update (all "frontend" instances)
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(max_over_time(src_conf_client_time_since_last_successful_update_seconds{job=~`(sourcegraph-)?frontend`,instance=~`${internalInstance:regex}`}[1m]))
Frontend: Codeintel: Precise code intelligence usage at a glance
frontend: codeintel_resolvers_total
Aggregate graphql operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_resolvers_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_resolvers_99th_percentile_duration
Aggregate successful graphql operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_resolvers_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_resolvers_errors_total
Aggregate graphql operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100402 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_resolvers_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_resolvers_error_rate
Aggregate graphql operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100403 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_resolvers_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_resolvers_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_resolvers_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_resolvers_total
Graphql operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100410 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_resolvers_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_resolvers_99th_percentile_duration
99th percentile successful graphql operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100411 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_resolvers_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_resolvers_errors_total
Graphql operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100412 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_resolvers_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_resolvers_error_rate
Graphql operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100413 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_resolvers_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_resolvers_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_resolvers_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: Auto-index enqueuer
frontend: codeintel_autoindex_enqueuer_total
Aggregate enqueuer operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindex_enqueuer_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_autoindex_enqueuer_99th_percentile_duration
Aggregate successful enqueuer operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100501 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_autoindex_enqueuer_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_autoindex_enqueuer_errors_total
Aggregate enqueuer operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100502 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindex_enqueuer_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_autoindex_enqueuer_error_rate
Aggregate enqueuer operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100503 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindex_enqueuer_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_autoindex_enqueuer_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_autoindex_enqueuer_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_autoindex_enqueuer_total
Enqueuer operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100510 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindex_enqueuer_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_autoindex_enqueuer_99th_percentile_duration
99th percentile successful enqueuer operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100511 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_autoindex_enqueuer_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_autoindex_enqueuer_errors_total
Enqueuer operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100512 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindex_enqueuer_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_autoindex_enqueuer_error_rate
Enqueuer operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100513 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindex_enqueuer_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_autoindex_enqueuer_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_autoindex_enqueuer_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: dbstore stats
frontend: codeintel_uploads_store_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100601 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploads_store_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_store_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100602 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_store_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100603 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_uploads_store_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_uploads_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_uploads_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100610 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_store_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100611 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_store_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_uploads_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100612 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100613 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_store_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Workerutil: lsif_indexes dbworker/store stats
frontend: workerutil_dbworker_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100700 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_total{domain='codeintel_index_jobs',job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: workerutil_dbworker_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100701 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_workerutil_dbworker_store_duration_seconds_bucket{domain='codeintel_index_jobs',job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: workerutil_dbworker_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100702 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_errors_total{domain='codeintel_index_jobs',job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: workerutil_dbworker_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100703 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_errors_total{domain='codeintel_index_jobs',job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_workerutil_dbworker_store_total{domain='codeintel_index_jobs',job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_workerutil_dbworker_store_errors_total{domain='codeintel_index_jobs',job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: lsifstore stats
frontend: codeintel_uploads_lsifstore_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100800 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_lsifstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_lsifstore_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100801 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploads_lsifstore_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_lsifstore_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100802 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_lsifstore_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100803 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_uploads_lsifstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_uploads_lsifstore_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100810 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_lsifstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_lsifstore_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100811 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_lsifstore_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_uploads_lsifstore_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100812 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_lsifstore_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100813 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_lsifstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: gitserver client
frontend: gitserver_client_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100900 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100901 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100902 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100903 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_gitserver_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: gitserver_client_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100910 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100911 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: gitserver_client_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100912 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100913 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_gitserver_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: uploadstore stats
frontend: codeintel_uploadstore_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploadstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploadstore_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploadstore_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploadstore_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101002 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploadstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploadstore_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101003 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploadstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_uploadstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_uploadstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_uploadstore_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploadstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploadstore_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101011 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploadstore_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_uploadstore_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101012 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploadstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploadstore_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101013 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploadstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_uploadstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_uploadstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Gitserver: Gitserver Client
frontend: gitserver_client_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_gitserver_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: gitserver_client_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101110 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op,scope)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: gitserver_client_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101112 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101113 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op,scope)(increase(src_gitserver_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Gitserver: Gitserver Repository Service Client
frontend: gitserver_repositoryservice_client_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_repositoryservice_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_repositoryservice_client_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_repositoryservice_client_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_repositoryservice_client_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_repositoryservice_client_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_gitserver_repositoryservice_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: gitserver_repositoryservice_client_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_repositoryservice_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_repositoryservice_client_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op,scope)(rate(src_gitserver_repositoryservice_client_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: gitserver_repositoryservice_client_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_repositoryservice_client_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101213 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op,scope)(increase(src_gitserver_repositoryservice_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op,scope)(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Batches: dbstore stats
frontend: batches_dbstore_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_dbstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_dbstore_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101301 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_batches_dbstore_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_dbstore_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101302 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_dbstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_dbstore_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101303 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_dbstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_batches_dbstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_batches_dbstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: batches_dbstore_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101310 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_dbstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_dbstore_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101311 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_batches_dbstore_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: batches_dbstore_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101312 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_dbstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_dbstore_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101313 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_dbstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_batches_dbstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_batches_dbstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Batches: service stats
frontend: batches_service_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_service_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_service_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_batches_service_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_service_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101402 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_service_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_service_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101403 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_service_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_batches_service_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_batches_service_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: batches_service_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101410 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_service_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_service_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101411 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_batches_service_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: batches_service_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101412 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_service_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_service_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101413 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_service_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_batches_service_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_batches_service_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Batches: HTTP API File Handler
frontend: batches_httpapi_total
Aggregate http handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_httpapi_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_httpapi_99th_percentile_duration
Aggregate successful http handler operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101501 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_batches_httpapi_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_httpapi_errors_total
Aggregate http handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101502 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_httpapi_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_httpapi_error_rate
Aggregate http handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101503 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_httpapi_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_batches_httpapi_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_batches_httpapi_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: batches_httpapi_total
Http handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101510 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_httpapi_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_httpapi_99th_percentile_duration
99th percentile successful http handler operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101511 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_batches_httpapi_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: batches_httpapi_errors_total
Http handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101512 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_httpapi_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_httpapi_error_rate
Http handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101513 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_httpapi_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_batches_httpapi_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_batches_httpapi_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Out-of-band migrations: up migration invocation (one batch processed)
frontend: oobmigration_total
Migration handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_oobmigration_total{op="up",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: oobmigration_99th_percentile_duration
Aggregate successful migration handler operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101601 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_oobmigration_duration_seconds_bucket{op="up",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: oobmigration_errors_total
Migration handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101602 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_oobmigration_errors_total{op="up",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: oobmigration_error_rate
Migration handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101603 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_oobmigration_errors_total{op="up",job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_oobmigration_total{op="up",job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_oobmigration_errors_total{op="up",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Out-of-band migrations: down migration invocation (one batch processed)
frontend: oobmigration_total
Migration handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101700 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_oobmigration_total{op="down",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: oobmigration_99th_percentile_duration
Aggregate successful migration handler operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101701 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_oobmigration_duration_seconds_bucket{op="down",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: oobmigration_errors_total
Migration handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101702 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_oobmigration_errors_total{op="down",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: oobmigration_error_rate
Migration handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101703 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_oobmigration_errors_total{op="down",job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_oobmigration_total{op="down",job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_oobmigration_errors_total{op="down",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Zoekt Configuration GRPC server metrics
frontend: zoekt_configuration_grpc_request_rate_all_methods
Request rate across all methods over 2m
The number of gRPC requests received per second across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101800 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m]))
frontend: zoekt_configuration_grpc_request_rate_per_method
Request rate per-method over 2m
The number of gRPC requests received per second broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101801 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])) by (grpc_method)
frontend: zoekt_configuration_error_percentage_all_methods
Error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101810 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_code!="OK",instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m]))) / (sum(rate(grpc_server_handled_total{instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m]))) ))
frontend: zoekt_configuration_grpc_error_percentage_per_method
Error percentage per-method over 2m
The percentage of gRPC requests that fail per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101811 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_method=~`${zoekt_configuration_method:regex}`,grpc_code!="OK",instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])) by (grpc_method)) / (sum(rate(grpc_server_handled_total{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])) by (grpc_method)) ))
frontend: zoekt_configuration_p99_response_time_per_method
99th percentile response time per method over 2m
The 99th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101820 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p90_response_time_per_method
90th percentile response time per method over 2m
The 90th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101821 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p75_response_time_per_method
75th percentile response time per method over 2m
The 75th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101822 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p99_9_response_size_per_method
99.9th percentile total response size per method over 2m
The 99.9th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101830 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p90_response_size_per_method
90th percentile total response size per method over 2m
The 90th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101831 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p75_response_size_per_method
75th percentile total response size per method over 2m
The 75th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101832 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p99_9_invididual_sent_message_size_per_method
99.9th percentile individual sent message size per method over 2m
The 99.9th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101840 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p90_invididual_sent_message_size_per_method
90th percentile individual sent message size per method over 2m
The 90th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101841 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p75_invididual_sent_message_size_per_method
75th percentile individual sent message size per method over 2m
The 75th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101842 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_grpc_response_stream_message_count_per_method
Average streaming response message count per-method over 2m
The average number of response messages sent during a streaming RPC method, broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101850 on your Sourcegraph instance.
Technical details
Query:
SHELL((sum(rate(grpc_server_msg_sent_total{grpc_type="server_stream",instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])) by (grpc_method))/(sum(rate(grpc_server_started_total{grpc_type="server_stream",instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])) by (grpc_method)))
frontend: zoekt_configuration_grpc_all_codes_per_method
Response codes rate per-method over 2m
The rate of all generated gRPC response codes per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101860 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_handled_total{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])) by (grpc_method, grpc_code)
Frontend: Zoekt Configuration GRPC "internal error" metrics
frontend: zoekt_configuration_grpc_clients_error_percentage_all_methods
Client baseline error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods (regardless of whether or not there was an internal error), aggregated across all "zoekt_configuration" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101900 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_code!="OK"}[2m])))) / ((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))))))
frontend: zoekt_configuration_grpc_clients_error_percentage_per_method
Client baseline error percentage per-method over 2m
The percentage of gRPC requests that fail per method (regardless of whether or not there was an internal error), aggregated across all "zoekt_configuration" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101901 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}",grpc_code!="OK"}[2m])) by (grpc_method))) / ((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}"}[2m])) by (grpc_method))))))
frontend: zoekt_configuration_grpc_clients_all_codes_per_method
Client baseline response codes rate per-method over 2m
The rate of all generated gRPC response codes per method (regardless of whether or not there was an internal error), aggregated across all "zoekt_configuration" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101902 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}"}[2m])) by (grpc_method, grpc_code))
frontend: zoekt_configuration_grpc_clients_internal_error_percentage_all_methods
Client-observed gRPC internal error percentage across all methods over 2m
The percentage of gRPC requests that appear to fail due to gRPC internal errors across all methods, aggregated across all "zoekt_configuration" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "zoekt_configuration" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101910 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_code!="OK",is_internal_error="true"}[2m])))) / ((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))))))
frontend: zoekt_configuration_grpc_clients_internal_error_percentage_per_method
Client-observed gRPC internal error percentage per-method over 2m
The percentage of gRPC requests that appear to fail to due to gRPC internal errors per method, aggregated across all "zoekt_configuration" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "zoekt_configuration" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101911 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}",grpc_code!="OK",is_internal_error="true"}[2m])) by (grpc_method))) / ((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}"}[2m])) by (grpc_method))))))
frontend: zoekt_configuration_grpc_clients_internal_error_all_codes_per_method
Client-observed gRPC internal error response code rate per-method over 2m
The rate of gRPC internal-error response codes per method, aggregated across all "zoekt_configuration" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "zoekt_configuration" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101912 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",is_internal_error="true",grpc_method=~"${zoekt_configuration_method:regex}"}[2m])) by (grpc_method, grpc_code))
Frontend: Zoekt Configuration GRPC retry metrics
frontend: zoekt_configuration_grpc_clients_retry_percentage_across_all_methods
Client retry percentage across all methods over 2m
The percentage of gRPC requests that were retried across all methods, aggregated across all "zoekt_configuration" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102000 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",is_retried="true"}[2m])))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))))))
frontend: zoekt_configuration_grpc_clients_retry_percentage_per_method
Client retry percentage per-method over 2m
The percentage of gRPC requests that were retried aggregated across all "zoekt_configuration" clients, broken out per method.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102001 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",is_retried="true",grpc_method=~"${zoekt_configuration_method:regex}"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}"}[2m])) by (grpc_method))))))
frontend: zoekt_configuration_grpc_clients_retry_count_per_method
Client retry count per-method over 2m
The count of gRPC requests that were retried aggregated across all "zoekt_configuration" clients, broken out per method
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102002 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}",is_retried="true"}[2m])) by (grpc_method))
Frontend: Internal Api GRPC server metrics
frontend: internal_api_grpc_request_rate_all_methods
Request rate across all methods over 2m
The number of gRPC requests received per second across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m]))
frontend: internal_api_grpc_request_rate_per_method
Request rate per-method over 2m
The number of gRPC requests received per second broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])) by (grpc_method)
frontend: internal_api_error_percentage_all_methods
Error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102110 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_code!="OK",instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m]))) / (sum(rate(grpc_server_handled_total{instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m]))) ))
frontend: internal_api_grpc_error_percentage_per_method
Error percentage per-method over 2m
The percentage of gRPC requests that fail per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102111 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_method=~`${internal_api_method:regex}`,grpc_code!="OK",instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])) by (grpc_method)) / (sum(rate(grpc_server_handled_total{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])) by (grpc_method)) ))
frontend: internal_api_p99_response_time_per_method
99th percentile response time per method over 2m
The 99th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102120 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p90_response_time_per_method
90th percentile response time per method over 2m
The 90th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102121 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p75_response_time_per_method
75th percentile response time per method over 2m
The 75th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102122 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p99_9_response_size_per_method
99.9th percentile total response size per method over 2m
The 99.9th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102130 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p90_response_size_per_method
90th percentile total response size per method over 2m
The 90th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102131 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p75_response_size_per_method
75th percentile total response size per method over 2m
The 75th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102132 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p99_9_invididual_sent_message_size_per_method
99.9th percentile individual sent message size per method over 2m
The 99.9th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102140 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p90_invididual_sent_message_size_per_method
90th percentile individual sent message size per method over 2m
The 90th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102141 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p75_invididual_sent_message_size_per_method
75th percentile individual sent message size per method over 2m
The 75th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102142 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_grpc_response_stream_message_count_per_method
Average streaming response message count per-method over 2m
The average number of response messages sent during a streaming RPC method, broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102150 on your Sourcegraph instance.
Technical details
Query:
SHELL((sum(rate(grpc_server_msg_sent_total{grpc_type="server_stream",instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])) by (grpc_method))/(sum(rate(grpc_server_started_total{grpc_type="server_stream",instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])) by (grpc_method)))
frontend: internal_api_grpc_all_codes_per_method
Response codes rate per-method over 2m
The rate of all generated gRPC response codes per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102160 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_handled_total{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])) by (grpc_method, grpc_code)
Frontend: Internal Api GRPC "internal error" metrics
frontend: internal_api_grpc_clients_error_percentage_all_methods
Client baseline error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods (regardless of whether or not there was an internal error), aggregated across all "internal_api" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102200 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_code!="OK"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))))))
frontend: internal_api_grpc_clients_error_percentage_per_method
Client baseline error percentage per-method over 2m
The percentage of gRPC requests that fail per method (regardless of whether or not there was an internal error), aggregated across all "internal_api" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102201 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}",grpc_code!="OK"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}"}[2m])) by (grpc_method))))))
frontend: internal_api_grpc_clients_all_codes_per_method
Client baseline response codes rate per-method over 2m
The rate of all generated gRPC response codes per method (regardless of whether or not there was an internal error), aggregated across all "internal_api" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102202 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}"}[2m])) by (grpc_method, grpc_code))
frontend: internal_api_grpc_clients_internal_error_percentage_all_methods
Client-observed gRPC internal error percentage across all methods over 2m
The percentage of gRPC requests that appear to fail due to gRPC internal errors across all methods, aggregated across all "internal_api" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "internal_api" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102210 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_code!="OK",is_internal_error="true"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))))))
frontend: internal_api_grpc_clients_internal_error_percentage_per_method
Client-observed gRPC internal error percentage per-method over 2m
The percentage of gRPC requests that appear to fail to due to gRPC internal errors per method, aggregated across all "internal_api" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "internal_api" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102211 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}",grpc_code!="OK",is_internal_error="true"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}"}[2m])) by (grpc_method))))))
frontend: internal_api_grpc_clients_internal_error_all_codes_per_method
Client-observed gRPC internal error response code rate per-method over 2m
The rate of gRPC internal-error response codes per method, aggregated across all "internal_api" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "internal_api" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102212 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",is_internal_error="true",grpc_method=~"${internal_api_method:regex}"}[2m])) by (grpc_method, grpc_code))
Frontend: Internal Api GRPC retry metrics
frontend: internal_api_grpc_clients_retry_percentage_across_all_methods
Client retry percentage across all methods over 2m
The percentage of gRPC requests that were retried across all methods, aggregated across all "internal_api" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102300 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"api.internalapi.v1.ConfigService",is_retried="true"}[2m])))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))))))
frontend: internal_api_grpc_clients_retry_percentage_per_method
Client retry percentage per-method over 2m
The percentage of gRPC requests that were retried aggregated across all "internal_api" clients, broken out per method.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102301 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"api.internalapi.v1.ConfigService",is_retried="true",grpc_method=~"${internal_api_method:regex}"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}"}[2m])) by (grpc_method))))))
frontend: internal_api_grpc_clients_retry_count_per_method
Client retry count per-method over 2m
The count of gRPC requests that were retried aggregated across all "internal_api" clients, broken out per method
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102302 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}",is_retried="true"}[2m])) by (grpc_method))
Frontend: Internal service requests
frontend: internal_indexed_search_error_responses
Internal indexed search error responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(code) (increase(src_zoekt_request_duration_seconds_count{code!~"2.."}[5m])) / ignoring(code) group_left sum(increase(src_zoekt_request_duration_seconds_count[5m])) * 100
frontend: internal_unindexed_search_error_responses
Internal unindexed search error responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(code) (increase(searcher_service_request_total{code!~"2.."}[5m])) / ignoring(code) group_left sum(increase(searcher_service_request_total[5m])) * 100
frontend: 99th_percentile_gitserver_duration
99th percentile successful gitserver query duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102410 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,category)(rate(src_gitserver_request_duration_seconds_bucket{job=~"(sourcegraph-)?frontend"}[5m])))
frontend: gitserver_error_responses
Gitserver error responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102411 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (category)(increase(src_gitserver_request_duration_seconds_count{job=~"(sourcegraph-)?frontend",code!~"2.."}[5m])) / ignoring(code) group_left sum by (category)(increase(src_gitserver_request_duration_seconds_count{job=~"(sourcegraph-)?frontend"}[5m])) * 100
frontend: observability_test_alert_warning
Warning test alert metric
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102420 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(owner) (observability_test_metric_warning)
frontend: observability_test_alert_critical
Critical test alert metric
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102421 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(owner) (observability_test_metric_critical)
Frontend: Authentication API requests
frontend: sign_in_rate
Rate of API requests to sign-in
Rate (QPS) of requests to sign-in
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(irate(src_http_request_duration_seconds_count{route="sign-in",method="post"}[5m]))
frontend: sign_in_latency_p99
99 percentile of sign-in latency
99% percentile of sign-in latency
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102501 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_http_request_duration_seconds_bucket{route="sign-in",method="post"}[5m])) by (le))
frontend: sign_in_error_rate
Percentage of sign-in requests by http code
Percentage of sign-in requests grouped by http code
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102502 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (code)(irate(src_http_request_duration_seconds_count{route="sign-in",method="post"}[5m]))/ ignoring (code) group_left sum(irate(src_http_request_duration_seconds_count{route="sign-in",method="post"}[5m]))*100
frontend: sign_up_rate
Rate of API requests to sign-up
Rate (QPS) of requests to sign-up
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102510 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(irate(src_http_request_duration_seconds_count{route="sign-up",method="post"}[5m]))
frontend: sign_up_latency_p99
99 percentile of sign-up latency
99% percentile of sign-up latency
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102511 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_http_request_duration_seconds_bucket{route="sign-up",method="post"}[5m])) by (le))
frontend: sign_up_code_percentage
Percentage of sign-up requests by http code
Percentage of sign-up requests grouped by http code
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102512 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (code)(irate(src_http_request_duration_seconds_count{route="sign-up",method="post"}[5m]))/ ignoring (code) group_left sum(irate(src_http_request_duration_seconds_count{route="sign-out"}[5m]))*100
frontend: sign_out_rate
Rate of API requests to sign-out
Rate (QPS) of requests to sign-out
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102520 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(irate(src_http_request_duration_seconds_count{route="sign-out"}[5m]))
frontend: sign_out_latency_p99
99 percentile of sign-out latency
99% percentile of sign-out latency
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102521 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_http_request_duration_seconds_bucket{route="sign-out"}[5m])) by (le))
frontend: sign_out_error_rate
Percentage of sign-out requests that return non-303 http code
Percentage of sign-out requests grouped by http code
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102522 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (code)(irate(src_http_request_duration_seconds_count{route="sign-out"}[5m]))/ ignoring (code) group_left sum(irate(src_http_request_duration_seconds_count{route="sign-out"}[5m]))*100
frontend: account_failed_sign_in_attempts
Rate of failed sign-in attempts
Failed sign-in attempts per minute
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102530 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_frontend_account_failed_sign_in_attempts_total[1m]))
frontend: account_lockouts
Rate of account lockouts
Account lockouts per minute
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102531 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_frontend_account_lockouts_total[1m]))
Frontend: External HTTP Request Rate
frontend: external_http_request_rate_by_host
Rate of external HTTP requests by host over 1m
Shows the rate of external HTTP requests made by Sourcegraph to other services, broken down by host.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (host) (rate(src_http_client_external_request_count{host=~`${httpRequestHost:regex}`}[1m]))
frontend: external_http_request_rate_by_host_by_code
Rate of external HTTP requests by host and response code over 1m
Shows the rate of external HTTP requests made by Sourcegraph to other services, broken down by host and response code.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102610 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (host, status_code) (rate(src_http_client_external_request_count{host=~`${httpRequestHost:regex}`}[1m]))
Frontend: Cody API requests
frontend: cody_api_rate
Rate of API requests to cody endpoints (excluding GraphQL)
Rate (QPS) of requests to cody related endpoints. completions.stream is for the conversational endpoints. completions.code is for the code auto-complete endpoints.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102700 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (route, code)(irate(src_http_request_duration_seconds_count{route=~"^completions.*"}[5m]))
Frontend: Cloud KMS and cache
frontend: cloudkms_cryptographic_requests
Cryptographic requests to Cloud KMS every 1m
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102800 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_cloudkms_cryptographic_total[1m]))
frontend: encryption_cache_hit_ratio
Average encryption cache hit ratio per workload
- Encryption cache hit ratio (hits/(hits+misses)) - minimum across all instances of a workload.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102801 on your Sourcegraph instance.
Technical details
Query:
SHELLmin by (kubernetes_name) (src_encryption_cache_hit_total/(src_encryption_cache_hit_total+src_encryption_cache_miss_total))
frontend: encryption_cache_evictions
Rate of encryption cache evictions - sum across all instances of a given workload
- Rate of encryption cache evictions (caused by cache exceeding its maximum size) - sum across all instances of a workload
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102802 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (kubernetes_name) (irate(src_encryption_cache_eviction_total[5m]))
Frontend: Periodic Goroutines
frontend: running_goroutines
Number of currently running periodic goroutines
The number of currently running periodic goroutines by name and job.
A value of 0 indicates the routine isnt running currently, it awaits its next schedule.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102900 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (src_periodic_goroutine_running{job=~".*frontend.*"})
frontend: goroutine_success_rate
Success rate for periodic goroutine executions
The rate of successful executions of each periodic goroutine. A low or zero value could indicate that a routine is stalled or encountering errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102901 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_total{job=~".*frontend.*"}[5m]))
frontend: goroutine_error_rate
Error rate for periodic goroutine executions
The rate of errors encountered by each periodic goroutine. A sustained high error rate may indicate a problem with the routine`s configuration or dependencies.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102910 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_errors_total{job=~".*frontend.*"}[5m]))
frontend: goroutine_error_percentage
Percentage of periodic goroutine executions that result in errors
The percentage of executions that result in errors for each periodic goroutine. A value above 5% indicates that a significant portion of routine executions are failing.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102911 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_errors_total{job=~".*frontend.*"}[5m])) / sum by (name, job_name) (rate(src_periodic_goroutine_total{job=~".*frontend.*"}[5m]) > 0) * 100
frontend: goroutine_handler_duration
95th percentile handler execution time
The 95th percentile execution time for each periodic goroutine handler. Longer durations might indicate increased load or processing time.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102920 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job_name, le) (rate(src_periodic_goroutine_duration_seconds_bucket{job=~".*frontend.*"}[5m])))
frontend: goroutine_loop_duration
95th percentile loop cycle time
The 95th percentile loop cycle time for each periodic goroutine (excluding sleep time). This represents how long a complete loop iteration takes before sleeping for the next interval.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102921 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job_name, le) (rate(src_periodic_goroutine_loop_duration_seconds_bucket{job=~".*frontend.*"}[5m])))
frontend: tenant_processing_duration
95th percentile tenant processing time
The 95th percentile processing time for individual tenants within periodic goroutines. Higher values indicate that tenant processing is taking longer and may affect overall performance.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102930 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job_name, le) (rate(src_periodic_goroutine_tenant_duration_seconds_bucket{job=~".*frontend.*"}[5m])))
frontend: tenant_processing_max
Maximum tenant processing time
The maximum processing time for individual tenants within periodic goroutines. Consistently high values might indicate problematic tenants or inefficient processing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102931 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name, job_name) (rate(src_periodic_goroutine_tenant_duration_seconds_sum{job=~".*frontend.*"}[5m]) / rate(src_periodic_goroutine_tenant_duration_seconds_count{job=~".*frontend.*"}[5m]))
frontend: tenant_count
Number of tenants processed per routine
The number of tenants processed by each periodic goroutine. Unexpected changes can indicate tenant configuration issues or scaling events.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102940 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name, job_name) (src_periodic_goroutine_tenant_count{job=~".*frontend.*"})
frontend: tenant_success_rate
Rate of successful tenant processing operations
The rate of successful tenant processing operations. A healthy routine should maintain a consistent processing rate.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102941 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_tenant_success_total{job=~".*frontend.*"}[5m]))
frontend: tenant_error_rate
Rate of tenant processing errors
The rate of tenant processing operations that result in errors. Consistent errors indicate problems with specific tenants.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102950 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_tenant_errors_total{job=~".*frontend.*"}[5m]))
frontend: tenant_error_percentage
Percentage of tenant operations resulting in errors
The percentage of tenant operations that result in errors. Values above 5% indicate significant tenant processing problems.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102951 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum by (name, job_name) (rate(src_periodic_goroutine_tenant_errors_total{job=~".*frontend.*"}[5m])) / (sum by (name, job_name) (rate(src_periodic_goroutine_tenant_success_total{job=~".*frontend.*"}[5m])) + sum by (name, job_name) (rate(src_periodic_goroutine_tenant_errors_total{job=~".*frontend.*"}[5m])))) * 100
Frontend: Database connections
frontend: max_open_conns
Maximum open
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_max_open{app_name="frontend"})
frontend: open_conns
Established
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_open{app_name="frontend"})
frontend: in_use
Used
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_in_use{app_name="frontend"})
frontend: idle
Idle
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_idle{app_name="frontend"})
frontend: mean_blocked_seconds_per_conn_request
Mean blocked seconds per conn request
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103020 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_blocked_seconds{app_name="frontend"}[5m])) / sum by (app_name, db_name) (increase(src_pgsql_conns_waited_for{app_name="frontend"}[5m]))
frontend: closed_max_idle
Closed by SetMaxIdleConns
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103030 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle{app_name="frontend"}[5m]))
frontend: closed_max_lifetime
Closed by SetConnMaxLifetime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103031 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_lifetime{app_name="frontend"}[5m]))
frontend: closed_max_idle_time
Closed by SetConnMaxIdleTime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103032 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle_time{app_name="frontend"}[5m]))
Frontend: (frontend|sourcegraph-frontend) (CPU, Memory)
frontend: cpu_usage_percentage
CPU usage
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103100 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}
frontend: memory_usage_percentage
Memory usage percentage (total)
An estimate for the active memory in use, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103101 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}
frontend: memory_working_set_bytes
Memory usage bytes (total)
An estimate for the active memory in use in bytes, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103102 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_memory_working_set_bytes{name=~"^(frontend|sourcegraph-frontend).*"})
frontend: memory_rss
Memory (RSS)
The total anonymous memory in use by the application, which includes Go stack and heap. This memory is is non-reclaimable, and high usage may trigger OOM kills. Note: the metric is named RSS because to match the cadvisor name, but anonymous is more accurate."
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103110 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_rss{name=~"^(frontend|sourcegraph-frontend).*"} / container_spec_memory_limit_bytes{name=~"^(frontend|sourcegraph-frontend).*"}) by (name) * 100.0
frontend: memory_total_active_file
Memory usage (active file)
This metric shows the total active file-backed memory currently in use by the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103111 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_total_active_file_bytes{name=~"^(frontend|sourcegraph-frontend).*"} / container_spec_memory_limit_bytes{name=~"^(frontend|sourcegraph-frontend).*"}) by (name) * 100.0
frontend: memory_kernel_usage
Memory usage (kernel)
The kernel usage metric shows the amount of memory used by the kernel on behalf of the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103112 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_kernel_usage{name=~"^(frontend|sourcegraph-frontend).*"} / container_spec_memory_limit_bytes{name=~"^(frontend|sourcegraph-frontend).*"}) by (name) * 100.0
Frontend: Container monitoring (not available on server)
frontend: container_missing
Container missing
This value is the number of times a container has not been seen for more than one minute. If you observe this value change independent of deployment events (such as an upgrade), it could indicate pods are being OOM killed or terminated for some other reasons.
- Kubernetes:
- Determine if the pod was OOM killed using
kubectl describe pod (frontend\|sourcegraph-frontend)(look forOOMKilled: true) and, if so, consider increasing the memory limit in the relevantDeployment.yaml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingkubectl logs -p (frontend\|sourcegraph-frontend).
- Determine if the pod was OOM killed using
- Docker Compose:
- Determine if the pod was OOM killed using
docker inspect -f '\{\{json .State\}\}' (frontend\|sourcegraph-frontend)(look for"OOMKilled":true) and, if so, consider increasing the memory limit of the (frontend|sourcegraph-frontend) container indocker-compose.yml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingdocker logs (frontend\|sourcegraph-frontend)(note this will include logs from the previous and currently running container).
- Determine if the pod was OOM killed using
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103200 on your Sourcegraph instance.
Technical details
Query:
SHELLcount by(name) ((time() - container_last_seen{name=~"^(frontend|sourcegraph-frontend).*"}) > 60)
frontend: container_cpu_usage
Container cpu usage total (1m average) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103201 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}
frontend: container_memory_usage
Container memory usage by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103202 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}
frontend: fs_io_operations
Filesystem reads and writes rate by instance over 1h
This value indicates the number of filesystem read and write operations by containers of this service. When extremely high, this can indicate a resource usage problem, or can cause problems with the service itself, especially if high values or spikes correlate with {{CONTAINER_NAME}} issues.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(name) (rate(container_fs_reads_total{name=~"^(frontend|sourcegraph-frontend).*"}[1h]) + rate(container_fs_writes_total{name=~"^(frontend|sourcegraph-frontend).*"}[1h]))
Frontend: Provisioning indicators (not available on server)
frontend: provisioning_container_cpu_usage_long_term
Container cpu usage total (90th percentile over 1d) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103300 on your Sourcegraph instance.
Technical details
Query:
SHELLquantile_over_time(0.9, cadvisor_container_cpu_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}[1d])
frontend: provisioning_container_memory_usage_long_term
Container memory usage (1d maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103301 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}[1d])
frontend: provisioning_container_cpu_usage_short_term
Container cpu usage total (5m maximum) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103310 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}[5m])
frontend: provisioning_container_memory_usage_short_term
Container memory usage (5m maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103311 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}[5m])
frontend: container_oomkill_events_total
Container OOMKILL events total by instance
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103312 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^(frontend|sourcegraph-frontend).*"})
Frontend: Golang runtime monitoring
frontend: go_goroutines
Maximum active goroutines
A high value here indicates a possible goroutine leak.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103400 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_goroutines{job=~".*(frontend|sourcegraph-frontend)"})
frontend: go_gc_duration_seconds
Maximum go garbage collection duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103401 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_gc_duration_seconds{job=~".*(frontend|sourcegraph-frontend)"})
Frontend: Kubernetes monitoring (only available on Kubernetes)
frontend: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*(frontend|sourcegraph-frontend)"}) / count by (app) (up{app=~".*(frontend|sourcegraph-frontend)"}) * 100
Frontend: Search: Ranking
frontend: total_search_clicks
Total number of search clicks over 6h
The total number of search clicks across all search types over a 6 hour window.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (ranked) (increase(src_search_ranking_result_clicked_count[6h]))
frontend: percent_clicks_on_top_search_result
Percent of clicks on top search result over 6h
The percent of clicks that were on the top search result, excluding searches with very few results (3 or fewer).
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103601 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (ranked) (increase(src_search_ranking_result_clicked_bucket{le="1",resultsLength=">3"}[6h])) / sum by (ranked) (increase(src_search_ranking_result_clicked_count[6h])) * 100
frontend: percent_clicks_on_top_3_search_results
Percent of clicks on top 3 search results over 6h
The percent of clicks that were on the first 3 search results, excluding searches with very few results (3 or fewer).
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103602 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (ranked) (increase(src_search_ranking_result_clicked_bucket{le="3",resultsLength=">3"}[6h])) / sum by (ranked) (increase(src_search_ranking_result_clicked_count[6h])) * 100
frontend: distribution_of_clicked_search_result_type_over_6h_in_percent
Distribution of clicked search result type over 6h
The distribution of clicked search results by result type. At every point in time, the values should sum to 100.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103610 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_search_ranking_result_clicked_count{type="repo"}[6h])) / sum(increase(src_search_ranking_result_clicked_count[6h])) * 100
frontend: percent_zoekt_searches_hitting_flush_limit
Percent of zoekt searches that hit the flush time limit
The percent of Zoekt searches that hit the flush time limit. These searches don`t visit all matches, so they could be missing relevant results, or be non-deterministic.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103611 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(zoekt_final_aggregate_size_count{reason="timer_expired"}[1d])) / sum(increase(zoekt_final_aggregate_size_count[1d])) * 100
Frontend: Email delivery
frontend: email_delivery_failures
Email delivery failure rate over 30 minutes
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103700 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_email_send{success="false"}[30m])) / sum(increase(src_email_send[30m])) * 100
frontend: email_deliveries_total
Total emails successfully delivered every 30 minutes
Total emails successfully delivered.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103710 on your Sourcegraph instance.
Technical details
Query:
SHELLsum (increase(src_email_send{success="true"}[30m]))
frontend: email_deliveries_by_source
Emails successfully delivered every 30 minutes by source
Emails successfully delivered by source, i.e. product feature.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103711 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (email_source) (increase(src_email_send{success="true"}[30m]))
Frontend: Sentinel queries (only on sourcegraph.com)
frontend: mean_successful_sentinel_duration_over_2h
Mean successful sentinel search duration over 2h
Mean search duration for all successful sentinel queries
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103800 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_search_response_latency_seconds_sum{source=~`searchblitz.*`, status=`success`}[2h])) / sum(rate(src_search_response_latency_seconds_count{source=~`searchblitz.*`, status=`success`}[2h]))
frontend: mean_sentinel_stream_latency_over_2h
Mean successful sentinel stream latency over 2h
Mean time to first result for all successful streaming sentinel queries
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103801 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_search_streaming_latency_seconds_sum{source=~"searchblitz.*"}[2h])) / sum(rate(src_search_streaming_latency_seconds_count{source=~"searchblitz.*"}[2h]))
frontend: 90th_percentile_successful_sentinel_duration_over_2h
90th percentile successful sentinel search duration over 2h
90th percentile search duration for all successful sentinel queries
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103810 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le)(label_replace(rate(src_search_response_latency_seconds_bucket{source=~"searchblitz.*", status="success"}[2h]), "source", "$1", "source", "searchblitz_(.*)")))
frontend: 90th_percentile_sentinel_stream_latency_over_2h
90th percentile successful sentinel stream latency over 2h
90th percentile time to first result for all successful streaming sentinel queries
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103811 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le)(label_replace(rate(src_search_streaming_latency_seconds_bucket{source=~"searchblitz.*"}[2h]), "source", "$1", "source", "searchblitz_(.*)")))
frontend: mean_successful_sentinel_duration_by_query
Mean successful sentinel search duration by query
Mean search duration for successful sentinel queries, broken down by query. Useful for debugging whether a slowdown is limited to a specific type of query.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103820 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_search_response_latency_seconds_sum{source=~"searchblitz.*", status="success"}[$sentinel_sampling_duration])) by (source) / sum(rate(src_search_response_latency_seconds_count{source=~"searchblitz.*", status="success"}[$sentinel_sampling_duration])) by (source)
frontend: mean_sentinel_stream_latency_by_query
Mean successful sentinel stream latency by query
Mean time to first result for successful streaming sentinel queries, broken down by query. Useful for debugging whether a slowdown is limited to a specific type of query.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103821 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_search_streaming_latency_seconds_sum{source=~"searchblitz.*"}[$sentinel_sampling_duration])) by (source) / sum(rate(src_search_streaming_latency_seconds_count{source=~"searchblitz.*"}[$sentinel_sampling_duration])) by (source)
frontend: 90th_percentile_successful_sentinel_duration_by_query
90th percentile successful sentinel search duration by query
90th percentile search duration for successful sentinel queries, broken down by query. Useful for debugging whether a slowdown is limited to a specific type of query.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103830 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum(rate(src_search_response_latency_seconds_bucket{source=~"searchblitz.*", status="success"}[$sentinel_sampling_duration])) by (le, source))
frontend: 90th_percentile_successful_stream_latency_by_query
90th percentile successful sentinel stream latency by query
90th percentile time to first result for successful streaming sentinel queries, broken down by query. Useful for debugging whether a slowdown is limited to a specific type of query.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103831 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum(rate(src_search_streaming_latency_seconds_bucket{source=~"searchblitz.*"}[$sentinel_sampling_duration])) by (le, source))
frontend: 90th_percentile_unsuccessful_duration_by_query
90th percentile unsuccessful sentinel search duration by query
90th percentile search duration of unsuccessful sentinel queries (by error or timeout), broken down by query. Useful for debugging how the performance of failed requests affect UX.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103840 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum(rate(src_search_response_latency_seconds_bucket{source=~`searchblitz.*`, status!=`success`}[$sentinel_sampling_duration])) by (le, source))
frontend: 75th_percentile_successful_sentinel_duration_by_query
75th percentile successful sentinel search duration by query
75th percentile search duration of successful sentinel queries, broken down by query. Useful for debugging whether a slowdown is limited to a specific type of query.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103850 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_search_response_latency_seconds_bucket{source=~"searchblitz.*", status="success"}[$sentinel_sampling_duration])) by (le, source))
frontend: 75th_percentile_successful_stream_latency_by_query
75th percentile successful sentinel stream latency by query
75th percentile time to first result for successful streaming sentinel queries, broken down by query. Useful for debugging whether a slowdown is limited to a specific type of query.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103851 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_search_streaming_latency_seconds_bucket{source=~"searchblitz.*"}[$sentinel_sampling_duration])) by (le, source))
frontend: 75th_percentile_unsuccessful_duration_by_query
75th percentile unsuccessful sentinel search duration by query
75th percentile search duration of unsuccessful sentinel queries (by error or timeout), broken down by query. Useful for debugging how the performance of failed requests affect UX.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103860 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_search_response_latency_seconds_bucket{source=~`searchblitz.*`, status!=`success`}[$sentinel_sampling_duration])) by (le, source))
frontend: unsuccessful_status_rate
Unsuccessful status rate
The rate of unsuccessful sentinel queries, broken down by failure type.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103870 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_graphql_search_response{source=~"searchblitz.*", status!="success"}[$sentinel_sampling_duration])) by (status)
Frontend: Incoming webhooks
frontend: p95_time_to_handle_incoming_webhooks
P95 time to handle incoming webhooks
p95 response time to incoming webhook requests from code hosts.
Increases in response time can point to too much load on the database to keep up with the incoming requests.
See this documentation page for more details on webhook requests: (https://sourcegraph.com/docs/admin/config/webhooks/incoming)
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103900 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum (rate(src_http_request_duration_seconds_bucket{route=~"webhooks|github.webhooks|gitlab.webhooks|bitbucketServer.webhooks|bitbucketCloud.webhooks"}[5m])) by (le, route))
Frontend: Search aggregations: proactive and expanded search aggregations
frontend: insights_aggregations_total
Aggregate search aggregations operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_insights_aggregations_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: insights_aggregations_99th_percentile_duration
Aggregate successful search aggregations operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_insights_aggregations_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: insights_aggregations_errors_total
Aggregate search aggregations operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104002 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_insights_aggregations_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: insights_aggregations_error_rate
Aggregate search aggregations operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104003 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_insights_aggregations_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_insights_aggregations_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_insights_aggregations_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: insights_aggregations_total
Search aggregations operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,extended_mode)(increase(src_insights_aggregations_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: insights_aggregations_99th_percentile_duration
99th percentile successful search aggregations operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104011 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op,extended_mode)(rate(src_insights_aggregations_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: insights_aggregations_errors_total
Search aggregations operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104012 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,extended_mode)(increase(src_insights_aggregations_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: insights_aggregations_error_rate
Search aggregations operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104013 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,extended_mode)(increase(src_insights_aggregations_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op,extended_mode)(increase(src_insights_aggregations_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op,extended_mode)(increase(src_insights_aggregations_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Git Server
Stores, manages, and operates Git repositories.
To see this dashboard, visit /-/debug/grafana/d/gitserver/gitserver on your Sourcegraph instance.
gitserver: go_routines
Go routines
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLgo_goroutines{app="gitserver", instance=~`${shard:regex}`}
gitserver: disk_space_remaining
Disk space remaining
Indicates disk space remaining for each gitserver instance. When disk space is low, gitserver may experience slowdowns or fails to fetch repositories.
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELL(src_gitserver_disk_space_available{instance=~`${shard:regex}`} / src_gitserver_disk_space_total{instance=~`${shard:regex}`}) * 100
gitserver: cpu_throttling_time
Container CPU throttling time %
- A high value indicates that the container is spending too much time waiting for CPU cycles.
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) ((rate(container_cpu_cfs_throttled_periods_total{container_label_io_kubernetes_container_name="gitserver", container_label_io_kubernetes_pod_name=~`${shard:regex}`}[5m]) / rate(container_cpu_cfs_periods_total{container_label_io_kubernetes_container_name="gitserver", container_label_io_kubernetes_pod_name=~`${shard:regex}`}[5m])) * 100)
gitserver: cpu_usage_seconds
Cpu usage seconds
- This value should not exceed 75% of the CPU limit over a longer period of time.
-
We cannot alert on this as we don`t know the resource allocation.
-
If this value is high for a longer time, consider increasing the CPU limit for the container.
-
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_cpu_usage_seconds_total{container_label_io_kubernetes_container_name="gitserver", container_label_io_kubernetes_pod_name=~`${shard:regex}`}[5m]))
gitserver: memory_major_page_faults
Gitserver page faults
The number of major page faults in a 5 minute window for gitserver. If this number increases significantly, it indicates that more git API calls need to load data from disk. There may not be enough memory to efficiently support the amount of API requests served concurrently.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100020 on your Sourcegraph instance.
Technical details
Query:
SHELLrate(container_memory_failures_total{failure_type="pgmajfault", name=~"^gitserver.*"}[5m])
gitserver: high_memory_git_commands
Number of git commands that exceeded the threshold for high memory usage
This graph tracks the number of git subcommands that gitserver ran that exceeded the threshold for high memory usage. This graph in itself is not an alert, but it is used to learn about the memory usage of gitserver.
If gitserver frequently serves requests where the status code is KILLED, this graph might help to correlate that with the high memory usage.
This graph spiking is not a problem necessarily. But when subcommands or the whole gitserver service are getting OOM killed and this graph shows spikes, increasing the memory might be useful.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100021 on your Sourcegraph instance.
Technical details
Query:
SHELLsort_desc(sum(sum_over_time(src_gitserver_exec_high_memory_usage_count{instance=~`${shard:regex}`}[2m])) by (cmd))
gitserver: running_git_commands
Git commands running on each gitserver instance
A high value signals load.
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100030 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance, cmd) (src_gitserver_exec_running{instance=~`${shard:regex}`})
gitserver: git_commands_received
Rate of git commands received
per second rate per command
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100031 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (cmd) (rate(src_gitserver_exec_duration_seconds_count{instance=~`${shard:regex}`}[5m]))
gitserver: echo_command_duration_test
Echo test command duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100040 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_gitserver_echo_duration_seconds)
gitserver: repo_corrupted
Number of times a repo corruption has been identified
A non-null value here indicates that a problem has been detected with the gitserver repository storage. Repository corruptions are never expected. This is a real issue. Gitserver should try to recover from them by recloning repositories, but this may take a while depending on repo size.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100041 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_repo_corrupted[5m]))
gitserver: repository_clone_queue_size
Repository clone queue size
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100050 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_gitserver_clone_queue)
gitserver: src_gitserver_client_concurrent_requests
Number of concurrent requests running against gitserver client
This metric is only for informational purposes. It indicates the current number of concurrently running requests by process against gitserver gRPC.
It does not indicate any problems with the instance, but can give a good indication of load spikes or request throttling.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100051 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (job, instance) (src_gitserver_client_concurrent_requests)
Git Server: Gitservice for internal cloning
gitserver: gitservice_request_duration
95th percentile gitservice request duration per shard
A high value means any internal service trying to clone a repo from gitserver is slowed down.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_gitserver_gitservice_duration_seconds_bucket{instance=~`${shard:regex}`}[5m])) by (le, gitservice))
gitserver: gitservice_request_rate
Gitservice request rate per shard
Per shard gitservice request rate
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_gitservice_duration_seconds_count{instance=~`${shard:regex}`}[5m])) by (gitservice)
gitserver: gitservice_requests_running
Gitservice requests running per shard
Per shard gitservice requests running
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_gitserver_gitservice_running{instance=~`${shard:regex}`}) by (gitservice)
Git Server: Gitserver cleanup jobs
gitserver: janitor_tasks_total
Total housekeeping tasks by type and status
The rate of housekeeping tasks performed in repositories, broken down by task type and success/failure status
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_janitor_tasks_total{instance=~`${shard:regex}`}[5m])) by (housekeeping_task, status)
gitserver: p90_janitor_tasks_latency_success_over_5m
90th percentile latency of successful tasks by type over 5m
The 90th percentile latency of successful housekeeping tasks, broken down by task type
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum(rate(src_gitserver_janitor_tasks_latency_bucket{instance=~`${shard:regex}`, status="success"}[5m])) by (le, housekeeping_task))
gitserver: p95_janitor_tasks_latency_success_over_5m
95th percentile latency of successful tasks by type over 5m
The 95th percentile latency of successful housekeeping tasks, broken down by task type
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_gitserver_janitor_tasks_latency_bucket{instance=~`${shard:regex}`, status="success"}[5m])) by (le, housekeeping_task))
gitserver: p99_janitor_tasks_latency_success_over_5m
99th percentile latency of successful tasks by type over 5m
The 99th percentile latency of successful housekeeping tasks, broken down by task type
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_gitserver_janitor_tasks_latency_bucket{instance=~`${shard:regex}`, status="success"}[5m])) by (le, housekeeping_task))
gitserver: p90_janitor_tasks_latency_failure_over_5m
90th percentile latency of failed tasks by type over 5m
The 90th percentile latency of failed housekeeping tasks, broken down by task type
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100220 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum(rate(src_gitserver_janitor_tasks_latency_bucket{instance=~`${shard:regex}`, status="failure"}[5m])) by (le, housekeeping_task))
gitserver: p95_janitor_tasks_latency_failure_over_5m
95th percentile latency of failed tasks by type over 5m
The 95th percentile latency of failed housekeeping tasks, broken down by task type
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100221 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_gitserver_janitor_tasks_latency_bucket{instance=~`${shard:regex}`, status="failure"}[5m])) by (le, housekeeping_task))
gitserver: p99_janitor_tasks_latency_failure_over_5m
99th percentile latency of failed tasks by type over 5m
The 99th percentile latency of failed housekeeping tasks, broken down by task type
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100222 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_gitserver_janitor_tasks_latency_bucket{instance=~`${shard:regex}`, status="failure"}[5m])) by (le, housekeeping_task))
gitserver: pruned_files_total_over_5m
Files pruned by type over 5m
The rate of files pruned during cleanup, broken down by file type
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100230 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_janitor_pruned_files_total{instance=~`${shard:regex}`}[5m])) by (filetype)
gitserver: data_structure_count_over_5m
Data structure counts over 5m
The count distribution of various Git data structures in repositories
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100240 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_gitserver_janitor_data_structure_count_bucket{instance=~`${shard:regex}`}[5m])) by (le, data_structure))
gitserver: janitor_data_structure_size
Data structure sizes
The size distribution of various Git data structures in repositories
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100250 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_gitserver_janitor_data_structure_size_bucket{instance=~`${shard:regex}`}[5m])) by (le, data_structure))
gitserver: janitor_time_since_optimization
Time since last optimization
The time elapsed since last optimization of various Git data structures
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100260 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_gitserver_janitor_time_since_last_optimization_seconds_bucket{instance=~`${shard:regex}`}[5m])) by (le, data_structure))
gitserver: janitor_data_structure_existence
Data structure existence
The rate at which data structures are reported to exist in repositories
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100270 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_janitor_data_structure_existence_total{instance=~`${shard:regex}`, exists="true"}[5m])) by (data_structure)
Git Server: Git Command Corruption Retries
gitserver: git_command_retry_attempts_rate
Rate of git command corruption retry attempts over 5m
The rate of git command retry attempts due to corruption detection. A non-zero value indicates that gitserver is detecting potential corruption and attempting retries. This metric helps track how often the retry mechanism is triggered.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_retry_attempts_total{instance=~`${shard:regex}`}[5m]))
gitserver: git_command_retry_success_rate
Rate of successful git command corruption retries over 5m
The rate of git commands that succeeded after retry attempts. This indicates how effective the retry mechanism is at resolving transient corruption issues.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_retry_success_total{instance=~`${shard:regex}`}[5m]))
gitserver: git_command_retry_failure_rate
Rate of failed git command corruption retries over 5m
The rate of git commands that failed even after all retry attempts were exhausted. These failures will result in repository corruption marking and potential recloning.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100310 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_retry_failure_total{instance=~`${shard:regex}`}[5m]))
gitserver: git_command_retry_different_error_rate
Rate of corruption retries that failed with non-corruption errors over 5m
The rate of retry attempts that failed with errors other than corruption. This indicates that repository state or environment changed between the original command and retry attempt. Common causes include network issues, permission changes, or concurrent repository modifications.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100311 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_retry_different_error_total{instance=~`${shard:regex}`}[5m]))
gitserver: git_command_retry_success_ratio
Ratio of successful corruption retries to total corruption retry attempts over 5m
The percentage of retry attempts that ultimately succeeded. A high ratio indicates that most corruption errors are transient and resolved by retries. A low ratio may indicate persistent corruption issues requiring investigation.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100312 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_retry_success_total{instance=~`${shard:regex}`}[5m])) / sum(rate(src_gitserver_retry_attempts_total{instance=~`${shard:regex}`}[5m]))
Git Server: Periodic Goroutines
gitserver: running_goroutines
Number of currently running periodic goroutines
The number of currently running periodic goroutines by name and job.
A value of 0 indicates the routine isnt running currently, it awaits its next schedule.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (src_periodic_goroutine_running{job=~".*gitserver.*"})
gitserver: goroutine_success_rate
Success rate for periodic goroutine executions
The rate of successful executions of each periodic goroutine. A low or zero value could indicate that a routine is stalled or encountering errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_total{job=~".*gitserver.*"}[5m]))
gitserver: goroutine_error_rate
Error rate for periodic goroutine executions
The rate of errors encountered by each periodic goroutine. A sustained high error rate may indicate a problem with the routine`s configuration or dependencies.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100410 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_errors_total{job=~".*gitserver.*"}[5m]))
gitserver: goroutine_error_percentage
Percentage of periodic goroutine executions that result in errors
The percentage of executions that result in errors for each periodic goroutine. A value above 5% indicates that a significant portion of routine executions are failing.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100411 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_errors_total{job=~".*gitserver.*"}[5m])) / sum by (name, job_name) (rate(src_periodic_goroutine_total{job=~".*gitserver.*"}[5m]) > 0) * 100
gitserver: goroutine_handler_duration
95th percentile handler execution time
The 95th percentile execution time for each periodic goroutine handler. Longer durations might indicate increased load or processing time.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100420 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job_name, le) (rate(src_periodic_goroutine_duration_seconds_bucket{job=~".*gitserver.*"}[5m])))
gitserver: goroutine_loop_duration
95th percentile loop cycle time
The 95th percentile loop cycle time for each periodic goroutine (excluding sleep time). This represents how long a complete loop iteration takes before sleeping for the next interval.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100421 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job_name, le) (rate(src_periodic_goroutine_loop_duration_seconds_bucket{job=~".*gitserver.*"}[5m])))
gitserver: tenant_processing_duration
95th percentile tenant processing time
The 95th percentile processing time for individual tenants within periodic goroutines. Higher values indicate that tenant processing is taking longer and may affect overall performance.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100430 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job_name, le) (rate(src_periodic_goroutine_tenant_duration_seconds_bucket{job=~".*gitserver.*"}[5m])))
gitserver: tenant_processing_max
Maximum tenant processing time
The maximum processing time for individual tenants within periodic goroutines. Consistently high values might indicate problematic tenants or inefficient processing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100431 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name, job_name) (rate(src_periodic_goroutine_tenant_duration_seconds_sum{job=~".*gitserver.*"}[5m]) / rate(src_periodic_goroutine_tenant_duration_seconds_count{job=~".*gitserver.*"}[5m]))
gitserver: tenant_count
Number of tenants processed per routine
The number of tenants processed by each periodic goroutine. Unexpected changes can indicate tenant configuration issues or scaling events.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100440 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name, job_name) (src_periodic_goroutine_tenant_count{job=~".*gitserver.*"})
gitserver: tenant_success_rate
Rate of successful tenant processing operations
The rate of successful tenant processing operations. A healthy routine should maintain a consistent processing rate.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100441 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_tenant_success_total{job=~".*gitserver.*"}[5m]))
gitserver: tenant_error_rate
Rate of tenant processing errors
The rate of tenant processing operations that result in errors. Consistent errors indicate problems with specific tenants.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100450 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_tenant_errors_total{job=~".*gitserver.*"}[5m]))
gitserver: tenant_error_percentage
Percentage of tenant operations resulting in errors
The percentage of tenant operations that result in errors. Values above 5% indicate significant tenant processing problems.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100451 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum by (name, job_name) (rate(src_periodic_goroutine_tenant_errors_total{job=~".*gitserver.*"}[5m])) / (sum by (name, job_name) (rate(src_periodic_goroutine_tenant_success_total{job=~".*gitserver.*"}[5m])) + sum by (name, job_name) (rate(src_periodic_goroutine_tenant_errors_total{job=~".*gitserver.*"}[5m])))) * 100
Git Server: Gitserver (CPU, Memory)
gitserver: cpu_usage_percentage
CPU usage
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^gitserver.*"}
gitserver: memory_usage_percentage
Memory usage percentage (total)
An estimate for the active memory in use, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100501 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^gitserver.*"}
gitserver: memory_working_set_bytes
Memory usage bytes (total)
An estimate for the active memory in use in bytes, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100502 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_memory_working_set_bytes{name=~"^gitserver.*"})
gitserver: memory_rss
Memory (RSS)
The total anonymous memory in use by the application, which includes Go stack and heap. This memory is is non-reclaimable, and high usage may trigger OOM kills. Note: the metric is named RSS because to match the cadvisor name, but anonymous is more accurate."
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100510 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_rss{name=~"^gitserver.*"} / container_spec_memory_limit_bytes{name=~"^gitserver.*"}) by (name) * 100.0
gitserver: memory_total_active_file
Memory usage (active file)
This metric shows the total active file-backed memory currently in use by the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100511 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_total_active_file_bytes{name=~"^gitserver.*"} / container_spec_memory_limit_bytes{name=~"^gitserver.*"}) by (name) * 100.0
gitserver: memory_kernel_usage
Memory usage (kernel)
The kernel usage metric shows the amount of memory used by the kernel on behalf of the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100512 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_kernel_usage{name=~"^gitserver.*"} / container_spec_memory_limit_bytes{name=~"^gitserver.*"}) by (name) * 100.0
Git Server: Network I/O pod metrics (only available on Kubernetes)
gitserver: network_sent_bytes_aggregate
Transmission rate over 5m (aggregate)
The rate of bytes sent over the network across all pods
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(container_network_transmit_bytes_total{container_label_io_kubernetes_pod_name=~`.*gitserver.*`}[5m]))
gitserver: network_received_packets_per_instance
Transmission rate over 5m (per instance)
The amount of bytes sent over the network by individual pods
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100601 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_network_transmit_bytes_total{container_label_io_kubernetes_pod_name=~`${instance:regex}`}[5m]))
gitserver: network_received_bytes_aggregate
Receive rate over 5m (aggregate)
The amount of bytes received from the network across pods
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100610 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(container_network_receive_bytes_total{container_label_io_kubernetes_pod_name=~`.*gitserver.*`}[5m]))
gitserver: network_received_bytes_per_instance
Receive rate over 5m (per instance)
The amount of bytes received from the network by individual pods
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100611 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_network_receive_bytes_total{container_label_io_kubernetes_pod_name=~`${instance:regex}`}[5m]))
gitserver: network_transmitted_packets_dropped_by_instance
Transmit packet drop rate over 5m (by instance)
An increase in dropped packets could be a leading indicator of network saturation.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100620 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_network_transmit_packets_dropped_total{container_label_io_kubernetes_pod_name=~`${instance:regex}`}[5m]))
gitserver: network_transmitted_packets_errors_per_instance
Errors encountered while transmitting over 5m (per instance)
An increase in transmission errors could indicate a networking issue
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100621 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_network_transmit_errors_total{container_label_io_kubernetes_pod_name=~`${instance:regex}`}[5m]))
gitserver: network_received_packets_dropped_by_instance
Receive packet drop rate over 5m (by instance)
An increase in dropped packets could be a leading indicator of network saturation.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100622 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_network_receive_packets_dropped_total{container_label_io_kubernetes_pod_name=~`${instance:regex}`}[5m]))
gitserver: network_transmitted_packets_errors_by_instance
Errors encountered while receiving over 5m (per instance)
An increase in errors while receiving could indicate a networking issue.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100623 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_network_receive_errors_total{container_label_io_kubernetes_pod_name=~`${instance:regex}`}[5m]))
Git Server: VCS Clone metrics
gitserver: vcs_syncer_999_successful_clone_duration
99.9th percentile successful Clone duration over 1m
The 99.9th percentile duration for successful Clone VCS operations. This is the time taken to clone a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100700 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (type, le) (rate(vcssyncer_clone_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="true"}[1m])))
gitserver: vcs_syncer_99_successful_clone_duration
99th percentile successful Clone duration over 1m
The 99th percentile duration for successful Clone VCS operations. This is the time taken to clone a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100701 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (type, le) (rate(vcssyncer_clone_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="true"}[1m])))
gitserver: vcs_syncer_95_successful_clone_duration
95th percentile successful Clone duration over 1m
The 95th percentile duration for successful Clone VCS operations. This is the time taken to clone a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100702 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (type, le) (rate(vcssyncer_clone_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="true"}[1m])))
gitserver: vcs_syncer_successful_clone_rate
Rate of successful Clone VCS operations over 1m
The rate of successful Clone VCS operations. This is the time taken to clone a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100710 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (type) (rate(vcssyncer_clone_duration_seconds_count{type=~`${vcsSyncerType:regex}`, success="true"}[1m]))
gitserver: vcs_syncer_999_failed_clone_duration
99.9th percentile failed Clone duration over 1m
The 99.9th percentile duration for failed Clone VCS operations. This is the time taken to clone a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100720 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (type, le) (rate(vcssyncer_clone_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="false"}[1m])))
gitserver: vcs_syncer_99_failed_clone_duration
99th percentile failed Clone duration over 1m
The 99th percentile duration for failed Clone VCS operations. This is the time taken to clone a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100721 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (type, le) (rate(vcssyncer_clone_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="false"}[1m])))
gitserver: vcs_syncer_95_failed_clone_duration
95th percentile failed Clone duration over 1m
The 95th percentile duration for failed Clone VCS operations. This is the time taken to clone a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100722 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (type, le) (rate(vcssyncer_clone_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="false"}[1m])))
gitserver: vcs_syncer_failed_clone_rate
Rate of failed Clone VCS operations over 1m
The rate of failed Clone VCS operations. This is the time taken to clone a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100730 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (type) (rate(vcssyncer_clone_duration_seconds_count{type=~`${vcsSyncerType:regex}`, success="false"}[1m]))
Git Server: VCS Fetch metrics
gitserver: vcs_syncer_999_successful_fetch_duration
99.9th percentile successful Fetch duration over 1m
The 99.9th percentile duration for successful Fetch VCS operations. This is the time taken to fetch a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100800 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (type, le) (rate(vcssyncer_fetch_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="true"}[1m])))
gitserver: vcs_syncer_99_successful_fetch_duration
99th percentile successful Fetch duration over 1m
The 99th percentile duration for successful Fetch VCS operations. This is the time taken to fetch a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100801 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (type, le) (rate(vcssyncer_fetch_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="true"}[1m])))
gitserver: vcs_syncer_95_successful_fetch_duration
95th percentile successful Fetch duration over 1m
The 95th percentile duration for successful Fetch VCS operations. This is the time taken to fetch a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100802 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (type, le) (rate(vcssyncer_fetch_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="true"}[1m])))
gitserver: vcs_syncer_successful_fetch_rate
Rate of successful Fetch VCS operations over 1m
The rate of successful Fetch VCS operations. This is the time taken to fetch a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100810 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (type) (rate(vcssyncer_fetch_duration_seconds_count{type=~`${vcsSyncerType:regex}`, success="true"}[1m]))
gitserver: vcs_syncer_999_failed_fetch_duration
99.9th percentile failed Fetch duration over 1m
The 99.9th percentile duration for failed Fetch VCS operations. This is the time taken to fetch a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100820 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (type, le) (rate(vcssyncer_fetch_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="false"}[1m])))
gitserver: vcs_syncer_99_failed_fetch_duration
99th percentile failed Fetch duration over 1m
The 99th percentile duration for failed Fetch VCS operations. This is the time taken to fetch a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100821 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (type, le) (rate(vcssyncer_fetch_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="false"}[1m])))
gitserver: vcs_syncer_95_failed_fetch_duration
95th percentile failed Fetch duration over 1m
The 95th percentile duration for failed Fetch VCS operations. This is the time taken to fetch a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100822 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (type, le) (rate(vcssyncer_fetch_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="false"}[1m])))
gitserver: vcs_syncer_failed_fetch_rate
Rate of failed Fetch VCS operations over 1m
The rate of failed Fetch VCS operations. This is the time taken to fetch a repository from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100830 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (type) (rate(vcssyncer_fetch_duration_seconds_count{type=~`${vcsSyncerType:regex}`, success="false"}[1m]))
Git Server: VCS Is_cloneable metrics
gitserver: vcs_syncer_999_successful_is_cloneable_duration
99.9th percentile successful Is_cloneable duration over 1m
The 99.9th percentile duration for successful Is_cloneable VCS operations. This is the time taken to check to see if a repository is cloneable from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100900 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (type, le) (rate(vcssyncer_is_cloneable_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="true"}[1m])))
gitserver: vcs_syncer_99_successful_is_cloneable_duration
99th percentile successful Is_cloneable duration over 1m
The 99th percentile duration for successful Is_cloneable VCS operations. This is the time taken to check to see if a repository is cloneable from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100901 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (type, le) (rate(vcssyncer_is_cloneable_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="true"}[1m])))
gitserver: vcs_syncer_95_successful_is_cloneable_duration
95th percentile successful Is_cloneable duration over 1m
The 95th percentile duration for successful Is_cloneable VCS operations. This is the time taken to check to see if a repository is cloneable from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100902 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (type, le) (rate(vcssyncer_is_cloneable_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="true"}[1m])))
gitserver: vcs_syncer_successful_is_cloneable_rate
Rate of successful Is_cloneable VCS operations over 1m
The rate of successful Is_cloneable VCS operations. This is the time taken to check to see if a repository is cloneable from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100910 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (type) (rate(vcssyncer_is_cloneable_duration_seconds_count{type=~`${vcsSyncerType:regex}`, success="true"}[1m]))
gitserver: vcs_syncer_999_failed_is_cloneable_duration
99.9th percentile failed Is_cloneable duration over 1m
The 99.9th percentile duration for failed Is_cloneable VCS operations. This is the time taken to check to see if a repository is cloneable from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100920 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (type, le) (rate(vcssyncer_is_cloneable_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="false"}[1m])))
gitserver: vcs_syncer_99_failed_is_cloneable_duration
99th percentile failed Is_cloneable duration over 1m
The 99th percentile duration for failed Is_cloneable VCS operations. This is the time taken to check to see if a repository is cloneable from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100921 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (type, le) (rate(vcssyncer_is_cloneable_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="false"}[1m])))
gitserver: vcs_syncer_95_failed_is_cloneable_duration
95th percentile failed Is_cloneable duration over 1m
The 95th percentile duration for failed Is_cloneable VCS operations. This is the time taken to check to see if a repository is cloneable from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100922 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (type, le) (rate(vcssyncer_is_cloneable_duration_seconds_bucket{type=~`${vcsSyncerType:regex}`, success="false"}[1m])))
gitserver: vcs_syncer_failed_is_cloneable_rate
Rate of failed Is_cloneable VCS operations over 1m
The rate of failed Is_cloneable VCS operations. This is the time taken to check to see if a repository is cloneable from the upstream source.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100930 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (type) (rate(vcssyncer_is_cloneable_duration_seconds_count{type=~`${vcsSyncerType:regex}`, success="false"}[1m]))
Git Server: Gitserver: Gitserver Backend
gitserver: concurrent_backend_operations
Number of concurrently running backend operations
The number of requests that are currently being handled by gitserver backend layer, at the point in time of scraping.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101000 on your Sourcegraph instance.
Technical details
Query:
SHELLsrc_gitserver_backend_concurrent_operations
gitserver: gitserver_backend_total
Aggregate operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_backend_total{job=~"^gitserver.*"}[5m]))
gitserver: gitserver_backend_99th_percentile_duration
Aggregate successful operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_backend_duration_seconds_bucket{job=~"^gitserver.*"}[5m]))
gitserver: gitserver_backend_errors_total
Aggregate operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101012 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_backend_errors_total{job=~"^gitserver.*"}[5m]))
gitserver: gitserver_backend_error_rate
Aggregate operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101013 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_backend_errors_total{job=~"^gitserver.*"}[5m])) / (sum(increase(src_gitserver_backend_total{job=~"^gitserver.*"}[5m])) + sum(increase(src_gitserver_backend_errors_total{job=~"^gitserver.*"}[5m]))) * 100
gitserver: gitserver_backend_total
operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101020 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_backend_total{job=~"^gitserver.*"}[5m]))
gitserver: gitserver_backend_99th_percentile_duration
99th percentile successful operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101021 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_gitserver_backend_duration_seconds_bucket{job=~"^gitserver.*"}[5m])))
gitserver: gitserver_backend_errors_total
operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101022 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_backend_errors_total{job=~"^gitserver.*"}[5m]))
gitserver: gitserver_backend_error_rate
operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101023 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_backend_errors_total{job=~"^gitserver.*"}[5m])) / (sum by (op)(increase(src_gitserver_backend_total{job=~"^gitserver.*"}[5m])) + sum by (op)(increase(src_gitserver_backend_errors_total{job=~"^gitserver.*"}[5m]))) * 100
Git Server: Gitserver: Gitserver Client
gitserver: gitserver_client_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_total{job=~"^*.*"}[5m]))
gitserver: gitserver_client_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^*.*"}[5m]))
gitserver: gitserver_client_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^*.*"}[5m]))
gitserver: gitserver_client_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^*.*"}[5m])) / (sum(increase(src_gitserver_client_total{job=~"^*.*"}[5m])) + sum(increase(src_gitserver_client_errors_total{job=~"^*.*"}[5m]))) * 100
gitserver: gitserver_client_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101110 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_total{job=~"^*.*"}[5m]))
gitserver: gitserver_client_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op,scope)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^*.*"}[5m])))
gitserver: gitserver_client_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101112 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^*.*"}[5m]))
gitserver: gitserver_client_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101113 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^*.*"}[5m])) / (sum by (op,scope)(increase(src_gitserver_client_total{job=~"^*.*"}[5m])) + sum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^*.*"}[5m]))) * 100
Git Server: Gitserver: Gitserver Repository Service Client
gitserver: gitserver_repositoryservice_client_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_repositoryservice_client_total{job=~"^*.*"}[5m]))
gitserver: gitserver_repositoryservice_client_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_repositoryservice_client_duration_seconds_bucket{job=~"^*.*"}[5m]))
gitserver: gitserver_repositoryservice_client_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^*.*"}[5m]))
gitserver: gitserver_repositoryservice_client_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^*.*"}[5m])) / (sum(increase(src_gitserver_repositoryservice_client_total{job=~"^*.*"}[5m])) + sum(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^*.*"}[5m]))) * 100
gitserver: gitserver_repositoryservice_client_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_repositoryservice_client_total{job=~"^*.*"}[5m]))
gitserver: gitserver_repositoryservice_client_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op,scope)(rate(src_gitserver_repositoryservice_client_duration_seconds_bucket{job=~"^*.*"}[5m])))
gitserver: gitserver_repositoryservice_client_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^*.*"}[5m]))
gitserver: gitserver_repositoryservice_client_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101213 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^*.*"}[5m])) / (sum by (op,scope)(increase(src_gitserver_repositoryservice_client_total{job=~"^*.*"}[5m])) + sum by (op,scope)(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^*.*"}[5m]))) * 100
Git Server: Repos disk I/O metrics
gitserver: repos_disk_reads_sec
Read request rate over 1m (per instance)
The number of read requests that were issued to the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101300 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_completed_total{instance=~`node-exporter.*`}[1m])))))
gitserver: repos_disk_writes_sec
Write request rate over 1m (per instance)
The number of write requests that were issued to the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101301 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_completed_total{instance=~`node-exporter.*`}[1m])))))
gitserver: repos_disk_read_throughput
Read throughput over 1m (per instance)
The amount of data that was read from the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101310 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_read_bytes_total{instance=~`node-exporter.*`}[1m])))))
gitserver: repos_disk_write_throughput
Write throughput over 1m (per instance)
The amount of data that was written to the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101311 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_written_bytes_total{instance=~`node-exporter.*`}[1m])))))
gitserver: repos_disk_read_duration
Average read duration over 1m (per instance)
The average time for read requests issued to the device to be served. This includes the time spent by the requests in queue and the time spent servicing them.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101320 on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_read_time_seconds_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_completed_total{instance=~`node-exporter.*`}[1m])))))))
gitserver: repos_disk_write_duration
Average write duration over 1m (per instance)
The average time for write requests issued to the device to be served. This includes the time spent by the requests in queue and the time spent servicing them.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101321 on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_write_time_seconds_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_completed_total{instance=~`node-exporter.*`}[1m])))))))
gitserver: repos_disk_read_request_size
Average read request size over 1m (per instance)
The average size of read requests that were issued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101330 on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_read_bytes_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_completed_total{instance=~`node-exporter.*`}[1m])))))))
gitserver: repos_disk_write_request_size)
Average write request size over 1m (per instance)
The average size of write requests that were issued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101331 on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_written_bytes_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_completed_total{instance=~`node-exporter.*`}[1m])))))))
gitserver: repos_disk_reads_merged_sec
Merged read request rate over 1m (per instance)
The number of read requests merged per second that were queued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101340 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_merged_total{instance=~`node-exporter.*`}[1m])))))
gitserver: repos_disk_writes_merged_sec
Merged writes request rate over 1m (per instance)
The number of write requests merged per second that were queued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101341 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_merged_total{instance=~`node-exporter.*`}[1m])))))
gitserver: repos_disk_average_queue_size
Average queue size over 1m (per instance)
The number of I/O operations that were being queued or being serviced. See https://blog.actorsfit.com/a?ID=00200-428fa2ac-e338-4540-848c-af9a3eb1ebd2 for background (avgqu-sz).
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101350 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_io_time_weighted_seconds_total{instance=~`node-exporter.*`}[1m])))))
Git Server: Git Service GRPC server metrics
gitserver: git_service_grpc_request_rate_all_methods
Request rate across all methods over 2m
The number of gRPC requests received per second across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m]))
gitserver: git_service_grpc_request_rate_per_method
Request rate per-method over 2m
The number of gRPC requests received per second broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{grpc_method=~`${git_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])) by (grpc_method)
gitserver: git_service_error_percentage_all_methods
Error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101410 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_code!="OK",instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m]))) / (sum(rate(grpc_server_handled_total{instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m]))) ))
gitserver: git_service_grpc_error_percentage_per_method
Error percentage per-method over 2m
The percentage of gRPC requests that fail per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101411 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_method=~`${git_service_method:regex}`,grpc_code!="OK",instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])) by (grpc_method)) / (sum(rate(grpc_server_handled_total{grpc_method=~`${git_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])) by (grpc_method)) ))
gitserver: git_service_p99_response_time_per_method
99th percentile response time per method over 2m
The 99th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101420 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${git_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: git_service_p90_response_time_per_method
90th percentile response time per method over 2m
The 90th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101421 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${git_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: git_service_p75_response_time_per_method
75th percentile response time per method over 2m
The 75th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101422 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${git_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: git_service_p99_9_response_size_per_method
99.9th percentile total response size per method over 2m
The 99.9th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101430 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${git_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: git_service_p90_response_size_per_method
90th percentile total response size per method over 2m
The 90th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101431 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${git_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: git_service_p75_response_size_per_method
75th percentile total response size per method over 2m
The 75th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101432 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${git_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: git_service_p99_9_invididual_sent_message_size_per_method
99.9th percentile individual sent message size per method over 2m
The 99.9th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101440 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${git_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: git_service_p90_invididual_sent_message_size_per_method
90th percentile individual sent message size per method over 2m
The 90th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101441 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${git_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: git_service_p75_invididual_sent_message_size_per_method
75th percentile individual sent message size per method over 2m
The 75th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101442 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${git_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: git_service_grpc_response_stream_message_count_per_method
Average streaming response message count per-method over 2m
The average number of response messages sent during a streaming RPC method, broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101450 on your Sourcegraph instance.
Technical details
Query:
SHELL((sum(rate(grpc_server_msg_sent_total{grpc_type="server_stream",instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])) by (grpc_method))/(sum(rate(grpc_server_started_total{grpc_type="server_stream",instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])) by (grpc_method)))
gitserver: git_service_grpc_all_codes_per_method
Response codes rate per-method over 2m
The rate of all generated gRPC response codes per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101460 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_handled_total{grpc_method=~`${git_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])) by (grpc_method, grpc_code)
Git Server: Git Service GRPC "internal error" metrics
gitserver: git_service_grpc_clients_error_percentage_all_methods
Client baseline error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods (regardless of whether or not there was an internal error), aggregated across all "git_service" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101500 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_code!="OK"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService"}[2m])))))))
gitserver: git_service_grpc_clients_error_percentage_per_method
Client baseline error percentage per-method over 2m
The percentage of gRPC requests that fail per method (regardless of whether or not there was an internal error), aggregated across all "git_service" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101501 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${git_service_method:regex}",grpc_code!="OK"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${git_service_method:regex}"}[2m])) by (grpc_method))))))
gitserver: git_service_grpc_clients_all_codes_per_method
Client baseline response codes rate per-method over 2m
The rate of all generated gRPC response codes per method (regardless of whether or not there was an internal error), aggregated across all "git_service" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101502 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${git_service_method:regex}"}[2m])) by (grpc_method, grpc_code))
gitserver: git_service_grpc_clients_internal_error_percentage_all_methods
Client-observed gRPC internal error percentage across all methods over 2m
The percentage of gRPC requests that appear to fail due to gRPC internal errors across all methods, aggregated across all "git_service" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "git_service" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101510 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_code!="OK",is_internal_error="true"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService"}[2m])))))))
gitserver: git_service_grpc_clients_internal_error_percentage_per_method
Client-observed gRPC internal error percentage per-method over 2m
The percentage of gRPC requests that appear to fail to due to gRPC internal errors per method, aggregated across all "git_service" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "git_service" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101511 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${git_service_method:regex}",grpc_code!="OK",is_internal_error="true"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${git_service_method:regex}"}[2m])) by (grpc_method))))))
gitserver: git_service_grpc_clients_internal_error_all_codes_per_method
Client-observed gRPC internal error response code rate per-method over 2m
The rate of gRPC internal-error response codes per method, aggregated across all "git_service" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "git_service" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101512 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",is_internal_error="true",grpc_method=~"${git_service_method:regex}"}[2m])) by (grpc_method, grpc_code))
Git Server: Git Service GRPC retry metrics
gitserver: git_service_grpc_clients_retry_percentage_across_all_methods
Client retry percentage across all methods over 2m
The percentage of gRPC requests that were retried across all methods, aggregated across all "git_service" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101600 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverService",is_retried="true"}[2m])))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverService"}[2m])))))))
gitserver: git_service_grpc_clients_retry_percentage_per_method
Client retry percentage per-method over 2m
The percentage of gRPC requests that were retried aggregated across all "git_service" clients, broken out per method.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101601 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverService",is_retried="true",grpc_method=~"${git_service_method:regex}"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${git_service_method:regex}"}[2m])) by (grpc_method))))))
gitserver: git_service_grpc_clients_retry_count_per_method
Client retry count per-method over 2m
The count of gRPC requests that were retried aggregated across all "git_service" clients, broken out per method
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101602 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${git_service_method:regex}",is_retried="true"}[2m])) by (grpc_method))
Git Server: Repository Service GRPC server metrics
gitserver: repository_service_grpc_request_rate_all_methods
Request rate across all methods over 2m
The number of gRPC requests received per second across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101700 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m]))
gitserver: repository_service_grpc_request_rate_per_method
Request rate per-method over 2m
The number of gRPC requests received per second broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101701 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{grpc_method=~`${repository_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])) by (grpc_method)
gitserver: repository_service_error_percentage_all_methods
Error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101710 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_code!="OK",instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m]))) / (sum(rate(grpc_server_handled_total{instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m]))) ))
gitserver: repository_service_grpc_error_percentage_per_method
Error percentage per-method over 2m
The percentage of gRPC requests that fail per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101711 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_method=~`${repository_service_method:regex}`,grpc_code!="OK",instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])) by (grpc_method)) / (sum(rate(grpc_server_handled_total{grpc_method=~`${repository_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])) by (grpc_method)) ))
gitserver: repository_service_p99_response_time_per_method
99th percentile response time per method over 2m
The 99th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101720 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${repository_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])))
gitserver: repository_service_p90_response_time_per_method
90th percentile response time per method over 2m
The 90th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101721 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${repository_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])))
gitserver: repository_service_p75_response_time_per_method
75th percentile response time per method over 2m
The 75th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101722 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${repository_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])))
gitserver: repository_service_p99_9_response_size_per_method
99.9th percentile total response size per method over 2m
The 99.9th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101730 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${repository_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])))
gitserver: repository_service_p90_response_size_per_method
90th percentile total response size per method over 2m
The 90th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101731 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${repository_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])))
gitserver: repository_service_p75_response_size_per_method
75th percentile total response size per method over 2m
The 75th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101732 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${repository_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])))
gitserver: repository_service_p99_9_invididual_sent_message_size_per_method
99.9th percentile individual sent message size per method over 2m
The 99.9th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101740 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${repository_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])))
gitserver: repository_service_p90_invididual_sent_message_size_per_method
90th percentile individual sent message size per method over 2m
The 90th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101741 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${repository_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])))
gitserver: repository_service_p75_invididual_sent_message_size_per_method
75th percentile individual sent message size per method over 2m
The 75th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101742 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${repository_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])))
gitserver: repository_service_grpc_response_stream_message_count_per_method
Average streaming response message count per-method over 2m
The average number of response messages sent during a streaming RPC method, broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101750 on your Sourcegraph instance.
Technical details
Query:
SHELL((sum(rate(grpc_server_msg_sent_total{grpc_type="server_stream",instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])) by (grpc_method))/(sum(rate(grpc_server_started_total{grpc_type="server_stream",instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])) by (grpc_method)))
gitserver: repository_service_grpc_all_codes_per_method
Response codes rate per-method over 2m
The rate of all generated gRPC response codes per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101760 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_handled_total{grpc_method=~`${repository_service_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])) by (grpc_method, grpc_code)
Git Server: Repository Service GRPC "internal error" metrics
gitserver: repository_service_grpc_clients_error_percentage_all_methods
Client baseline error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods (regardless of whether or not there was an internal error), aggregated across all "repository_service" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101800 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverRepositoryService",grpc_code!="OK"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])))))))
gitserver: repository_service_grpc_clients_error_percentage_per_method
Client baseline error percentage per-method over 2m
The percentage of gRPC requests that fail per method (regardless of whether or not there was an internal error), aggregated across all "repository_service" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101801 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverRepositoryService",grpc_method=~"${repository_service_method:regex}",grpc_code!="OK"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverRepositoryService",grpc_method=~"${repository_service_method:regex}"}[2m])) by (grpc_method))))))
gitserver: repository_service_grpc_clients_all_codes_per_method
Client baseline response codes rate per-method over 2m
The rate of all generated gRPC response codes per method (regardless of whether or not there was an internal error), aggregated across all "repository_service" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101802 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverRepositoryService",grpc_method=~"${repository_service_method:regex}"}[2m])) by (grpc_method, grpc_code))
gitserver: repository_service_grpc_clients_internal_error_percentage_all_methods
Client-observed gRPC internal error percentage across all methods over 2m
The percentage of gRPC requests that appear to fail due to gRPC internal errors across all methods, aggregated across all "repository_service" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "repository_service" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101810 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverRepositoryService",grpc_code!="OK",is_internal_error="true"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])))))))
gitserver: repository_service_grpc_clients_internal_error_percentage_per_method
Client-observed gRPC internal error percentage per-method over 2m
The percentage of gRPC requests that appear to fail to due to gRPC internal errors per method, aggregated across all "repository_service" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "repository_service" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101811 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverRepositoryService",grpc_method=~"${repository_service_method:regex}",grpc_code!="OK",is_internal_error="true"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverRepositoryService",grpc_method=~"${repository_service_method:regex}"}[2m])) by (grpc_method))))))
gitserver: repository_service_grpc_clients_internal_error_all_codes_per_method
Client-observed gRPC internal error response code rate per-method over 2m
The rate of gRPC internal-error response codes per method, aggregated across all "repository_service" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "repository_service" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101812 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverRepositoryService",is_internal_error="true",grpc_method=~"${repository_service_method:regex}"}[2m])) by (grpc_method, grpc_code))
Git Server: Repository Service GRPC retry metrics
gitserver: repository_service_grpc_clients_retry_percentage_across_all_methods
Client retry percentage across all methods over 2m
The percentage of gRPC requests that were retried across all methods, aggregated across all "repository_service" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101900 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverRepositoryService",is_retried="true"}[2m])))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverRepositoryService"}[2m])))))))
gitserver: repository_service_grpc_clients_retry_percentage_per_method
Client retry percentage per-method over 2m
The percentage of gRPC requests that were retried aggregated across all "repository_service" clients, broken out per method.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101901 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverRepositoryService",is_retried="true",grpc_method=~"${repository_service_method:regex}"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverRepositoryService",grpc_method=~"${repository_service_method:regex}"}[2m])) by (grpc_method))))))
gitserver: repository_service_grpc_clients_retry_count_per_method
Client retry count per-method over 2m
The count of gRPC requests that were retried aggregated across all "repository_service" clients, broken out per method
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101902 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverRepositoryService",grpc_method=~"${repository_service_method:regex}",is_retried="true"}[2m])) by (grpc_method))
Git Server: Site configuration client update latency
gitserver: gitserver_site_configuration_duration_since_last_successful_update_by_instance
Duration since last successful site configuration update (by instance)
The duration since the configuration client used by the "gitserver" service last successfully updated its site configuration. Long durations could indicate issues updating the site configuration.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102000 on your Sourcegraph instance.
Technical details
Query:
SHELLsrc_conf_client_time_since_last_successful_update_seconds{job=~`.*gitserver`,instance=~`${shard:regex}`}
gitserver: gitserver_site_configuration_duration_since_last_successful_update_by_instance
Maximum duration since last successful site configuration update (all "gitserver" instances)
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102001 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(max_over_time(src_conf_client_time_since_last_successful_update_seconds{job=~`.*gitserver`,instance=~`${shard:regex}`}[1m]))
Git Server: HTTP handlers
gitserver: healthy_request_rate
Requests per second, by route, when status code is 200
The number of healthy HTTP requests per second to internal HTTP api
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (route) (rate(src_http_request_duration_seconds_count{app="gitserver",code=~"2.."}[5m]))
gitserver: unhealthy_request_rate
Requests per second, by route, when status code is not 200
The number of unhealthy HTTP requests per second to internal HTTP api
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (route) (rate(src_http_request_duration_seconds_count{app="gitserver",code!~"2.."}[5m]))
gitserver: request_rate_by_code
Requests per second, by status code
The number of HTTP requests per second by code
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (code) (rate(src_http_request_duration_seconds_count{app="gitserver"}[5m]))
gitserver: 95th_percentile_healthy_requests
95th percentile duration by route, when status code is 200
The 95th percentile duration by route when the status code is 200
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102110 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_http_request_duration_seconds_bucket{app="gitserver",code=~"2.."}[5m])) by (le, route))
gitserver: 95th_percentile_unhealthy_requests
95th percentile duration by route, when status code is not 200
The 95th percentile duration by route when the status code is not 200
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_http_request_duration_seconds_bucket{app="gitserver",code!~"2.."}[5m])) by (le, route))
Git Server: Database connections
gitserver: max_open_conns
Maximum open
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_max_open{app_name="gitserver"})
gitserver: open_conns
Established
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_open{app_name="gitserver"})
gitserver: in_use
Used
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_in_use{app_name="gitserver"})
gitserver: idle
Idle
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102211 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_idle{app_name="gitserver"})
gitserver: mean_blocked_seconds_per_conn_request
Mean blocked seconds per conn request
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102220 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_blocked_seconds{app_name="gitserver"}[5m])) / sum by (app_name, db_name) (increase(src_pgsql_conns_waited_for{app_name="gitserver"}[5m]))
gitserver: closed_max_idle
Closed by SetMaxIdleConns
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102230 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle{app_name="gitserver"}[5m]))
gitserver: closed_max_lifetime
Closed by SetConnMaxLifetime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102231 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_lifetime{app_name="gitserver"}[5m]))
gitserver: closed_max_idle_time
Closed by SetConnMaxIdleTime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102232 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle_time{app_name="gitserver"}[5m]))
Git Server: Container monitoring (not available on server)
gitserver: container_missing
Container missing
This value is the number of times a container has not been seen for more than one minute. If you observe this value change independent of deployment events (such as an upgrade), it could indicate pods are being OOM killed or terminated for some other reasons.
- Kubernetes:
- Determine if the pod was OOM killed using
kubectl describe pod gitserver(look forOOMKilled: true) and, if so, consider increasing the memory limit in the relevantDeployment.yaml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingkubectl logs -p gitserver.
- Determine if the pod was OOM killed using
- Docker Compose:
- Determine if the pod was OOM killed using
docker inspect -f '\{\{json .State\}\}' gitserver(look for"OOMKilled":true) and, if so, consider increasing the memory limit of the gitserver container indocker-compose.yml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingdocker logs gitserver(note this will include logs from the previous and currently running container).
- Determine if the pod was OOM killed using
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102300 on your Sourcegraph instance.
Technical details
Query:
SHELLcount by(name) ((time() - container_last_seen{name=~"^gitserver.*"}) > 60)
gitserver: container_cpu_usage
Container cpu usage total (1m average) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102301 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^gitserver.*"}
gitserver: container_memory_usage
Container memory usage by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102302 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^gitserver.*"}
gitserver: fs_io_operations
Filesystem reads and writes rate by instance over 1h
This value indicates the number of filesystem read and write operations by containers of this service. When extremely high, this can indicate a resource usage problem, or can cause problems with the service itself, especially if high values or spikes correlate with {{CONTAINER_NAME}} issues.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102303 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(name) (rate(container_fs_reads_total{name=~"^gitserver.*"}[1h]) + rate(container_fs_writes_total{name=~"^gitserver.*"}[1h]))
Git Server: Provisioning indicators (not available on server)
gitserver: provisioning_container_cpu_usage_long_term
Container cpu usage total (90th percentile over 1d) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102400 on your Sourcegraph instance.
Technical details
Query:
SHELLquantile_over_time(0.9, cadvisor_container_cpu_usage_percentage_total{name=~"^gitserver.*"}[1d])
gitserver: provisioning_container_memory_usage_long_term
Container memory usage (1d maximum) by instance
Git Server is expected to use up all the memory it is provided.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102401 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^gitserver.*"}[1d])
gitserver: provisioning_container_cpu_usage_short_term
Container cpu usage total (5m maximum) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102410 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^gitserver.*"}[5m])
gitserver: provisioning_container_memory_usage_short_term
Container memory usage (5m maximum) by instance
Git Server is expected to use up all the memory it is provided.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102411 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^gitserver.*"}[5m])
gitserver: container_oomkill_events_total
Container OOMKILL events total by instance
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102412 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^gitserver.*"})
Git Server: Golang runtime monitoring
gitserver: go_goroutines
Maximum active goroutines
A high value here indicates a possible goroutine leak.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102500 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_goroutines{job=~".*gitserver"})
gitserver: go_gc_duration_seconds
Maximum go garbage collection duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102501 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_gc_duration_seconds{job=~".*gitserver"})
Git Server: Kubernetes monitoring (only available on Kubernetes)
gitserver: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=102600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*gitserver"}) / count by (app) (up{app=~".*gitserver"}) * 100
Postgres
Postgres metrics, exported from postgres_exporter (not available on server).
To see this dashboard, visit /-/debug/grafana/d/postgres/postgres on your Sourcegraph instance.
postgres: connections
Active connections
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (job) (pg_stat_activity_count{datname!~"template.*|postgres|cloudsqladmin"}) OR sum by (job) (pg_stat_activity_count{job="codeinsights-db", datname!~"template.*|cloudsqladmin"})
postgres: usage_connections_percentage
Connection in use
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(pg_stat_activity_count) by (job) / (sum(pg_settings_max_connections) by (job) - sum(pg_settings_superuser_reserved_connections) by (job)) * 100
postgres: transaction_durations
Maximum transaction durations
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100002 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (job) (pg_stat_activity_max_tx_duration{datname!~"template.*|postgres|cloudsqladmin",job!="codeintel-db"}) OR sum by (job) (pg_stat_activity_max_tx_duration{job="codeinsights-db", datname!~"template.*|cloudsqladmin"})
Postgres: Database and collector status
postgres: postgres_up
Database availability
A non-zero value indicates the database is online.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLpg_up
postgres: invalid_indexes
Invalid indexes (unusable by the query planner)
A non-zero value indicates the that Postgres failed to build an index. Expect degraded performance until the index is manually rebuilt.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (relname)(pg_invalid_index_count)
postgres: pg_exporter_err
Errors scraping postgres exporter
This value indicates issues retrieving metrics from postgres_exporter.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLpg_exporter_last_scrape_error
postgres: migration_in_progress
Active schema migration
A 0 value indicates that no migration is in progress.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLpg_sg_migration_status
Postgres: Object size and bloat
postgres: pg_table_size
Table size
Total size of this table
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (relname)(pg_table_bloat_size)
postgres: pg_table_bloat_ratio
Table bloat ratio
Estimated bloat ratio of this table (high bloat = high overhead)
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (relname)(pg_table_bloat_ratio) * 100
postgres: pg_index_size
Index size
Total size of this index
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (relname)(pg_index_bloat_size)
postgres: pg_index_bloat_ratio
Index bloat ratio
Estimated bloat ratio of this index (high bloat = high overhead)
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (relname)(pg_index_bloat_ratio) * 100
Postgres: Provisioning indicators (not available on server)
postgres: provisioning_container_cpu_usage_long_term
Container cpu usage total (90th percentile over 1d) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLquantile_over_time(0.9, cadvisor_container_cpu_usage_percentage_total{name=~"^(pgsql|codeintel-db|codeinsights).*"}[1d])
postgres: provisioning_container_memory_usage_long_term
Container memory usage (1d maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^(pgsql|codeintel-db|codeinsights).*"}[1d])
postgres: provisioning_container_cpu_usage_short_term
Container cpu usage total (5m maximum) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100310 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^(pgsql|codeintel-db|codeinsights).*"}[5m])
postgres: provisioning_container_memory_usage_short_term
Container memory usage (5m maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100311 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^(pgsql|codeintel-db|codeinsights).*"}[5m])
postgres: container_oomkill_events_total
Container OOMKILL events total by instance
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100312 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^(pgsql|codeintel-db|codeinsights).*"})
Postgres: Kubernetes monitoring (only available on Kubernetes)
postgres: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/postgres/postgres?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*(pgsql|codeintel-db|codeinsights)"}) / count by (app) (up{app=~".*(pgsql|codeintel-db|codeinsights)"}) * 100
Precise Code Intel Worker
Handles conversion of uploaded precise code intelligence bundles.
To see this dashboard, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker on your Sourcegraph instance.
Precise Code Intel Worker: Codeintel: LSIF uploads
precise-code-intel-worker: codeintel_upload_handlers
Handler active handlers
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_codeintel_upload_processor_handlers{job=~"^precise-code-intel-worker.*"})
precise-code-intel-worker: codeintel_upload_processor_upload_size
Sum of upload sizes in bytes being processed by each precise code-intel worker instance
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(instance) (src_codeintel_upload_processor_upload_size{job="precise-code-intel-worker"})
precise-code-intel-worker: codeintel_upload_processor_total
Handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_upload_processor_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_upload_processor_99th_percentile_duration
Aggregate successful handler operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_upload_processor_duration_seconds_bucket{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_upload_processor_errors_total
Handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100012 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_upload_processor_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_upload_processor_error_rate
Handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100013 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_upload_processor_errors_total{job=~"^precise-code-intel-worker.*"}[5m])) / (sum(increase(src_codeintel_upload_processor_total{job=~"^precise-code-intel-worker.*"}[5m])) + sum(increase(src_codeintel_upload_processor_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))) * 100
Precise Code Intel Worker: Codeintel: dbstore stats
precise-code-intel-worker: codeintel_uploads_store_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploads_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploads_store_duration_seconds_bucket{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploads_store_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploads_store_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_errors_total{job=~"^precise-code-intel-worker.*"}[5m])) / (sum(increase(src_codeintel_uploads_store_total{job=~"^precise-code-intel-worker.*"}[5m])) + sum(increase(src_codeintel_uploads_store_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))) * 100
precise-code-intel-worker: codeintel_uploads_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploads_store_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_store_duration_seconds_bucket{job=~"^precise-code-intel-worker.*"}[5m])))
precise-code-intel-worker: codeintel_uploads_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100112 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploads_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100113 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^precise-code-intel-worker.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_store_total{job=~"^precise-code-intel-worker.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))) * 100
Precise Code Intel Worker: Codeintel: lsifstore stats
precise-code-intel-worker: codeintel_uploads_lsifstore_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_lsifstore_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploads_lsifstore_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploads_lsifstore_duration_seconds_bucket{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploads_lsifstore_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploads_lsifstore_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^precise-code-intel-worker.*"}[5m])) / (sum(increase(src_codeintel_uploads_lsifstore_total{job=~"^precise-code-intel-worker.*"}[5m])) + sum(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))) * 100
precise-code-intel-worker: codeintel_uploads_lsifstore_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_lsifstore_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploads_lsifstore_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_lsifstore_duration_seconds_bucket{job=~"^precise-code-intel-worker.*"}[5m])))
precise-code-intel-worker: codeintel_uploads_lsifstore_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploads_lsifstore_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100213 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^precise-code-intel-worker.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_lsifstore_total{job=~"^precise-code-intel-worker.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))) * 100
Precise Code Intel Worker: Workerutil: lsif_uploads dbworker/store stats
precise-code-intel-worker: workerutil_dbworker_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_total{domain='codeintel_upload',job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: workerutil_dbworker_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_workerutil_dbworker_store_duration_seconds_bucket{domain='codeintel_upload',job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: workerutil_dbworker_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100302 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_errors_total{domain='codeintel_upload',job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: workerutil_dbworker_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100303 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_errors_total{domain='codeintel_upload',job=~"^precise-code-intel-worker.*"}[5m])) / (sum(increase(src_workerutil_dbworker_store_total{domain='codeintel_upload',job=~"^precise-code-intel-worker.*"}[5m])) + sum(increase(src_workerutil_dbworker_store_errors_total{domain='codeintel_upload',job=~"^precise-code-intel-worker.*"}[5m]))) * 100
Precise Code Intel Worker: Codeintel: gitserver client
precise-code-intel-worker: gitserver_client_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: gitserver_client_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: gitserver_client_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100402 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: gitserver_client_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100403 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^precise-code-intel-worker.*"}[5m])) / (sum(increase(src_gitserver_client_total{job=~"^precise-code-intel-worker.*"}[5m])) + sum(increase(src_gitserver_client_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))) * 100
precise-code-intel-worker: gitserver_client_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100410 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_client_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: gitserver_client_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100411 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^precise-code-intel-worker.*"}[5m])))
precise-code-intel-worker: gitserver_client_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100412 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_client_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: gitserver_client_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100413 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_client_errors_total{job=~"^precise-code-intel-worker.*"}[5m])) / (sum by (op)(increase(src_gitserver_client_total{job=~"^precise-code-intel-worker.*"}[5m])) + sum by (op)(increase(src_gitserver_client_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))) * 100
Precise Code Intel Worker: Codeintel: uploadstore stats
precise-code-intel-worker: codeintel_uploadstore_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploadstore_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploadstore_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100501 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploadstore_duration_seconds_bucket{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploadstore_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100502 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploadstore_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploadstore_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100503 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploadstore_errors_total{job=~"^precise-code-intel-worker.*"}[5m])) / (sum(increase(src_codeintel_uploadstore_total{job=~"^precise-code-intel-worker.*"}[5m])) + sum(increase(src_codeintel_uploadstore_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))) * 100
precise-code-intel-worker: codeintel_uploadstore_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100510 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploadstore_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploadstore_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100511 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploadstore_duration_seconds_bucket{job=~"^precise-code-intel-worker.*"}[5m])))
precise-code-intel-worker: codeintel_uploadstore_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100512 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploadstore_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))
precise-code-intel-worker: codeintel_uploadstore_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100513 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploadstore_errors_total{job=~"^precise-code-intel-worker.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploadstore_total{job=~"^precise-code-intel-worker.*"}[5m])) + sum by (op)(increase(src_codeintel_uploadstore_errors_total{job=~"^precise-code-intel-worker.*"}[5m]))) * 100
Precise Code Intel Worker: Database connections
precise-code-intel-worker: max_open_conns
Maximum open
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_max_open{app_name="precise-code-intel-worker"})
precise-code-intel-worker: open_conns
Established
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100601 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_open{app_name="precise-code-intel-worker"})
precise-code-intel-worker: in_use
Used
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100610 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_in_use{app_name="precise-code-intel-worker"})
precise-code-intel-worker: idle
Idle
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100611 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_idle{app_name="precise-code-intel-worker"})
precise-code-intel-worker: mean_blocked_seconds_per_conn_request
Mean blocked seconds per conn request
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100620 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_blocked_seconds{app_name="precise-code-intel-worker"}[5m])) / sum by (app_name, db_name) (increase(src_pgsql_conns_waited_for{app_name="precise-code-intel-worker"}[5m]))
precise-code-intel-worker: closed_max_idle
Closed by SetMaxIdleConns
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100630 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle{app_name="precise-code-intel-worker"}[5m]))
precise-code-intel-worker: closed_max_lifetime
Closed by SetConnMaxLifetime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100631 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_lifetime{app_name="precise-code-intel-worker"}[5m]))
precise-code-intel-worker: closed_max_idle_time
Closed by SetConnMaxIdleTime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100632 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle_time{app_name="precise-code-intel-worker"}[5m]))
Precise Code Intel Worker: Precise-code-intel-worker (CPU, Memory)
precise-code-intel-worker: cpu_usage_percentage
CPU usage
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100700 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^precise-code-intel-worker.*"}
precise-code-intel-worker: memory_usage_percentage
Memory usage percentage (total)
An estimate for the active memory in use, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100701 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^precise-code-intel-worker.*"}
precise-code-intel-worker: memory_working_set_bytes
Memory usage bytes (total)
An estimate for the active memory in use in bytes, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100702 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_memory_working_set_bytes{name=~"^precise-code-intel-worker.*"})
precise-code-intel-worker: memory_rss
Memory (RSS)
The total anonymous memory in use by the application, which includes Go stack and heap. This memory is is non-reclaimable, and high usage may trigger OOM kills. Note: the metric is named RSS because to match the cadvisor name, but anonymous is more accurate."
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100710 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_rss{name=~"^precise-code-intel-worker.*"} / container_spec_memory_limit_bytes{name=~"^precise-code-intel-worker.*"}) by (name) * 100.0
precise-code-intel-worker: memory_total_active_file
Memory usage (active file)
This metric shows the total active file-backed memory currently in use by the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100711 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_total_active_file_bytes{name=~"^precise-code-intel-worker.*"} / container_spec_memory_limit_bytes{name=~"^precise-code-intel-worker.*"}) by (name) * 100.0
precise-code-intel-worker: memory_kernel_usage
Memory usage (kernel)
The kernel usage metric shows the amount of memory used by the kernel on behalf of the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100712 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_kernel_usage{name=~"^precise-code-intel-worker.*"} / container_spec_memory_limit_bytes{name=~"^precise-code-intel-worker.*"}) by (name) * 100.0
Precise Code Intel Worker: Container monitoring (not available on server)
precise-code-intel-worker: container_missing
Container missing
This value is the number of times a container has not been seen for more than one minute. If you observe this value change independent of deployment events (such as an upgrade), it could indicate pods are being OOM killed or terminated for some other reasons.
- Kubernetes:
- Determine if the pod was OOM killed using
kubectl describe pod precise-code-intel-worker(look forOOMKilled: true) and, if so, consider increasing the memory limit in the relevantDeployment.yaml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingkubectl logs -p precise-code-intel-worker.
- Determine if the pod was OOM killed using
- Docker Compose:
- Determine if the pod was OOM killed using
docker inspect -f '\{\{json .State\}\}' precise-code-intel-worker(look for"OOMKilled":true) and, if so, consider increasing the memory limit of the precise-code-intel-worker container indocker-compose.yml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingdocker logs precise-code-intel-worker(note this will include logs from the previous and currently running container).
- Determine if the pod was OOM killed using
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100800 on your Sourcegraph instance.
Technical details
Query:
SHELLcount by(name) ((time() - container_last_seen{name=~"^precise-code-intel-worker.*"}) > 60)
precise-code-intel-worker: container_cpu_usage
Container cpu usage total (1m average) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100801 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^precise-code-intel-worker.*"}
precise-code-intel-worker: container_memory_usage
Container memory usage by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100802 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^precise-code-intel-worker.*"}
precise-code-intel-worker: fs_io_operations
Filesystem reads and writes rate by instance over 1h
This value indicates the number of filesystem read and write operations by containers of this service. When extremely high, this can indicate a resource usage problem, or can cause problems with the service itself, especially if high values or spikes correlate with {{CONTAINER_NAME}} issues.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100803 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(name) (rate(container_fs_reads_total{name=~"^precise-code-intel-worker.*"}[1h]) + rate(container_fs_writes_total{name=~"^precise-code-intel-worker.*"}[1h]))
Precise Code Intel Worker: Provisioning indicators (not available on server)
precise-code-intel-worker: provisioning_container_cpu_usage_long_term
Container cpu usage total (90th percentile over 1d) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100900 on your Sourcegraph instance.
Technical details
Query:
SHELLquantile_over_time(0.9, cadvisor_container_cpu_usage_percentage_total{name=~"^precise-code-intel-worker.*"}[1d])
precise-code-intel-worker: provisioning_container_memory_usage_long_term
Container memory usage (1d maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100901 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^precise-code-intel-worker.*"}[1d])
precise-code-intel-worker: provisioning_container_cpu_usage_short_term
Container cpu usage total (5m maximum) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100910 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^precise-code-intel-worker.*"}[5m])
precise-code-intel-worker: provisioning_container_memory_usage_short_term
Container memory usage (5m maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100911 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^precise-code-intel-worker.*"}[5m])
precise-code-intel-worker: container_oomkill_events_total
Container OOMKILL events total by instance
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=100912 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^precise-code-intel-worker.*"})
Precise Code Intel Worker: Golang runtime monitoring
precise-code-intel-worker: go_goroutines
Maximum active goroutines
A high value here indicates a possible goroutine leak.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=101000 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_goroutines{job=~".*precise-code-intel-worker"})
precise-code-intel-worker: go_gc_duration_seconds
Maximum go garbage collection duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=101001 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_gc_duration_seconds{job=~".*precise-code-intel-worker"})
Precise Code Intel Worker: Kubernetes monitoring (only available on Kubernetes)
precise-code-intel-worker: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/precise-code-intel-worker/precise-code-intel-worker?viewPanel=101100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*precise-code-intel-worker"}) / count by (app) (up{app=~".*precise-code-intel-worker"}) * 100
Syntactic Indexing
Handles syntactic indexing of repositories.
To see this dashboard, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing on your Sourcegraph instance.
Syntactic Indexing: Syntactic indexing scheduling: summary
####syntactic-indexing:
Syntactic indexing jobs proposed for insertion over 5m
Syntactic indexing jobs are proposed for insertion into the queue based on round-robin scheduling across recently modified repos.
This should be equal to the sum of inserted + updated + skipped, but is shown separately for clarity.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_syntactic_enqueuer_jobs_proposed[5m]))
####syntactic-indexing:
Syntactic indexing jobs inserted over 5m
Syntactic indexing jobs are inserted into the queue if there is a proposed repo commit pair (R, X) such that there is no existing job for R in the queue.
If this number is close to the number of proposed jobs, it may indicate that the scheduler is not able to keep up with the rate of incoming commits.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_syntactic_enqueuer_jobs_inserted[5m]))
####syntactic-indexing:
Syntactic indexing jobs updated in-place over 5m
Syntactic indexing jobs are updated in-place when the scheduler attempts to enqueue a repo commit pair (R, X) and discovers that the queue already had some other repo commit pair (R, Y) where Y is an ancestor of X. In that case, the job is updated in-place to point to X, to reflect the fact that users looking at the tip of the default branch of R are more likely to benefit from newer commits being indexed.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100002 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_syntactic_enqueuer_jobs_updated[5m]))
####syntactic-indexing:
Syntactic indexing jobs skipped over 5m
Syntactic indexing jobs insertion is skipped when the scheduler attempts to enqueue a repo commit pair (R, X) and discovers that the queue already had the same job (most likely) or another job (R, Y) where Y is not an ancestor of X.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100003 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_syntactic_enqueuer_jobs_skipped[5m]))
Syntactic Indexing: Workerutil: syntactic_scip_indexing_jobs dbworker/store stats
syntactic-indexing: workerutil_dbworker_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_total{domain='syntactic_scip_indexing_jobs',job=~"^syntactic-code-intel-worker.*"}[5m]))
syntactic-indexing: workerutil_dbworker_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_workerutil_dbworker_store_duration_seconds_bucket{domain='syntactic_scip_indexing_jobs',job=~"^syntactic-code-intel-worker.*"}[5m]))
syntactic-indexing: workerutil_dbworker_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_errors_total{domain='syntactic_scip_indexing_jobs',job=~"^syntactic-code-intel-worker.*"}[5m]))
syntactic-indexing: workerutil_dbworker_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_errors_total{domain='syntactic_scip_indexing_jobs',job=~"^syntactic-code-intel-worker.*"}[5m])) / (sum(increase(src_workerutil_dbworker_store_total{domain='syntactic_scip_indexing_jobs',job=~"^syntactic-code-intel-worker.*"}[5m])) + sum(increase(src_workerutil_dbworker_store_errors_total{domain='syntactic_scip_indexing_jobs',job=~"^syntactic-code-intel-worker.*"}[5m]))) * 100
Syntactic Indexing: Codeintel: gitserver client
syntactic-indexing: gitserver_client_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_total{job=~"^syntactic-code-intel-worker.*"}[5m]))
syntactic-indexing: gitserver_client_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^syntactic-code-intel-worker.*"}[5m]))
syntactic-indexing: gitserver_client_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^syntactic-code-intel-worker.*"}[5m]))
syntactic-indexing: gitserver_client_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^syntactic-code-intel-worker.*"}[5m])) / (sum(increase(src_gitserver_client_total{job=~"^syntactic-code-intel-worker.*"}[5m])) + sum(increase(src_gitserver_client_errors_total{job=~"^syntactic-code-intel-worker.*"}[5m]))) * 100
syntactic-indexing: gitserver_client_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_client_total{job=~"^syntactic-code-intel-worker.*"}[5m]))
syntactic-indexing: gitserver_client_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^syntactic-code-intel-worker.*"}[5m])))
syntactic-indexing: gitserver_client_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_client_errors_total{job=~"^syntactic-code-intel-worker.*"}[5m]))
syntactic-indexing: gitserver_client_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100213 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_client_errors_total{job=~"^syntactic-code-intel-worker.*"}[5m])) / (sum by (op)(increase(src_gitserver_client_total{job=~"^syntactic-code-intel-worker.*"}[5m])) + sum by (op)(increase(src_gitserver_client_errors_total{job=~"^syntactic-code-intel-worker.*"}[5m]))) * 100
Syntactic Indexing: Database connections
syntactic-indexing: max_open_conns
Maximum open
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_max_open{app_name="syntactic-code-intel-worker"})
syntactic-indexing: open_conns
Established
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_open{app_name="syntactic-code-intel-worker"})
syntactic-indexing: in_use
Used
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100310 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_in_use{app_name="syntactic-code-intel-worker"})
syntactic-indexing: idle
Idle
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100311 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_idle{app_name="syntactic-code-intel-worker"})
syntactic-indexing: mean_blocked_seconds_per_conn_request
Mean blocked seconds per conn request
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100320 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_blocked_seconds{app_name="syntactic-code-intel-worker"}[5m])) / sum by (app_name, db_name) (increase(src_pgsql_conns_waited_for{app_name="syntactic-code-intel-worker"}[5m]))
syntactic-indexing: closed_max_idle
Closed by SetMaxIdleConns
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100330 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle{app_name="syntactic-code-intel-worker"}[5m]))
syntactic-indexing: closed_max_lifetime
Closed by SetConnMaxLifetime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100331 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_lifetime{app_name="syntactic-code-intel-worker"}[5m]))
syntactic-indexing: closed_max_idle_time
Closed by SetConnMaxIdleTime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100332 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle_time{app_name="syntactic-code-intel-worker"}[5m]))
Syntactic Indexing: Syntactic-code-intel-worker (CPU, Memory)
syntactic-indexing: cpu_usage_percentage
CPU usage
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^syntactic-code-intel-worker.*"}
syntactic-indexing: memory_usage_percentage
Memory usage percentage (total)
An estimate for the active memory in use, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^syntactic-code-intel-worker.*"}
syntactic-indexing: memory_working_set_bytes
Memory usage bytes (total)
An estimate for the active memory in use in bytes, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100402 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_memory_working_set_bytes{name=~"^syntactic-code-intel-worker.*"})
syntactic-indexing: memory_rss
Memory (RSS)
The total anonymous memory in use by the application, which includes Go stack and heap. This memory is is non-reclaimable, and high usage may trigger OOM kills. Note: the metric is named RSS because to match the cadvisor name, but anonymous is more accurate."
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100410 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_rss{name=~"^syntactic-code-intel-worker.*"} / container_spec_memory_limit_bytes{name=~"^syntactic-code-intel-worker.*"}) by (name) * 100.0
syntactic-indexing: memory_total_active_file
Memory usage (active file)
This metric shows the total active file-backed memory currently in use by the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100411 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_total_active_file_bytes{name=~"^syntactic-code-intel-worker.*"} / container_spec_memory_limit_bytes{name=~"^syntactic-code-intel-worker.*"}) by (name) * 100.0
syntactic-indexing: memory_kernel_usage
Memory usage (kernel)
The kernel usage metric shows the amount of memory used by the kernel on behalf of the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100412 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_kernel_usage{name=~"^syntactic-code-intel-worker.*"} / container_spec_memory_limit_bytes{name=~"^syntactic-code-intel-worker.*"}) by (name) * 100.0
Syntactic Indexing: Container monitoring (not available on server)
syntactic-indexing: container_missing
Container missing
This value is the number of times a container has not been seen for more than one minute. If you observe this value change independent of deployment events (such as an upgrade), it could indicate pods are being OOM killed or terminated for some other reasons.
- Kubernetes:
- Determine if the pod was OOM killed using
kubectl describe pod syntactic-code-intel-worker(look forOOMKilled: true) and, if so, consider increasing the memory limit in the relevantDeployment.yaml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingkubectl logs -p syntactic-code-intel-worker.
- Determine if the pod was OOM killed using
- Docker Compose:
- Determine if the pod was OOM killed using
docker inspect -f '\{\{json .State\}\}' syntactic-code-intel-worker(look for"OOMKilled":true) and, if so, consider increasing the memory limit of the syntactic-code-intel-worker container indocker-compose.yml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingdocker logs syntactic-code-intel-worker(note this will include logs from the previous and currently running container).
- Determine if the pod was OOM killed using
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLcount by(name) ((time() - container_last_seen{name=~"^syntactic-code-intel-worker.*"}) > 60)
syntactic-indexing: container_cpu_usage
Container cpu usage total (1m average) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100501 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^syntactic-code-intel-worker.*"}
syntactic-indexing: container_memory_usage
Container memory usage by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100502 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^syntactic-code-intel-worker.*"}
syntactic-indexing: fs_io_operations
Filesystem reads and writes rate by instance over 1h
This value indicates the number of filesystem read and write operations by containers of this service. When extremely high, this can indicate a resource usage problem, or can cause problems with the service itself, especially if high values or spikes correlate with {{CONTAINER_NAME}} issues.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100503 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(name) (rate(container_fs_reads_total{name=~"^syntactic-code-intel-worker.*"}[1h]) + rate(container_fs_writes_total{name=~"^syntactic-code-intel-worker.*"}[1h]))
Syntactic Indexing: Provisioning indicators (not available on server)
syntactic-indexing: provisioning_container_cpu_usage_long_term
Container cpu usage total (90th percentile over 1d) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100600 on your Sourcegraph instance.
Technical details
Query:
SHELLquantile_over_time(0.9, cadvisor_container_cpu_usage_percentage_total{name=~"^syntactic-code-intel-worker.*"}[1d])
syntactic-indexing: provisioning_container_memory_usage_long_term
Container memory usage (1d maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100601 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^syntactic-code-intel-worker.*"}[1d])
syntactic-indexing: provisioning_container_cpu_usage_short_term
Container cpu usage total (5m maximum) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100610 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^syntactic-code-intel-worker.*"}[5m])
syntactic-indexing: provisioning_container_memory_usage_short_term
Container memory usage (5m maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100611 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^syntactic-code-intel-worker.*"}[5m])
syntactic-indexing: container_oomkill_events_total
Container OOMKILL events total by instance
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100612 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^syntactic-code-intel-worker.*"})
Syntactic Indexing: Golang runtime monitoring
syntactic-indexing: go_goroutines
Maximum active goroutines
A high value here indicates a possible goroutine leak.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100700 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_goroutines{job=~".*syntactic-code-intel-worker"})
syntactic-indexing: go_gc_duration_seconds
Maximum go garbage collection duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100701 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_gc_duration_seconds{job=~".*syntactic-code-intel-worker"})
Syntactic Indexing: Kubernetes monitoring (only available on Kubernetes)
syntactic-indexing: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntactic-indexing/syntactic-indexing?viewPanel=100800 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*syntactic-code-intel-worker"}) / count by (app) (up{app=~".*syntactic-code-intel-worker"}) * 100
Redis
Metrics from both redis databases.
To see this dashboard, visit /-/debug/grafana/d/redis/redis on your Sourcegraph instance.
Redis: Redis Store
redis: redis-store_up
Redis-store availability
A value of 1 indicates the service is currently running
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLredis_up{app="redis-store"}
Redis: Redis Cache
redis: redis-cache_up
Redis-cache availability
A value of 1 indicates the service is currently running
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLredis_up{app="redis-cache"}
Redis: Provisioning indicators (not available on server)
redis: provisioning_container_cpu_usage_long_term
Container cpu usage total (90th percentile over 1d) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLquantile_over_time(0.9, cadvisor_container_cpu_usage_percentage_total{name=~"^redis-cache.*"}[1d])
redis: provisioning_container_memory_usage_long_term
Container memory usage (1d maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^redis-cache.*"}[1d])
redis: provisioning_container_cpu_usage_short_term
Container cpu usage total (5m maximum) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^redis-cache.*"}[5m])
redis: provisioning_container_memory_usage_short_term
Container memory usage (5m maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^redis-cache.*"}[5m])
redis: container_oomkill_events_total
Container OOMKILL events total by instance
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^redis-cache.*"})
Redis: Provisioning indicators (not available on server)
redis: provisioning_container_cpu_usage_long_term
Container cpu usage total (90th percentile over 1d) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLquantile_over_time(0.9, cadvisor_container_cpu_usage_percentage_total{name=~"^redis-store.*"}[1d])
redis: provisioning_container_memory_usage_long_term
Container memory usage (1d maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^redis-store.*"}[1d])
redis: provisioning_container_cpu_usage_short_term
Container cpu usage total (5m maximum) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100310 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^redis-store.*"}[5m])
redis: provisioning_container_memory_usage_short_term
Container memory usage (5m maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100311 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^redis-store.*"}[5m])
redis: container_oomkill_events_total
Container OOMKILL events total by instance
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100312 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^redis-store.*"})
Redis: Kubernetes monitoring (only available on Kubernetes)
redis: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*redis-cache"}) / count by (app) (up{app=~".*redis-cache"}) * 100
Redis: Kubernetes monitoring (only available on Kubernetes)
redis: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/redis/redis?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*redis-store"}) / count by (app) (up{app=~".*redis-store"}) * 100
Worker
Manages background processes.
To see this dashboard, visit /-/debug/grafana/d/worker/worker on your Sourcegraph instance.
Worker: Active jobs
worker: worker_job_count
Number of worker instances running each job
The number of worker instances running each job type. It is necessary for each job type to be managed by at least one worker instance.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (job_name) (src_worker_jobs{job=~"^worker.*"})
worker: worker_job_codeintel-upload-janitor_count
Number of worker instances running the codeintel-upload-janitor job
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum (src_worker_jobs{job=~"^worker.*", job_name="codeintel-upload-janitor"})
worker: worker_job_codeintel-commitgraph-updater_count
Number of worker instances running the codeintel-commitgraph-updater job
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum (src_worker_jobs{job=~"^worker.*", job_name="codeintel-commitgraph-updater"})
worker: worker_job_codeintel-autoindexing-scheduler_count
Number of worker instances running the codeintel-autoindexing-scheduler job
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100012 on your Sourcegraph instance.
Technical details
Query:
SHELLsum (src_worker_jobs{job=~"^worker.*", job_name="codeintel-autoindexing-scheduler"})
Worker: Database record encrypter
worker: records_encrypted_at_rest_percentage
Percentage of database records encrypted at rest
Percentage of encrypted database records
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELL(max(src_records_encrypted_at_rest_total) by (tableName)) / ((max(src_records_encrypted_at_rest_total) by (tableName)) + (max(src_records_unencrypted_at_rest_total) by (tableName))) * 100
worker: records_encrypted_total
Database records encrypted every 5m
Number of encrypted database records every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (tableName)(increase(src_records_encrypted_total{job=~"^worker.*"}[5m]))
worker: records_decrypted_total
Database records decrypted every 5m
Number of encrypted database records every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (tableName)(increase(src_records_decrypted_total{job=~"^worker.*"}[5m]))
worker: record_encryption_errors_total
Encryption operation errors every 5m
Number of database record encryption/decryption errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_record_encryption_errors_total{job=~"^worker.*"}[5m]))
Worker: Codeintel: Repository commit graph updates
worker: codeintel_commit_graph_processor_total
Update operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_commit_graph_processor_total{job=~"^worker.*"}[5m]))
worker: codeintel_commit_graph_processor_99th_percentile_duration
Aggregate successful update operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_commit_graph_processor_duration_seconds_bucket{job=~"^worker.*"}[5m]))
worker: codeintel_commit_graph_processor_errors_total
Update operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_commit_graph_processor_errors_total{job=~"^worker.*"}[5m]))
worker: codeintel_commit_graph_processor_error_rate
Update operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_commit_graph_processor_errors_total{job=~"^worker.*"}[5m])) / (sum(increase(src_codeintel_commit_graph_processor_total{job=~"^worker.*"}[5m])) + sum(increase(src_codeintel_commit_graph_processor_errors_total{job=~"^worker.*"}[5m]))) * 100
Worker: Codeintel: Auto-index scheduler
worker: codeintel_autoindexing_total
Auto-indexing job scheduler operations every 10m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_total{op='HandleIndexSchedule',job=~"^worker.*"}[10m]))
worker: codeintel_autoindexing_99th_percentile_duration
Aggregate successful auto-indexing job scheduler operation duration distribution over 10m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_autoindexing_duration_seconds_bucket{op='HandleIndexSchedule',job=~"^worker.*"}[10m]))
worker: codeintel_autoindexing_errors_total
Auto-indexing job scheduler operation errors every 10m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100302 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_errors_total{op='HandleIndexSchedule',job=~"^worker.*"}[10m]))
worker: codeintel_autoindexing_error_rate
Auto-indexing job scheduler operation error rate over 10m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100303 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_errors_total{op='HandleIndexSchedule',job=~"^worker.*"}[10m])) / (sum(increase(src_codeintel_autoindexing_total{op='HandleIndexSchedule',job=~"^worker.*"}[10m])) + sum(increase(src_codeintel_autoindexing_errors_total{op='HandleIndexSchedule',job=~"^worker.*"}[10m]))) * 100
Worker: Codeintel: dbstore stats
worker: codeintel_uploads_store_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_total{job=~"^worker.*"}[5m]))
worker: codeintel_uploads_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploads_store_duration_seconds_bucket{job=~"^worker.*"}[5m]))
worker: codeintel_uploads_store_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100402 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_errors_total{job=~"^worker.*"}[5m]))
worker: codeintel_uploads_store_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100403 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_errors_total{job=~"^worker.*"}[5m])) / (sum(increase(src_codeintel_uploads_store_total{job=~"^worker.*"}[5m])) + sum(increase(src_codeintel_uploads_store_errors_total{job=~"^worker.*"}[5m]))) * 100
worker: codeintel_uploads_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100410 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_total{job=~"^worker.*"}[5m]))
worker: codeintel_uploads_store_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100411 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_store_duration_seconds_bucket{job=~"^worker.*"}[5m])))
worker: codeintel_uploads_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100412 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^worker.*"}[5m]))
worker: codeintel_uploads_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100413 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^worker.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_store_total{job=~"^worker.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^worker.*"}[5m]))) * 100
Worker: Codeintel: lsifstore stats
worker: codeintel_uploads_lsifstore_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_lsifstore_total{job=~"^worker.*"}[5m]))
worker: codeintel_uploads_lsifstore_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100501 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploads_lsifstore_duration_seconds_bucket{job=~"^worker.*"}[5m]))
worker: codeintel_uploads_lsifstore_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100502 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^worker.*"}[5m]))
worker: codeintel_uploads_lsifstore_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100503 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^worker.*"}[5m])) / (sum(increase(src_codeintel_uploads_lsifstore_total{job=~"^worker.*"}[5m])) + sum(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^worker.*"}[5m]))) * 100
worker: codeintel_uploads_lsifstore_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100510 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_lsifstore_total{job=~"^worker.*"}[5m]))
worker: codeintel_uploads_lsifstore_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100511 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_lsifstore_duration_seconds_bucket{job=~"^worker.*"}[5m])))
worker: codeintel_uploads_lsifstore_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100512 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^worker.*"}[5m]))
worker: codeintel_uploads_lsifstore_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100513 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^worker.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_lsifstore_total{job=~"^worker.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^worker.*"}[5m]))) * 100
Worker: Codeintel: gitserver client
worker: gitserver_client_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_total{job=~"^worker.*"}[5m]))
worker: gitserver_client_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100601 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^worker.*"}[5m]))
worker: gitserver_client_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100602 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^worker.*"}[5m]))
worker: gitserver_client_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100603 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^worker.*"}[5m])) / (sum(increase(src_gitserver_client_total{job=~"^worker.*"}[5m])) + sum(increase(src_gitserver_client_errors_total{job=~"^worker.*"}[5m]))) * 100
worker: gitserver_client_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100610 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_client_total{job=~"^worker.*"}[5m]))
worker: gitserver_client_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100611 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^worker.*"}[5m])))
worker: gitserver_client_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100612 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_client_errors_total{job=~"^worker.*"}[5m]))
worker: gitserver_client_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100613 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_client_errors_total{job=~"^worker.*"}[5m])) / (sum by (op)(increase(src_gitserver_client_total{job=~"^worker.*"}[5m])) + sum by (op)(increase(src_gitserver_client_errors_total{job=~"^worker.*"}[5m]))) * 100
Worker: Repositories
worker: syncer_sync_last_time
Time since last sync
A high value here indicates issues synchronizing repo metadata. If the value is persistently high, make sure all external services have valid tokens.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100700 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(timestamp(vector(time()))) - max(src_repoupdater_syncer_sync_last_time)
worker: src_repoupdater_max_sync_backoff
Time since oldest sync
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100701 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_repoupdater_max_sync_backoff)
worker: src_repoupdater_syncer_sync_errors_total
Site level external service sync error rate
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100702 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (family) (rate(src_repoupdater_syncer_sync_errors_total{owner!="user",reason!="invalid_npm_path",reason!="internal_rate_limit"}[5m]))
worker: syncer_sync_start
Repo metadata sync was started
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100710 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (family) (rate(src_repoupdater_syncer_start_sync{family="Syncer.SyncExternalService"}[9h0m0s]))
worker: syncer_sync_duration
95th repositories sync duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100711 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, max by (le, family, success) (rate(src_repoupdater_syncer_sync_duration_seconds_bucket[1m])))
worker: source_duration
95th repositories source duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100712 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, max by (le) (rate(src_repoupdater_source_duration_seconds_bucket[1m])))
worker: syncer_synced_repos
Repositories synced
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100720 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(rate(src_repoupdater_syncer_synced_repos_total[1m]))
worker: sourced_repos
Repositories sourced
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100721 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(rate(src_repoupdater_source_repos_total[1m]))
worker: sched_auto_fetch
Repositories scheduled due to hitting a deadline
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100730 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(rate(src_repoupdater_sched_auto_fetch[1m]))
worker: sched_manual_fetch
Repositories scheduled due to user traffic
Check worker logs if this value is persistently high. This does not indicate anything if there are no user added code hosts.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100731 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(rate(src_repoupdater_sched_manual_fetch[1m]))
worker: sched_loops
Scheduler loops
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100740 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(rate(src_repoupdater_sched_loops[1m]))
worker: src_repoupdater_stale_repos
Repos that haven't been fetched in more than 8 hours
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100741 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_repoupdater_stale_repos)
worker: sched_error
Repositories schedule error rate
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100742 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(rate(src_repoupdater_sched_error[1m]))
Worker: Repo state syncer
worker: state_syncer_running
State syncer is running
1, if the state syncer is currently running
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100800 on your Sourcegraph instance.
Technical details
Query:
SHELLmax (src_repo_statesyncer_running)
worker: repos_deleted_total
Total number of repos deleted
The total number of repos deleted across all gitservers by the state syncer. A high number here is not necessarily an issue, dig deeper into the other charts in this section to make a call if those deletions were correct.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100801 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_repo_statesyncer_repos_deleted)
worker: repos_deleted_from_primary_total
Total number of repos deleted from primary
The total number of repos deleted from the primary shard. Check the reasons for why they were deleted.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100802 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (reason) (src_repo_statesyncer_repos_deleted{is_primary="true"})
worker: repos_deleted_from_secondary_total
Total number of repos deleted from secondary
The total number of repos deleted from secondary shards. Check the reasons for why they were deleted.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100803 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (reason) (src_repo_statesyncer_repos_deleted{is_primary="false"})
Worker: External services
worker: src_repoupdater_external_services_total
The total number of external services
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100900 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_repoupdater_external_services_total)
worker: repoupdater_queued_sync_jobs_total
The total number of queued sync jobs
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100910 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_repoupdater_queued_sync_jobs_total)
worker: repoupdater_completed_sync_jobs_total
The total number of completed sync jobs
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100911 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_repoupdater_completed_sync_jobs_total)
worker: repoupdater_errored_sync_jobs_percentage
The percentage of external services that have failed their most recent sync
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100912 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_repoupdater_errored_sync_jobs_percentage)
worker: github_graphql_rate_limit_remaining
Remaining calls to GitHub graphql API before hitting the rate limit
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100920 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (src_github_rate_limit_remaining_v2{resource="graphql"})
worker: github_rest_rate_limit_remaining
Remaining calls to GitHub rest API before hitting the rate limit
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100921 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (src_github_rate_limit_remaining_v2{resource="rest"})
worker: github_search_rate_limit_remaining
Remaining calls to GitHub search API before hitting the rate limit
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100922 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (src_github_rate_limit_remaining_v2{resource="search"})
worker: github_graphql_rate_limit_wait_duration
Time spent waiting for the GitHub graphql API rate limiter
Indicates how long we`re waiting on the rate limit once it has been exceeded
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100930 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(name) (rate(src_github_rate_limit_wait_duration_seconds{resource="graphql"}[5m]))
worker: github_rest_rate_limit_wait_duration
Time spent waiting for the GitHub rest API rate limiter
Indicates how long we`re waiting on the rate limit once it has been exceeded
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100931 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(name) (rate(src_github_rate_limit_wait_duration_seconds{resource="rest"}[5m]))
worker: github_search_rate_limit_wait_duration
Time spent waiting for the GitHub search API rate limiter
Indicates how long we`re waiting on the rate limit once it has been exceeded
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100932 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(name) (rate(src_github_rate_limit_wait_duration_seconds{resource="search"}[5m]))
worker: gitlab_rest_rate_limit_remaining
Remaining calls to GitLab rest API before hitting the rate limit
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100940 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (src_gitlab_rate_limit_remaining{resource="rest"})
worker: gitlab_rest_rate_limit_wait_duration
Time spent waiting for the GitLab rest API rate limiter
Indicates how long we`re waiting on the rate limit once it has been exceeded
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100941 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (rate(src_gitlab_rate_limit_wait_duration_seconds{resource="rest"}[5m]))
worker: src_internal_rate_limit_wait_duration_bucket
95th percentile time spent successfully waiting on our internal rate limiter
Indicates how long we`re waiting on our internal rate limiter when communicating with a code host
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100950 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_internal_rate_limit_wait_duration_bucket{failed="false"}[5m])) by (le, urn))
worker: src_internal_rate_limit_wait_error_count
Rate of failures waiting on our internal rate limiter
The rate at which we fail our internal rate limiter.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=100951 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (urn) (rate(src_internal_rate_limit_wait_duration_count{failed="true"}[5m]))
Worker: Permissions
worker: user_success_syncs_total
Total number of user permissions syncs
Indicates the total number of user permissions sync completed.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_repo_perms_syncer_success_syncs{type="user"})
worker: user_success_syncs
Number of user permissions syncs [5m]
Indicates the number of users permissions syncs completed.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_repo_perms_syncer_success_syncs{type="user"}[5m]))
worker: user_initial_syncs
Number of first user permissions syncs [5m]
Indicates the number of permissions syncs done for the first time for the user.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101002 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_repo_perms_syncer_initial_syncs{type="user"}[5m]))
worker: repo_success_syncs_total
Total number of repo permissions syncs
Indicates the total number of repo permissions sync completed.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_repo_perms_syncer_success_syncs{type="repo"})
worker: repo_success_syncs
Number of repo permissions syncs over 5m
Indicates the number of repos permissions syncs completed.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_repo_perms_syncer_success_syncs{type="repo"}[5m]))
worker: repo_initial_syncs
Number of first repo permissions syncs over 5m
Indicates the number of permissions syncs done for the first time for the repo.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101012 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_repo_perms_syncer_initial_syncs{type="repo"}[5m]))
worker: users_consecutive_sync_delay
Max duration between two consecutive permissions sync for user
Indicates the max delay between two consecutive permissions sync for a user during the period.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101020 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(max_over_time (src_repo_perms_syncer_perms_consecutive_sync_delay{type="user"} [1m]))
worker: repos_consecutive_sync_delay
Max duration between two consecutive permissions sync for repo
Indicates the max delay between two consecutive permissions sync for a repo during the period.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101021 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(max_over_time (src_repo_perms_syncer_perms_consecutive_sync_delay{type="repo"} [1m]))
worker: users_first_sync_delay
Max duration between user creation and first permissions sync
Indicates the max delay between user creation and their permissions sync
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101030 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(max_over_time(src_repo_perms_syncer_perms_first_sync_delay{type="user"}[1m]))
worker: repos_first_sync_delay
Max duration between repo creation and first permissions sync over 1m
Indicates the max delay between repo creation and their permissions sync
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101031 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(max_over_time(src_repo_perms_syncer_perms_first_sync_delay{type="repo"}[1m]))
worker: permissions_found_count
Number of permissions found during user/repo permissions sync
Indicates the number permissions found during users/repos permissions sync.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101040 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (type) (src_repo_perms_syncer_perms_found)
worker: permissions_found_avg
Average number of permissions found during permissions sync per user/repo
Indicates the average number permissions found during permissions sync per user/repo.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101041 on your Sourcegraph instance.
Technical details
Query:
SHELLavg by (type) (src_repo_perms_syncer_perms_found)
worker: perms_syncer_outdated_perms
Number of entities with outdated permissions
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101050 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (type) (src_repo_perms_syncer_outdated_perms)
worker: perms_syncer_sync_duration
95th permissions sync duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101060 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, max by (le, type) (rate(src_repo_perms_syncer_sync_duration_seconds_bucket[1m])))
worker: perms_syncer_sync_errors
Permissions sync error rate
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101070 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (type) (ceil(rate(src_repo_perms_syncer_sync_errors_total[1m])))
worker: perms_syncer_scheduled_repos_total
Total number of repos scheduled for permissions sync
Indicates how many repositories have been scheduled for a permissions sync. More about repository permissions synchronization here
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101071 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(rate(src_repo_perms_syncer_schedule_repos_total[1m]))
Worker: Gitserver: Gitserver Client
worker: gitserver_client_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_total{job=~"^worker.*"}[5m]))
worker: gitserver_client_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^worker.*"}[5m]))
worker: gitserver_client_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^worker.*"}[5m]))
worker: gitserver_client_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^worker.*"}[5m])) / (sum(increase(src_gitserver_client_total{job=~"^worker.*"}[5m])) + sum(increase(src_gitserver_client_errors_total{job=~"^worker.*"}[5m]))) * 100
worker: gitserver_client_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101110 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_total{job=~"^worker.*"}[5m]))
worker: gitserver_client_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op,scope)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^worker.*"}[5m])))
worker: gitserver_client_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101112 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^worker.*"}[5m]))
worker: gitserver_client_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101113 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^worker.*"}[5m])) / (sum by (op,scope)(increase(src_gitserver_client_total{job=~"^worker.*"}[5m])) + sum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^worker.*"}[5m]))) * 100
Worker: Gitserver: Gitserver Repository Service Client
worker: gitserver_repositoryservice_client_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_repositoryservice_client_total{job=~"^worker.*"}[5m]))
worker: gitserver_repositoryservice_client_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_repositoryservice_client_duration_seconds_bucket{job=~"^worker.*"}[5m]))
worker: gitserver_repositoryservice_client_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^worker.*"}[5m]))
worker: gitserver_repositoryservice_client_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^worker.*"}[5m])) / (sum(increase(src_gitserver_repositoryservice_client_total{job=~"^worker.*"}[5m])) + sum(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^worker.*"}[5m]))) * 100
worker: gitserver_repositoryservice_client_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_repositoryservice_client_total{job=~"^worker.*"}[5m]))
worker: gitserver_repositoryservice_client_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op,scope)(rate(src_gitserver_repositoryservice_client_duration_seconds_bucket{job=~"^worker.*"}[5m])))
worker: gitserver_repositoryservice_client_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^worker.*"}[5m]))
worker: gitserver_repositoryservice_client_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101213 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^worker.*"}[5m])) / (sum by (op,scope)(increase(src_gitserver_repositoryservice_client_total{job=~"^worker.*"}[5m])) + sum by (op,scope)(increase(src_gitserver_repositoryservice_client_errors_total{job=~"^worker.*"}[5m]))) * 100
Worker: Batches: dbstore stats
worker: batches_dbstore_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_dbstore_total{job=~"^worker.*"}[5m]))
worker: batches_dbstore_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101301 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_batches_dbstore_duration_seconds_bucket{job=~"^worker.*"}[5m]))
worker: batches_dbstore_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101302 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_dbstore_errors_total{job=~"^worker.*"}[5m]))
worker: batches_dbstore_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101303 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_dbstore_errors_total{job=~"^worker.*"}[5m])) / (sum(increase(src_batches_dbstore_total{job=~"^worker.*"}[5m])) + sum(increase(src_batches_dbstore_errors_total{job=~"^worker.*"}[5m]))) * 100
worker: batches_dbstore_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101310 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_dbstore_total{job=~"^worker.*"}[5m]))
worker: batches_dbstore_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101311 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_batches_dbstore_duration_seconds_bucket{job=~"^worker.*"}[5m])))
worker: batches_dbstore_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101312 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_dbstore_errors_total{job=~"^worker.*"}[5m]))
worker: batches_dbstore_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101313 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_dbstore_errors_total{job=~"^worker.*"}[5m])) / (sum by (op)(increase(src_batches_dbstore_total{job=~"^worker.*"}[5m])) + sum by (op)(increase(src_batches_dbstore_errors_total{job=~"^worker.*"}[5m]))) * 100
Worker: Batches: service stats
worker: batches_service_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_service_total{job=~"^worker.*"}[5m]))
worker: batches_service_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_batches_service_duration_seconds_bucket{job=~"^worker.*"}[5m]))
worker: batches_service_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101402 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_service_errors_total{job=~"^worker.*"}[5m]))
worker: batches_service_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101403 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_service_errors_total{job=~"^worker.*"}[5m])) / (sum(increase(src_batches_service_total{job=~"^worker.*"}[5m])) + sum(increase(src_batches_service_errors_total{job=~"^worker.*"}[5m]))) * 100
worker: batches_service_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101410 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_service_total{job=~"^worker.*"}[5m]))
worker: batches_service_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101411 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_batches_service_duration_seconds_bucket{job=~"^worker.*"}[5m])))
worker: batches_service_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101412 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_service_errors_total{job=~"^worker.*"}[5m]))
worker: batches_service_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101413 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_service_errors_total{job=~"^worker.*"}[5m])) / (sum by (op)(increase(src_batches_service_total{job=~"^worker.*"}[5m])) + sum by (op)(increase(src_batches_service_errors_total{job=~"^worker.*"}[5m]))) * 100
Worker: Codeinsights: insights queue processor
worker: query_runner_worker_handlers
Handler active handlers
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_query_runner_worker_processor_handlers{job=~"^worker.*"})
worker: query_runner_worker_processor_total
Handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101510 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_query_runner_worker_processor_total{job=~"^worker.*"}[5m]))
worker: query_runner_worker_processor_99th_percentile_duration
Aggregate successful handler operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101511 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_query_runner_worker_processor_duration_seconds_bucket{job=~"^worker.*"}[5m]))
worker: query_runner_worker_processor_errors_total
Handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101512 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_query_runner_worker_processor_errors_total{job=~"^worker.*"}[5m]))
worker: query_runner_worker_processor_error_rate
Handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101513 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_query_runner_worker_processor_errors_total{job=~"^worker.*"}[5m])) / (sum(increase(src_query_runner_worker_processor_total{job=~"^worker.*"}[5m])) + sum(increase(src_query_runner_worker_processor_errors_total{job=~"^worker.*"}[5m]))) * 100
Worker: Codeinsights: dbstore stats
worker: workerutil_dbworker_store_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_total{domain='insights_query_runner_jobs',job=~"^worker.*"}[5m]))
worker: workerutil_dbworker_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101601 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_workerutil_dbworker_store_duration_seconds_bucket{domain='insights_query_runner_jobs',job=~"^worker.*"}[5m]))
worker: workerutil_dbworker_store_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101602 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_errors_total{domain='insights_query_runner_jobs',job=~"^worker.*"}[5m]))
worker: workerutil_dbworker_store_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101603 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_errors_total{domain='insights_query_runner_jobs',job=~"^worker.*"}[5m])) / (sum(increase(src_workerutil_dbworker_store_total{domain='insights_query_runner_jobs',job=~"^worker.*"}[5m])) + sum(increase(src_workerutil_dbworker_store_errors_total{domain='insights_query_runner_jobs',job=~"^worker.*"}[5m]))) * 100
worker: workerutil_dbworker_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101610 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_workerutil_dbworker_store_total{domain='insights_query_runner_jobs',job=~"^worker.*"}[5m]))
worker: workerutil_dbworker_store_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101611 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_workerutil_dbworker_store_duration_seconds_bucket{domain='insights_query_runner_jobs',job=~"^worker.*"}[5m])))
worker: workerutil_dbworker_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101612 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_workerutil_dbworker_store_errors_total{domain='insights_query_runner_jobs',job=~"^worker.*"}[5m]))
worker: workerutil_dbworker_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101613 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_workerutil_dbworker_store_errors_total{domain='insights_query_runner_jobs',job=~"^worker.*"}[5m])) / (sum by (op)(increase(src_workerutil_dbworker_store_total{domain='insights_query_runner_jobs',job=~"^worker.*"}[5m])) + sum by (op)(increase(src_workerutil_dbworker_store_errors_total{domain='insights_query_runner_jobs',job=~"^worker.*"}[5m]))) * 100
Worker: Completion Credits Entitlement Usage Aggregator: Completion credits entitlement usage aggregations
worker: completioncredits_aggregator_total
Completion credits entitlement usage aggregator operations every 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101700 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_completioncredits_aggregator_total{job=~"^worker.*"}[30m]))
worker: completioncredits_aggregator_99th_percentile_duration
Aggregate successful completion credits entitlement usage aggregator operation duration distribution over 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101701 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_completioncredits_aggregator_duration_seconds_bucket{job=~"^worker.*"}[30m]))
worker: completioncredits_aggregator_errors_total
Completion credits entitlement usage aggregator operation errors every 30m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101702 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_completioncredits_aggregator_errors_total{job=~"^worker.*"}[30m]))
worker: completioncredits_aggregator_error_rate
Completion credits entitlement usage aggregator operation error rate over 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101703 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_completioncredits_aggregator_errors_total{job=~"^worker.*"}[30m])) / (sum(increase(src_completioncredits_aggregator_total{job=~"^worker.*"}[30m])) + sum(increase(src_completioncredits_aggregator_errors_total{job=~"^worker.*"}[30m]))) * 100
Worker: Periodic Goroutines
worker: running_goroutines
Number of currently running periodic goroutines
The number of currently running periodic goroutines by name and job.
A value of 0 indicates the routine isnt running currently, it awaits its next schedule.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101800 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (src_periodic_goroutine_running{job=~".*worker.*"})
worker: goroutine_success_rate
Success rate for periodic goroutine executions
The rate of successful executions of each periodic goroutine. A low or zero value could indicate that a routine is stalled or encountering errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101801 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_total{job=~".*worker.*"}[5m]))
worker: goroutine_error_rate
Error rate for periodic goroutine executions
The rate of errors encountered by each periodic goroutine. A sustained high error rate may indicate a problem with the routine`s configuration or dependencies.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101810 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_errors_total{job=~".*worker.*"}[5m]))
worker: goroutine_error_percentage
Percentage of periodic goroutine executions that result in errors
The percentage of executions that result in errors for each periodic goroutine. A value above 5% indicates that a significant portion of routine executions are failing.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101811 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_errors_total{job=~".*worker.*"}[5m])) / sum by (name, job_name) (rate(src_periodic_goroutine_total{job=~".*worker.*"}[5m]) > 0) * 100
worker: goroutine_handler_duration
95th percentile handler execution time
The 95th percentile execution time for each periodic goroutine handler. Longer durations might indicate increased load or processing time.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101820 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job_name, le) (rate(src_periodic_goroutine_duration_seconds_bucket{job=~".*worker.*"}[5m])))
worker: goroutine_loop_duration
95th percentile loop cycle time
The 95th percentile loop cycle time for each periodic goroutine (excluding sleep time). This represents how long a complete loop iteration takes before sleeping for the next interval.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101821 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job_name, le) (rate(src_periodic_goroutine_loop_duration_seconds_bucket{job=~".*worker.*"}[5m])))
worker: tenant_processing_duration
95th percentile tenant processing time
The 95th percentile processing time for individual tenants within periodic goroutines. Higher values indicate that tenant processing is taking longer and may affect overall performance.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101830 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job_name, le) (rate(src_periodic_goroutine_tenant_duration_seconds_bucket{job=~".*worker.*"}[5m])))
worker: tenant_processing_max
Maximum tenant processing time
The maximum processing time for individual tenants within periodic goroutines. Consistently high values might indicate problematic tenants or inefficient processing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101831 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name, job_name) (rate(src_periodic_goroutine_tenant_duration_seconds_sum{job=~".*worker.*"}[5m]) / rate(src_periodic_goroutine_tenant_duration_seconds_count{job=~".*worker.*"}[5m]))
worker: tenant_count
Number of tenants processed per routine
The number of tenants processed by each periodic goroutine. Unexpected changes can indicate tenant configuration issues or scaling events.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101840 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name, job_name) (src_periodic_goroutine_tenant_count{job=~".*worker.*"})
worker: tenant_success_rate
Rate of successful tenant processing operations
The rate of successful tenant processing operations. A healthy routine should maintain a consistent processing rate.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101841 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_tenant_success_total{job=~".*worker.*"}[5m]))
worker: tenant_error_rate
Rate of tenant processing errors
The rate of tenant processing operations that result in errors. Consistent errors indicate problems with specific tenants.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101850 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_tenant_errors_total{job=~".*worker.*"}[5m]))
worker: tenant_error_percentage
Percentage of tenant operations resulting in errors
The percentage of tenant operations that result in errors. Values above 5% indicate significant tenant processing problems.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101851 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum by (name, job_name) (rate(src_periodic_goroutine_tenant_errors_total{job=~".*worker.*"}[5m])) / (sum by (name, job_name) (rate(src_periodic_goroutine_tenant_success_total{job=~".*worker.*"}[5m])) + sum by (name, job_name) (rate(src_periodic_goroutine_tenant_errors_total{job=~".*worker.*"}[5m])))) * 100
Worker: Database connections
worker: max_open_conns
Maximum open
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101900 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_max_open{app_name="worker"})
worker: open_conns
Established
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101901 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_open{app_name="worker"})
worker: in_use
Used
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101910 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_in_use{app_name="worker"})
worker: idle
Idle
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101911 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_idle{app_name="worker"})
worker: mean_blocked_seconds_per_conn_request
Mean blocked seconds per conn request
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101920 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_blocked_seconds{app_name="worker"}[5m])) / sum by (app_name, db_name) (increase(src_pgsql_conns_waited_for{app_name="worker"}[5m]))
worker: closed_max_idle
Closed by SetMaxIdleConns
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101930 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle{app_name="worker"}[5m]))
worker: closed_max_lifetime
Closed by SetConnMaxLifetime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101931 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_lifetime{app_name="worker"}[5m]))
worker: closed_max_idle_time
Closed by SetConnMaxIdleTime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=101932 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle_time{app_name="worker"}[5m]))
Worker: Worker (CPU, Memory)
worker: cpu_usage_percentage
CPU usage
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102000 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^worker.*"}
worker: memory_usage_percentage
Memory usage percentage (total)
An estimate for the active memory in use, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102001 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^worker.*"}
worker: memory_working_set_bytes
Memory usage bytes (total)
An estimate for the active memory in use in bytes, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102002 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_memory_working_set_bytes{name=~"^worker.*"})
worker: memory_rss
Memory (RSS)
The total anonymous memory in use by the application, which includes Go stack and heap. This memory is is non-reclaimable, and high usage may trigger OOM kills. Note: the metric is named RSS because to match the cadvisor name, but anonymous is more accurate."
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102010 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_rss{name=~"^worker.*"} / container_spec_memory_limit_bytes{name=~"^worker.*"}) by (name) * 100.0
worker: memory_total_active_file
Memory usage (active file)
This metric shows the total active file-backed memory currently in use by the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102011 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_total_active_file_bytes{name=~"^worker.*"} / container_spec_memory_limit_bytes{name=~"^worker.*"}) by (name) * 100.0
worker: memory_kernel_usage
Memory usage (kernel)
The kernel usage metric shows the amount of memory used by the kernel on behalf of the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102012 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_kernel_usage{name=~"^worker.*"} / container_spec_memory_limit_bytes{name=~"^worker.*"}) by (name) * 100.0
Worker: Container monitoring (not available on server)
worker: container_missing
Container missing
This value is the number of times a container has not been seen for more than one minute. If you observe this value change independent of deployment events (such as an upgrade), it could indicate pods are being OOM killed or terminated for some other reasons.
- Kubernetes:
- Determine if the pod was OOM killed using
kubectl describe pod worker(look forOOMKilled: true) and, if so, consider increasing the memory limit in the relevantDeployment.yaml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingkubectl logs -p worker.
- Determine if the pod was OOM killed using
- Docker Compose:
- Determine if the pod was OOM killed using
docker inspect -f '\{\{json .State\}\}' worker(look for"OOMKilled":true) and, if so, consider increasing the memory limit of the worker container indocker-compose.yml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingdocker logs worker(note this will include logs from the previous and currently running container).
- Determine if the pod was OOM killed using
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102100 on your Sourcegraph instance.
Technical details
Query:
SHELLcount by(name) ((time() - container_last_seen{name=~"^worker.*"}) > 60)
worker: container_cpu_usage
Container cpu usage total (1m average) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102101 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^worker.*"}
worker: container_memory_usage
Container memory usage by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102102 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^worker.*"}
worker: fs_io_operations
Filesystem reads and writes rate by instance over 1h
This value indicates the number of filesystem read and write operations by containers of this service. When extremely high, this can indicate a resource usage problem, or can cause problems with the service itself, especially if high values or spikes correlate with {{CONTAINER_NAME}} issues.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(name) (rate(container_fs_reads_total{name=~"^worker.*"}[1h]) + rate(container_fs_writes_total{name=~"^worker.*"}[1h]))
Worker: Provisioning indicators (not available on server)
worker: provisioning_container_cpu_usage_long_term
Container cpu usage total (90th percentile over 1d) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102200 on your Sourcegraph instance.
Technical details
Query:
SHELLquantile_over_time(0.9, cadvisor_container_cpu_usage_percentage_total{name=~"^worker.*"}[1d])
worker: provisioning_container_memory_usage_long_term
Container memory usage (1d maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102201 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^worker.*"}[1d])
worker: provisioning_container_cpu_usage_short_term
Container cpu usage total (5m maximum) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102210 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^worker.*"}[5m])
worker: provisioning_container_memory_usage_short_term
Container memory usage (5m maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102211 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^worker.*"}[5m])
worker: container_oomkill_events_total
Container OOMKILL events total by instance
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102212 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^worker.*"})
Worker: Golang runtime monitoring
worker: go_goroutines
Maximum active goroutines
A high value here indicates a possible goroutine leak.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102300 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_goroutines{job=~".*worker"})
worker: go_gc_duration_seconds
Maximum go garbage collection duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102301 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_gc_duration_seconds{job=~".*worker"})
Worker: Kubernetes monitoring (only available on Kubernetes)
worker: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*worker"}) / count by (app) (up{app=~".*worker"}) * 100
Worker: Own: repo indexer dbstore
worker: workerutil_dbworker_store_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_total{domain='own_background_worker_store',job=~"^worker.*"}[5m]))
worker: workerutil_dbworker_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102501 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_workerutil_dbworker_store_duration_seconds_bucket{domain='own_background_worker_store',job=~"^worker.*"}[5m]))
worker: workerutil_dbworker_store_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102502 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_errors_total{domain='own_background_worker_store',job=~"^worker.*"}[5m]))
worker: workerutil_dbworker_store_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102503 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_errors_total{domain='own_background_worker_store',job=~"^worker.*"}[5m])) / (sum(increase(src_workerutil_dbworker_store_total{domain='own_background_worker_store',job=~"^worker.*"}[5m])) + sum(increase(src_workerutil_dbworker_store_errors_total{domain='own_background_worker_store',job=~"^worker.*"}[5m]))) * 100
worker: workerutil_dbworker_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102510 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_workerutil_dbworker_store_total{domain='own_background_worker_store',job=~"^worker.*"}[5m]))
worker: workerutil_dbworker_store_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102511 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_workerutil_dbworker_store_duration_seconds_bucket{domain='own_background_worker_store',job=~"^worker.*"}[5m])))
worker: workerutil_dbworker_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102512 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_workerutil_dbworker_store_errors_total{domain='own_background_worker_store',job=~"^worker.*"}[5m]))
worker: workerutil_dbworker_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102513 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_workerutil_dbworker_store_errors_total{domain='own_background_worker_store',job=~"^worker.*"}[5m])) / (sum by (op)(increase(src_workerutil_dbworker_store_total{domain='own_background_worker_store',job=~"^worker.*"}[5m])) + sum by (op)(increase(src_workerutil_dbworker_store_errors_total{domain='own_background_worker_store',job=~"^worker.*"}[5m]))) * 100
Worker: Own: repo indexer worker queue
worker: own_background_worker_handlers
Handler active handlers
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_own_background_worker_processor_handlers{job=~"^worker.*"})
worker: own_background_worker_processor_total
Handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102610 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_own_background_worker_processor_total{job=~"^worker.*"}[5m]))
worker: own_background_worker_processor_99th_percentile_duration
Aggregate successful handler operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102611 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_own_background_worker_processor_duration_seconds_bucket{job=~"^worker.*"}[5m]))
worker: own_background_worker_processor_errors_total
Handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102612 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_own_background_worker_processor_errors_total{job=~"^worker.*"}[5m]))
worker: own_background_worker_processor_error_rate
Handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102613 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_own_background_worker_processor_errors_total{job=~"^worker.*"}[5m])) / (sum(increase(src_own_background_worker_processor_total{job=~"^worker.*"}[5m])) + sum(increase(src_own_background_worker_processor_errors_total{job=~"^worker.*"}[5m]))) * 100
Worker: Own: index job scheduler
worker: own_background_index_scheduler_total
Own index job scheduler operations every 10m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102700 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_own_background_index_scheduler_total{job=~"^worker.*"}[10m]))
worker: own_background_index_scheduler_99th_percentile_duration
99th percentile successful own index job scheduler operation duration over 10m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102701 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_own_background_index_scheduler_duration_seconds_bucket{job=~"^worker.*"}[10m])))
worker: own_background_index_scheduler_errors_total
Own index job scheduler operation errors every 10m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102702 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_own_background_index_scheduler_errors_total{job=~"^worker.*"}[10m]))
worker: own_background_index_scheduler_error_rate
Own index job scheduler operation error rate over 10m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102703 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_own_background_index_scheduler_errors_total{job=~"^worker.*"}[10m])) / (sum by (op)(increase(src_own_background_index_scheduler_total{job=~"^worker.*"}[10m])) + sum by (op)(increase(src_own_background_index_scheduler_errors_total{job=~"^worker.*"}[10m]))) * 100
Worker: Site configuration client update latency
worker: worker_site_configuration_duration_since_last_successful_update_by_instance
Duration since last successful site configuration update (by instance)
The duration since the configuration client used by the "worker" service last successfully updated its site configuration. Long durations could indicate issues updating the site configuration.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102800 on your Sourcegraph instance.
Technical details
Query:
SHELLsrc_conf_client_time_since_last_successful_update_seconds{job=~`^worker.*`,instance=~`${instance:regex}`}
worker: worker_site_configuration_duration_since_last_successful_update_by_instance
Maximum duration since last successful site configuration update (all "worker" instances)
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/worker/worker?viewPanel=102801 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(max_over_time(src_conf_client_time_since_last_successful_update_seconds{job=~`^worker.*`,instance=~`${instance:regex}`}[1m]))
Searcher
Performs unindexed searches (diff and commit search, text search for unindexed branches).
To see this dashboard, visit /-/debug/grafana/d/searcher/searcher on your Sourcegraph instance.
searcher: traffic
Requests per second by code over 10m
This graph is the average number of requests per second searcher is experiencing over the last 10 minutes.
The code is the HTTP Status code. 200 is success. We have a special code "canceled" which is common when doing a large search request and we find enough results before searching all possible repos.
Note: A search query is translated into an unindexed search query per unique (repo, commit). This means a single user query may result in thousands of requests to searcher.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (code) (rate(searcher_service_request_total{instance=~`${instance:regex}`}[10m]))
searcher: replica_traffic
Requests per second per replica over 10m
This graph is the average number of requests per second searcher is experiencing over the last 10 minutes broken down per replica.
The code is the HTTP Status code. 200 is success. We have a special code "canceled" which is common when doing a large search request and we find enough results before searching all possible repos.
Note: A search query is translated into an unindexed search query per unique (repo, commit). This means a single user query may result in thousands of requests to searcher.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance) (rate(searcher_service_request_total{instance=~`${instance:regex}`}[10m]))
searcher: concurrent_requests
Amount of in-flight unindexed search requests (per instance)
This graph is the amount of in-flight unindexed search requests per instance. Consistently high numbers here indicate you may need to scale out searcher.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance) (searcher_service_running{instance=~`${instance:regex}`})
searcher: unindexed_search_request_errors
Unindexed search request errors every 5m by code
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (code)(increase(searcher_service_request_total{code!="200",code!="canceled",instance=~`${instance:regex}`}[5m])) / ignoring(code) group_left sum(increase(searcher_service_request_total{instance=~`${instance:regex}`}[5m])) * 100
Searcher: Cache store
searcher: store_fetching
Amount of in-flight unindexed search requests fetching code from gitserver (per instance)
Before we can search a commit we fetch the code from gitserver then cache it for future search requests. This graph is the current number of search requests which are in the state of fetching code from gitserver.
Generally this number should remain low since fetching code is fast, but expect bursts. In the case of instances with a monorepo you would expect this number to stay low for the duration of fetching the code (which in some cases can take many minutes).
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance) (searcher_store_fetching{instance=~`${instance:regex}`})
searcher: store_fetching_waiting
Amount of in-flight unindexed search requests waiting to fetch code from gitserver (per instance)
We limit the number of requests which can fetch code to prevent overwhelming gitserver. This gauge is the number of requests waiting to be allowed to speak to gitserver.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance) (searcher_store_fetch_queue_size{instance=~`${instance:regex}`})
searcher: store_fetching_fail
Amount of unindexed search requests that failed while fetching code from gitserver over 10m (per instance)
This graph should be zero since fetching happens in the background and will not be influenced by user timeouts/etc. Expected upticks in this graph are during gitserver rollouts. If you regularly see this graph have non-zero values please reach out to support.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance) (rate(searcher_store_fetch_failed{instance=~`${instance:regex}`}[10m]))
Searcher: Index use
searcher: searcher_hybrid_final_state_total
Hybrid search final state over 10m
This graph is about our interactions with the search index (zoekt) to help complete unindexed search requests. Searcher will use indexed search for the files that have not changed between the unindexed commit and the index.
This graph should mostly be "success". The next most common state should be "search-canceled" which happens when result limits are hit or the user starts a new search. Finally the next most common should be "diff-too-large", which happens if the commit is too far from the indexed commit. Otherwise other state should be rare and likely are a sign for further investigation.
Note: On sourcegraph.com "zoekt-list-missing" is also common due to it indexing a subset of repositories. Otherwise every other state should occur rarely.
For a full list of possible state see recordHybridFinalState.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (state)(increase(searcher_hybrid_final_state_total{instance=~`${instance:regex}`}[10m]))
searcher: searcher_hybrid_retry_total
Hybrid search retrying over 10m
Expectation is that this graph should mostly be 0. It will trigger if a user manages to do a search and the underlying index changes while searching or Zoekt goes down. So occasional bursts can be expected, but if this graph is regularly above 0 it is a sign for further investigation.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (reason)(increase(searcher_hybrid_retry_total{instance=~`${instance:regex}`}[10m]))
Searcher: Cache disk I/O metrics
searcher: cache_disk_reads_sec
Read request rate over 1m (per instance)
The number of read requests that were issued to the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), searcher could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device searcher is using, not the load searcher is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_completed_total{instance=~`node-exporter.*`}[1m])))))
searcher: cache_disk_writes_sec
Write request rate over 1m (per instance)
The number of write requests that were issued to the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), searcher could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device searcher is using, not the load searcher is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_completed_total{instance=~`node-exporter.*`}[1m])))))
searcher: cache_disk_read_throughput
Read throughput over 1m (per instance)
The amount of data that was read from the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), searcher could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device searcher is using, not the load searcher is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100310 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_read_bytes_total{instance=~`node-exporter.*`}[1m])))))
searcher: cache_disk_write_throughput
Write throughput over 1m (per instance)
The amount of data that was written to the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), searcher could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device searcher is using, not the load searcher is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100311 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_written_bytes_total{instance=~`node-exporter.*`}[1m])))))
searcher: cache_disk_read_duration
Average read duration over 1m (per instance)
The average time for read requests issued to the device to be served. This includes the time spent by the requests in queue and the time spent servicing them.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), searcher could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device searcher is using, not the load searcher is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100320 on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_read_time_seconds_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_completed_total{instance=~`node-exporter.*`}[1m])))))))
searcher: cache_disk_write_duration
Average write duration over 1m (per instance)
The average time for write requests issued to the device to be served. This includes the time spent by the requests in queue and the time spent servicing them.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), searcher could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device searcher is using, not the load searcher is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100321 on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_write_time_seconds_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_completed_total{instance=~`node-exporter.*`}[1m])))))))
searcher: cache_disk_read_request_size
Average read request size over 1m (per instance)
The average size of read requests that were issued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), searcher could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device searcher is using, not the load searcher is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100330 on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_read_bytes_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_completed_total{instance=~`node-exporter.*`}[1m])))))))
searcher: cache_disk_write_request_size)
Average write request size over 1m (per instance)
The average size of write requests that were issued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), searcher could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device searcher is using, not the load searcher is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100331 on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_written_bytes_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_completed_total{instance=~`node-exporter.*`}[1m])))))))
searcher: cache_disk_reads_merged_sec
Merged read request rate over 1m (per instance)
The number of read requests merged per second that were queued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), searcher could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device searcher is using, not the load searcher is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100340 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_merged_total{instance=~`node-exporter.*`}[1m])))))
searcher: cache_disk_writes_merged_sec
Merged writes request rate over 1m (per instance)
The number of write requests merged per second that were queued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), searcher could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device searcher is using, not the load searcher is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100341 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_merged_total{instance=~`node-exporter.*`}[1m])))))
searcher: cache_disk_average_queue_size
Average queue size over 1m (per instance)
The number of I/O operations that were being queued or being serviced. See https://blog.actorsfit.com/a?ID=00200-428fa2ac-e338-4540-848c-af9a3eb1ebd2 for background (avgqu-sz).
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), searcher could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device searcher is using, not the load searcher is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100350 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (searcher_mount_point_info{mount_name="cacheDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_io_time_weighted_seconds_total{instance=~`node-exporter.*`}[1m])))))
Searcher: Searcher GRPC server metrics
searcher: searcher_grpc_request_rate_all_methods
Request rate across all methods over 2m
The number of gRPC requests received per second across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m]))
searcher: searcher_grpc_request_rate_per_method
Request rate per-method over 2m
The number of gRPC requests received per second broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{grpc_method=~`${searcher_method:regex}`,instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])) by (grpc_method)
searcher: searcher_error_percentage_all_methods
Error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100410 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_code!="OK",instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m]))) / (sum(rate(grpc_server_handled_total{instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m]))) ))
searcher: searcher_grpc_error_percentage_per_method
Error percentage per-method over 2m
The percentage of gRPC requests that fail per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100411 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_method=~`${searcher_method:regex}`,grpc_code!="OK",instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])) by (grpc_method)) / (sum(rate(grpc_server_handled_total{grpc_method=~`${searcher_method:regex}`,instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])) by (grpc_method)) ))
searcher: searcher_p99_response_time_per_method
99th percentile response time per method over 2m
The 99th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100420 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${searcher_method:regex}`,instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])))
searcher: searcher_p90_response_time_per_method
90th percentile response time per method over 2m
The 90th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100421 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${searcher_method:regex}`,instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])))
searcher: searcher_p75_response_time_per_method
75th percentile response time per method over 2m
The 75th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100422 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${searcher_method:regex}`,instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])))
searcher: searcher_p99_9_response_size_per_method
99.9th percentile total response size per method over 2m
The 99.9th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100430 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${searcher_method:regex}`,instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])))
searcher: searcher_p90_response_size_per_method
90th percentile total response size per method over 2m
The 90th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100431 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${searcher_method:regex}`,instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])))
searcher: searcher_p75_response_size_per_method
75th percentile total response size per method over 2m
The 75th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100432 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${searcher_method:regex}`,instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])))
searcher: searcher_p99_9_invididual_sent_message_size_per_method
99.9th percentile individual sent message size per method over 2m
The 99.9th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100440 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${searcher_method:regex}`,instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])))
searcher: searcher_p90_invididual_sent_message_size_per_method
90th percentile individual sent message size per method over 2m
The 90th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100441 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${searcher_method:regex}`,instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])))
searcher: searcher_p75_invididual_sent_message_size_per_method
75th percentile individual sent message size per method over 2m
The 75th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100442 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${searcher_method:regex}`,instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])))
searcher: searcher_grpc_response_stream_message_count_per_method
Average streaming response message count per-method over 2m
The average number of response messages sent during a streaming RPC method, broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100450 on your Sourcegraph instance.
Technical details
Query:
SHELL((sum(rate(grpc_server_msg_sent_total{grpc_type="server_stream",instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])) by (grpc_method))/(sum(rate(grpc_server_started_total{grpc_type="server_stream",instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])) by (grpc_method)))
searcher: searcher_grpc_all_codes_per_method
Response codes rate per-method over 2m
The rate of all generated gRPC response codes per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100460 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_handled_total{grpc_method=~`${searcher_method:regex}`,instance=~`${instance:regex}`,grpc_service=~"searcher.v1.SearcherService"}[2m])) by (grpc_method, grpc_code)
Searcher: Searcher GRPC "internal error" metrics
searcher: searcher_grpc_clients_error_percentage_all_methods
Client baseline error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods (regardless of whether or not there was an internal error), aggregated across all "searcher" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"searcher.v1.SearcherService",grpc_code!="OK"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"searcher.v1.SearcherService"}[2m])))))))
searcher: searcher_grpc_clients_error_percentage_per_method
Client baseline error percentage per-method over 2m
The percentage of gRPC requests that fail per method (regardless of whether or not there was an internal error), aggregated across all "searcher" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100501 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"searcher.v1.SearcherService",grpc_method=~"${searcher_method:regex}",grpc_code!="OK"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"searcher.v1.SearcherService",grpc_method=~"${searcher_method:regex}"}[2m])) by (grpc_method))))))
searcher: searcher_grpc_clients_all_codes_per_method
Client baseline response codes rate per-method over 2m
The rate of all generated gRPC response codes per method (regardless of whether or not there was an internal error), aggregated across all "searcher" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100502 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"searcher.v1.SearcherService",grpc_method=~"${searcher_method:regex}"}[2m])) by (grpc_method, grpc_code))
searcher: searcher_grpc_clients_internal_error_percentage_all_methods
Client-observed gRPC internal error percentage across all methods over 2m
The percentage of gRPC requests that appear to fail due to gRPC internal errors across all methods, aggregated across all "searcher" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "searcher" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100510 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"searcher.v1.SearcherService",grpc_code!="OK",is_internal_error="true"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"searcher.v1.SearcherService"}[2m])))))))
searcher: searcher_grpc_clients_internal_error_percentage_per_method
Client-observed gRPC internal error percentage per-method over 2m
The percentage of gRPC requests that appear to fail to due to gRPC internal errors per method, aggregated across all "searcher" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "searcher" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100511 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"searcher.v1.SearcherService",grpc_method=~"${searcher_method:regex}",grpc_code!="OK",is_internal_error="true"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"searcher.v1.SearcherService",grpc_method=~"${searcher_method:regex}"}[2m])) by (grpc_method))))))
searcher: searcher_grpc_clients_internal_error_all_codes_per_method
Client-observed gRPC internal error response code rate per-method over 2m
The rate of gRPC internal-error response codes per method, aggregated across all "searcher" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "searcher" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100512 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"searcher.v1.SearcherService",is_internal_error="true",grpc_method=~"${searcher_method:regex}"}[2m])) by (grpc_method, grpc_code))
Searcher: Searcher GRPC retry metrics
searcher: searcher_grpc_clients_retry_percentage_across_all_methods
Client retry percentage across all methods over 2m
The percentage of gRPC requests that were retried across all methods, aggregated across all "searcher" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100600 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"searcher.v1.SearcherService",is_retried="true"}[2m])))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"searcher.v1.SearcherService"}[2m])))))))
searcher: searcher_grpc_clients_retry_percentage_per_method
Client retry percentage per-method over 2m
The percentage of gRPC requests that were retried aggregated across all "searcher" clients, broken out per method.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100601 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"searcher.v1.SearcherService",is_retried="true",grpc_method=~"${searcher_method:regex}"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"searcher.v1.SearcherService",grpc_method=~"${searcher_method:regex}"}[2m])) by (grpc_method))))))
searcher: searcher_grpc_clients_retry_count_per_method
Client retry count per-method over 2m
The count of gRPC requests that were retried aggregated across all "searcher" clients, broken out per method
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100602 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"searcher.v1.SearcherService",grpc_method=~"${searcher_method:regex}",is_retried="true"}[2m])) by (grpc_method))
Searcher: Codeintel: Symbols API
searcher: codeintel_symbols_api_total
Aggregate API operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100700 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_symbols_api_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_api_99th_percentile_duration
Aggregate successful API operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100701 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_symbols_api_duration_seconds_bucket{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_api_errors_total
Aggregate API operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100702 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_symbols_api_errors_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_api_error_rate
Aggregate API operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100703 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_symbols_api_errors_total{job=~"^searcher.*"}[5m])) / (sum(increase(src_codeintel_symbols_api_total{job=~"^searcher.*"}[5m])) + sum(increase(src_codeintel_symbols_api_errors_total{job=~"^searcher.*"}[5m]))) * 100
searcher: codeintel_symbols_api_total
API operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100710 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,parseAmount)(increase(src_codeintel_symbols_api_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_api_99th_percentile_duration
99th percentile successful API operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100711 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op,parseAmount)(rate(src_codeintel_symbols_api_duration_seconds_bucket{job=~"^searcher.*"}[5m])))
searcher: codeintel_symbols_api_errors_total
API operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100712 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,parseAmount)(increase(src_codeintel_symbols_api_errors_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_api_error_rate
API operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100713 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,parseAmount)(increase(src_codeintel_symbols_api_errors_total{job=~"^searcher.*"}[5m])) / (sum by (op,parseAmount)(increase(src_codeintel_symbols_api_total{job=~"^searcher.*"}[5m])) + sum by (op,parseAmount)(increase(src_codeintel_symbols_api_errors_total{job=~"^searcher.*"}[5m]))) * 100
Searcher: Codeintel: Symbols parser
searcher: searcher
In-flight parse jobs
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100800 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_codeintel_symbols_parsing{job=~"^searcher.*"})
searcher: searcher
Parser queue size
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100801 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_codeintel_symbols_parse_queue_size{job=~"^searcher.*"})
searcher: searcher
Parse queue timeouts
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100802 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_codeintel_symbols_parse_queue_timeouts_total{job=~"^searcher.*"})
searcher: searcher
Parse failures every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100803 on your Sourcegraph instance.
Technical details
Query:
SHELLrate(src_codeintel_symbols_parse_failed_total{job=~"^searcher.*"}[5m])
searcher: codeintel_symbols_parser_total
Aggregate parser operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100810 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_symbols_parser_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_parser_99th_percentile_duration
Aggregate successful parser operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100811 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_symbols_parser_duration_seconds_bucket{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_parser_errors_total
Aggregate parser operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100812 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_symbols_parser_errors_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_parser_error_rate
Aggregate parser operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100813 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_symbols_parser_errors_total{job=~"^searcher.*"}[5m])) / (sum(increase(src_codeintel_symbols_parser_total{job=~"^searcher.*"}[5m])) + sum(increase(src_codeintel_symbols_parser_errors_total{job=~"^searcher.*"}[5m]))) * 100
searcher: codeintel_symbols_parser_total
Parser operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100820 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_symbols_parser_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_parser_99th_percentile_duration
99th percentile successful parser operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100821 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_symbols_parser_duration_seconds_bucket{job=~"^searcher.*"}[5m])))
searcher: codeintel_symbols_parser_errors_total
Parser operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100822 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_symbols_parser_errors_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_parser_error_rate
Parser operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100823 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_symbols_parser_errors_total{job=~"^searcher.*"}[5m])) / (sum by (op)(increase(src_codeintel_symbols_parser_total{job=~"^searcher.*"}[5m])) + sum by (op)(increase(src_codeintel_symbols_parser_errors_total{job=~"^searcher.*"}[5m]))) * 100
Searcher: Codeintel: Symbols cache janitor
searcher: searcher
Size in bytes of the on-disk cache
no
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100900 on your Sourcegraph instance.
Technical details
Query:
SHELLsrc_diskcache_store_symbols_cache_size_bytes
searcher: searcher
Cache eviction operations every 5m
no
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100901 on your Sourcegraph instance.
Technical details
Query:
SHELLrate(src_diskcache_store_symbols_evictions_total[5m])
searcher: searcher
Cache eviction operation errors every 5m
no
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=100902 on your Sourcegraph instance.
Technical details
Query:
SHELLrate(src_diskcache_store_symbols_errors_total[5m])
Searcher: Codeintel: Symbols repository fetcher
searcher: searcher
In-flight repository fetch operations
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101000 on your Sourcegraph instance.
Technical details
Query:
SHELLsrc_codeintel_symbols_fetching
searcher: searcher
Repository fetch queue size
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101001 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_codeintel_symbols_fetch_queue_size{job=~"^searcher.*"})
searcher: codeintel_symbols_repository_fetcher_total
Aggregate fetcher operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_symbols_repository_fetcher_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_repository_fetcher_99th_percentile_duration
Aggregate successful fetcher operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_symbols_repository_fetcher_duration_seconds_bucket{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_repository_fetcher_errors_total
Aggregate fetcher operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101012 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_symbols_repository_fetcher_errors_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_repository_fetcher_error_rate
Aggregate fetcher operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101013 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_symbols_repository_fetcher_errors_total{job=~"^searcher.*"}[5m])) / (sum(increase(src_codeintel_symbols_repository_fetcher_total{job=~"^searcher.*"}[5m])) + sum(increase(src_codeintel_symbols_repository_fetcher_errors_total{job=~"^searcher.*"}[5m]))) * 100
searcher: codeintel_symbols_repository_fetcher_total
Fetcher operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101020 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_symbols_repository_fetcher_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_repository_fetcher_99th_percentile_duration
99th percentile successful fetcher operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101021 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_symbols_repository_fetcher_duration_seconds_bucket{job=~"^searcher.*"}[5m])))
searcher: codeintel_symbols_repository_fetcher_errors_total
Fetcher operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101022 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_symbols_repository_fetcher_errors_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_repository_fetcher_error_rate
Fetcher operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101023 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_symbols_repository_fetcher_errors_total{job=~"^searcher.*"}[5m])) / (sum by (op)(increase(src_codeintel_symbols_repository_fetcher_total{job=~"^searcher.*"}[5m])) + sum by (op)(increase(src_codeintel_symbols_repository_fetcher_errors_total{job=~"^searcher.*"}[5m]))) * 100
Searcher: Codeintel: Symbols gitserver client
searcher: codeintel_symbols_gitserver_total
Aggregate gitserver client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_symbols_gitserver_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_gitserver_99th_percentile_duration
Aggregate successful gitserver client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_symbols_gitserver_duration_seconds_bucket{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_gitserver_errors_total
Aggregate gitserver client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_symbols_gitserver_errors_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_gitserver_error_rate
Aggregate gitserver client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_symbols_gitserver_errors_total{job=~"^searcher.*"}[5m])) / (sum(increase(src_codeintel_symbols_gitserver_total{job=~"^searcher.*"}[5m])) + sum(increase(src_codeintel_symbols_gitserver_errors_total{job=~"^searcher.*"}[5m]))) * 100
searcher: codeintel_symbols_gitserver_total
Gitserver client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101110 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_symbols_gitserver_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_gitserver_99th_percentile_duration
99th percentile successful gitserver client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_symbols_gitserver_duration_seconds_bucket{job=~"^searcher.*"}[5m])))
searcher: codeintel_symbols_gitserver_errors_total
Gitserver client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101112 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_symbols_gitserver_errors_total{job=~"^searcher.*"}[5m]))
searcher: codeintel_symbols_gitserver_error_rate
Gitserver client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101113 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_symbols_gitserver_errors_total{job=~"^searcher.*"}[5m])) / (sum by (op)(increase(src_codeintel_symbols_gitserver_total{job=~"^searcher.*"}[5m])) + sum by (op)(increase(src_codeintel_symbols_gitserver_errors_total{job=~"^searcher.*"}[5m]))) * 100
Searcher: Rockskip
searcher: p95_rockskip_search_request_duration
95th percentile search request duration over 5m
The 95th percentile duration of search requests to Rockskip in seconds. Lower is better.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101200 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_rockskip_service_search_request_duration_seconds_bucket[5m])) by (le))
searcher: rockskip_in_flight_search_requests
Number of in-flight search requests
The number of search requests currently being processed by Rockskip. If there is not much traffic and the requests are served very fast relative to the polling window of Prometheus, it possible that that this number is 0 even if there are search requests being processed.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_rockskip_service_in_flight_search_requests)
searcher: rockskip_search_request_errors
Search request errors every 5m
The number of search requests that returned an error in the last 5 minutes. The errors tracked here are all application errors, grpc errors are not included. We generally want this to be 0.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_rockskip_service_search_request_errors[5m]))
searcher: p95_rockskip_index_job_duration
95th percentile index job duration over 5m
The 95th percentile duration of index jobs in seconds. The range of values is very large, because the metric measure quick delta updates as well as full index jobs. Lower is better.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101210 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_rockskip_service_index_job_duration_seconds_bucket[5m])) by (le))
searcher: rockskip_in_flight_index_jobs
Number of in-flight index jobs
The number of index jobs currently being processed by Rockskip. This includes delta updates as well as full index jobs.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101211 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_rockskip_service_in_flight_index_jobs)
searcher: rockskip_index_job_errors
Index job errors every 5m
The number of index jobs that returned an error in the last 5 minutes. If the errors are persistent, users will see alerts in the UI. The service logs will contain more detailed information about the kind of errors. We generally want this to be 0.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_rockskip_service_index_job_errors[5m]))
searcher: rockskip_number_of_repos_indexed
Number of repositories indexed by Rockskip
The number of repositories indexed by Rockskip. Apart from an initial transient phase in which many repos are being indexed, this number should be low and relatively stable and only increase by small increments. To verify if this number makes sense, compare ROCKSKIP_MIN_REPO_SIZE_MB with the repository sizes reported by gitserver_repos table.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101220 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_rockskip_service_repos_indexed)
searcher: p95_rockskip_index_queue_age
95th percentile index queue delay over 5m
The 95th percentile age of index jobs in seconds. A high delay might indicate a resource issue. Consider increasing indexing bandwidth by either increasing the number of queues or the number of symbol services.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101221 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_rockskip_service_index_queue_age_seconds_bucket[5m])) by (le))
searcher: rockskip_file_parsing_requests
File parsing requests every 5m
The number of search requests in the last 5 minutes that were handled by parsing a single file, as opposed to searching the Rockskip index. This is an optimization to speed up symbol sidebar queries.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101222 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_rockskip_service_file_parsing_requests[5m]))
Searcher: Site configuration client update latency
searcher: searcher_site_configuration_duration_since_last_successful_update_by_instance
Duration since last successful site configuration update (by instance)
The duration since the configuration client used by the "searcher" service last successfully updated its site configuration. Long durations could indicate issues updating the site configuration.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101300 on your Sourcegraph instance.
Technical details
Query:
SHELLsrc_conf_client_time_since_last_successful_update_seconds{job=~`.*searcher`,instance=~`${instance:regex}`}
searcher: searcher_site_configuration_duration_since_last_successful_update_by_instance
Maximum duration since last successful site configuration update (all "searcher" instances)
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101301 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(max_over_time(src_conf_client_time_since_last_successful_update_seconds{job=~`.*searcher`,instance=~`${instance:regex}`}[1m]))
Searcher: Periodic Goroutines
searcher: running_goroutines
Number of currently running periodic goroutines
The number of currently running periodic goroutines by name and job.
A value of 0 indicates the routine isnt running currently, it awaits its next schedule.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (src_periodic_goroutine_running{job=~".*searcher.*"})
searcher: goroutine_success_rate
Success rate for periodic goroutine executions
The rate of successful executions of each periodic goroutine. A low or zero value could indicate that a routine is stalled or encountering errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_total{job=~".*searcher.*"}[5m]))
searcher: goroutine_error_rate
Error rate for periodic goroutine executions
The rate of errors encountered by each periodic goroutine. A sustained high error rate may indicate a problem with the routine`s configuration or dependencies.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101410 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_errors_total{job=~".*searcher.*"}[5m]))
searcher: goroutine_error_percentage
Percentage of periodic goroutine executions that result in errors
The percentage of executions that result in errors for each periodic goroutine. A value above 5% indicates that a significant portion of routine executions are failing.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101411 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_errors_total{job=~".*searcher.*"}[5m])) / sum by (name, job_name) (rate(src_periodic_goroutine_total{job=~".*searcher.*"}[5m]) > 0) * 100
searcher: goroutine_handler_duration
95th percentile handler execution time
The 95th percentile execution time for each periodic goroutine handler. Longer durations might indicate increased load or processing time.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101420 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job_name, le) (rate(src_periodic_goroutine_duration_seconds_bucket{job=~".*searcher.*"}[5m])))
searcher: goroutine_loop_duration
95th percentile loop cycle time
The 95th percentile loop cycle time for each periodic goroutine (excluding sleep time). This represents how long a complete loop iteration takes before sleeping for the next interval.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101421 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job_name, le) (rate(src_periodic_goroutine_loop_duration_seconds_bucket{job=~".*searcher.*"}[5m])))
searcher: tenant_processing_duration
95th percentile tenant processing time
The 95th percentile processing time for individual tenants within periodic goroutines. Higher values indicate that tenant processing is taking longer and may affect overall performance.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101430 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job_name, le) (rate(src_periodic_goroutine_tenant_duration_seconds_bucket{job=~".*searcher.*"}[5m])))
searcher: tenant_processing_max
Maximum tenant processing time
The maximum processing time for individual tenants within periodic goroutines. Consistently high values might indicate problematic tenants or inefficient processing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101431 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name, job_name) (rate(src_periodic_goroutine_tenant_duration_seconds_sum{job=~".*searcher.*"}[5m]) / rate(src_periodic_goroutine_tenant_duration_seconds_count{job=~".*searcher.*"}[5m]))
searcher: tenant_count
Number of tenants processed per routine
The number of tenants processed by each periodic goroutine. Unexpected changes can indicate tenant configuration issues or scaling events.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101440 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name, job_name) (src_periodic_goroutine_tenant_count{job=~".*searcher.*"})
searcher: tenant_success_rate
Rate of successful tenant processing operations
The rate of successful tenant processing operations. A healthy routine should maintain a consistent processing rate.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101441 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_tenant_success_total{job=~".*searcher.*"}[5m]))
searcher: tenant_error_rate
Rate of tenant processing errors
The rate of tenant processing operations that result in errors. Consistent errors indicate problems with specific tenants.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101450 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job_name) (rate(src_periodic_goroutine_tenant_errors_total{job=~".*searcher.*"}[5m]))
searcher: tenant_error_percentage
Percentage of tenant operations resulting in errors
The percentage of tenant operations that result in errors. Values above 5% indicate significant tenant processing problems.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101451 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum by (name, job_name) (rate(src_periodic_goroutine_tenant_errors_total{job=~".*searcher.*"}[5m])) / (sum by (name, job_name) (rate(src_periodic_goroutine_tenant_success_total{job=~".*searcher.*"}[5m])) + sum by (name, job_name) (rate(src_periodic_goroutine_tenant_errors_total{job=~".*searcher.*"}[5m])))) * 100
Searcher: Database connections
searcher: max_open_conns
Maximum open
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_max_open{app_name="searcher"})
searcher: open_conns
Established
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101501 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_open{app_name="searcher"})
searcher: in_use
Used
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101510 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_in_use{app_name="searcher"})
searcher: idle
Idle
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101511 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_idle{app_name="searcher"})
searcher: mean_blocked_seconds_per_conn_request
Mean blocked seconds per conn request
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101520 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_blocked_seconds{app_name="searcher"}[5m])) / sum by (app_name, db_name) (increase(src_pgsql_conns_waited_for{app_name="searcher"}[5m]))
searcher: closed_max_idle
Closed by SetMaxIdleConns
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101530 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle{app_name="searcher"}[5m]))
searcher: closed_max_lifetime
Closed by SetConnMaxLifetime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101531 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_lifetime{app_name="searcher"}[5m]))
searcher: closed_max_idle_time
Closed by SetConnMaxIdleTime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101532 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle_time{app_name="searcher"}[5m]))
Searcher: Searcher (CPU, Memory)
searcher: cpu_usage_percentage
CPU usage
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101600 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^searcher.*"}
searcher: memory_usage_percentage
Memory usage percentage (total)
An estimate for the active memory in use, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101601 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^searcher.*"}
searcher: memory_working_set_bytes
Memory usage bytes (total)
An estimate for the active memory in use in bytes, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101602 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_memory_working_set_bytes{name=~"^searcher.*"})
searcher: memory_rss
Memory (RSS)
The total anonymous memory in use by the application, which includes Go stack and heap. This memory is is non-reclaimable, and high usage may trigger OOM kills. Note: the metric is named RSS because to match the cadvisor name, but anonymous is more accurate."
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101610 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_rss{name=~"^searcher.*"} / container_spec_memory_limit_bytes{name=~"^searcher.*"}) by (name) * 100.0
searcher: memory_total_active_file
Memory usage (active file)
This metric shows the total active file-backed memory currently in use by the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101611 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_total_active_file_bytes{name=~"^searcher.*"} / container_spec_memory_limit_bytes{name=~"^searcher.*"}) by (name) * 100.0
searcher: memory_kernel_usage
Memory usage (kernel)
The kernel usage metric shows the amount of memory used by the kernel on behalf of the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101612 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_kernel_usage{name=~"^searcher.*"} / container_spec_memory_limit_bytes{name=~"^searcher.*"}) by (name) * 100.0
Searcher: Container monitoring (not available on server)
searcher: container_missing
Container missing
This value is the number of times a container has not been seen for more than one minute. If you observe this value change independent of deployment events (such as an upgrade), it could indicate pods are being OOM killed or terminated for some other reasons.
- Kubernetes:
- Determine if the pod was OOM killed using
kubectl describe pod searcher(look forOOMKilled: true) and, if so, consider increasing the memory limit in the relevantDeployment.yaml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingkubectl logs -p searcher.
- Determine if the pod was OOM killed using
- Docker Compose:
- Determine if the pod was OOM killed using
docker inspect -f '\{\{json .State\}\}' searcher(look for"OOMKilled":true) and, if so, consider increasing the memory limit of the searcher container indocker-compose.yml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingdocker logs searcher(note this will include logs from the previous and currently running container).
- Determine if the pod was OOM killed using
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101700 on your Sourcegraph instance.
Technical details
Query:
SHELLcount by(name) ((time() - container_last_seen{name=~"^searcher.*"}) > 60)
searcher: container_cpu_usage
Container cpu usage total (1m average) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101701 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^searcher.*"}
searcher: container_memory_usage
Container memory usage by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101702 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^searcher.*"}
searcher: fs_io_operations
Filesystem reads and writes rate by instance over 1h
This value indicates the number of filesystem read and write operations by containers of this service. When extremely high, this can indicate a resource usage problem, or can cause problems with the service itself, especially if high values or spikes correlate with {{CONTAINER_NAME}} issues.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101703 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(name) (rate(container_fs_reads_total{name=~"^searcher.*"}[1h]) + rate(container_fs_writes_total{name=~"^searcher.*"}[1h]))
Searcher: Provisioning indicators (not available on server)
searcher: provisioning_container_cpu_usage_long_term
Container cpu usage total (90th percentile over 1d) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101800 on your Sourcegraph instance.
Technical details
Query:
SHELLquantile_over_time(0.9, cadvisor_container_cpu_usage_percentage_total{name=~"^searcher.*"}[1d])
searcher: provisioning_container_memory_usage_long_term
Container memory usage (1d maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101801 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^searcher.*"}[1d])
searcher: provisioning_container_cpu_usage_short_term
Container cpu usage total (5m maximum) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101810 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^searcher.*"}[5m])
searcher: provisioning_container_memory_usage_short_term
Container memory usage (5m maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101811 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^searcher.*"}[5m])
searcher: container_oomkill_events_total
Container OOMKILL events total by instance
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101812 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^searcher.*"})
Searcher: Golang runtime monitoring
searcher: go_goroutines
Maximum active goroutines
A high value here indicates a possible goroutine leak.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101900 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_goroutines{job=~".*searcher"})
searcher: go_gc_duration_seconds
Maximum go garbage collection duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=101901 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_gc_duration_seconds{job=~".*searcher"})
Searcher: Kubernetes monitoring (only available on Kubernetes)
searcher: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/searcher/searcher?viewPanel=102000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*searcher"}) / count by (app) (up{app=~".*searcher"}) * 100
Syntect Server
Handles syntax highlighting for code files.
To see this dashboard, visit /-/debug/grafana/d/syntect-server/syntect-server on your Sourcegraph instance.
syntect-server: syntax_highlighting_errors
Syntax highlighting errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_syntax_highlighting_requests{status="error"}[5m])) / sum(increase(src_syntax_highlighting_requests[5m])) * 100
syntect-server: syntax_highlighting_timeouts
Syntax highlighting timeouts every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_syntax_highlighting_requests{status="timeout"}[5m])) / sum(increase(src_syntax_highlighting_requests[5m])) * 100
syntect-server: syntax_highlighting_panics
Syntax highlighting panics every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_syntax_highlighting_requests{status="panic"}[5m]))
syntect-server: syntax_highlighting_worker_deaths
Syntax highlighter worker deaths every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_syntax_highlighting_requests{status="hss_worker_timeout"}[5m]))
Syntect Server: Syntect-server (CPU, Memory)
syntect-server: cpu_usage_percentage
CPU usage
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^syntect-server.*"}
syntect-server: memory_usage_percentage
Memory usage percentage (total)
An estimate for the active memory in use, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^syntect-server.*"}
syntect-server: memory_working_set_bytes
Memory usage bytes (total)
An estimate for the active memory in use in bytes, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_memory_working_set_bytes{name=~"^syntect-server.*"})
syntect-server: memory_rss
Memory (RSS)
The total anonymous memory in use by the application, which includes Go stack and heap. This memory is is non-reclaimable, and high usage may trigger OOM kills. Note: the metric is named RSS because to match the cadvisor name, but anonymous is more accurate."
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_rss{name=~"^syntect-server.*"} / container_spec_memory_limit_bytes{name=~"^syntect-server.*"}) by (name) * 100.0
syntect-server: memory_total_active_file
Memory usage (active file)
This metric shows the total active file-backed memory currently in use by the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_total_active_file_bytes{name=~"^syntect-server.*"} / container_spec_memory_limit_bytes{name=~"^syntect-server.*"}) by (name) * 100.0
syntect-server: memory_kernel_usage
Memory usage (kernel)
The kernel usage metric shows the amount of memory used by the kernel on behalf of the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100112 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_kernel_usage{name=~"^syntect-server.*"} / container_spec_memory_limit_bytes{name=~"^syntect-server.*"}) by (name) * 100.0
Syntect Server: Container monitoring (not available on server)
syntect-server: container_missing
Container missing
This value is the number of times a container has not been seen for more than one minute. If you observe this value change independent of deployment events (such as an upgrade), it could indicate pods are being OOM killed or terminated for some other reasons.
- Kubernetes:
- Determine if the pod was OOM killed using
kubectl describe pod syntect-server(look forOOMKilled: true) and, if so, consider increasing the memory limit in the relevantDeployment.yaml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingkubectl logs -p syntect-server.
- Determine if the pod was OOM killed using
- Docker Compose:
- Determine if the pod was OOM killed using
docker inspect -f '\{\{json .State\}\}' syntect-server(look for"OOMKilled":true) and, if so, consider increasing the memory limit of the syntect-server container indocker-compose.yml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingdocker logs syntect-server(note this will include logs from the previous and currently running container).
- Determine if the pod was OOM killed using
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLcount by(name) ((time() - container_last_seen{name=~"^syntect-server.*"}) > 60)
syntect-server: container_cpu_usage
Container cpu usage total (1m average) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^syntect-server.*"}
syntect-server: container_memory_usage
Container memory usage by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^syntect-server.*"}
syntect-server: fs_io_operations
Filesystem reads and writes rate by instance over 1h
This value indicates the number of filesystem read and write operations by containers of this service. When extremely high, this can indicate a resource usage problem, or can cause problems with the service itself, especially if high values or spikes correlate with {{CONTAINER_NAME}} issues.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(name) (rate(container_fs_reads_total{name=~"^syntect-server.*"}[1h]) + rate(container_fs_writes_total{name=~"^syntect-server.*"}[1h]))
Syntect Server: Provisioning indicators (not available on server)
syntect-server: provisioning_container_cpu_usage_long_term
Container cpu usage total (90th percentile over 1d) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLquantile_over_time(0.9, cadvisor_container_cpu_usage_percentage_total{name=~"^syntect-server.*"}[1d])
syntect-server: provisioning_container_memory_usage_long_term
Container memory usage (1d maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^syntect-server.*"}[1d])
syntect-server: provisioning_container_cpu_usage_short_term
Container cpu usage total (5m maximum) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100310 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^syntect-server.*"}[5m])
syntect-server: provisioning_container_memory_usage_short_term
Container memory usage (5m maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100311 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^syntect-server.*"}[5m])
syntect-server: container_oomkill_events_total
Container OOMKILL events total by instance
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100312 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^syntect-server.*"})
Syntect Server: Kubernetes monitoring (only available on Kubernetes)
syntect-server: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/syntect-server/syntect-server?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*syntect-server"}) / count by (app) (up{app=~".*syntect-server"}) * 100
Zoekt
Indexes repositories, populates the search index, and responds to indexed search queries.
To see this dashboard, visit /-/debug/grafana/d/zoekt/zoekt on your Sourcegraph instance.
zoekt: total_repos_aggregate
Total number of repos (aggregate)
Sudden changes can be caused by indexing configuration changes.
Additionally, a discrepancy between "index_num_assigned" and "index_queue_cap" could indicate a bug.
Legend:
- index_num_assigned: # of repos assigned to Zoekt
- index_num_indexed: # of repos Zoekt has indexed
- index_queue_cap: # of repos Zoekt is aware of, including those that it has finished indexing
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (__name__) ({__name__=~"index_num_assigned|index_num_indexed|index_queue_cap"})
zoekt: total_repos_per_instance
Total number of repos (per instance)
Sudden changes can be caused by indexing configuration changes.
Additionally, a discrepancy between "index_num_assigned" and "index_queue_cap" could indicate a bug.
Legend:
- index_num_assigned: # of repos assigned to Zoekt
- index_num_indexed: # of repos Zoekt has indexed
- index_queue_cap: # of repos Zoekt is aware of, including those that it has finished processing
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (__name__, instance) ({__name__=~"index_num_assigned|index_num_indexed|index_queue_cap",instance=~"${instance:regex}"})
zoekt: repos_stopped_tracking_total_aggregate
The number of repositories we stopped tracking over 5m (aggregate)
Repositories we stop tracking are soft-deleted during the next cleanup job.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(index_num_stopped_tracking_total[5m]))
zoekt: repos_stopped_tracking_total_per_instance
The number of repositories we stopped tracking over 5m (per instance)
Repositories we stop tracking are soft-deleted during the next cleanup job.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance) (increase(index_num_stopped_tracking_total{instance=~`${instance:regex}`}[5m]))
zoekt: average_resolve_revision_duration
Average resolve revision duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100020 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(resolve_revision_seconds_sum[5m])) / sum(rate(resolve_revision_seconds_count[5m]))
zoekt: get_index_options_error_increase
The number of repositories we failed to get indexing options over 5m
When considering indexing a repository we ask for the index configuration from frontend per repository. The most likely reason this would fail is failing to resolve branch names to git SHAs.
This value can spike up during deployments/etc. Only if you encounter sustained periods of errors is there an underlying issue. When sustained this indicates repositories will not get updated indexes.
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100021 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(get_index_options_error_total[5m]))
Zoekt: Zoekt-indexserver (CPU, Memory)
zoekt: cpu_usage_percentage
CPU usage
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^zoekt-indexserver.*"}
zoekt: memory_usage_percentage
Memory usage percentage (total)
An estimate for the active memory in use, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^zoekt-indexserver.*"}
zoekt: memory_working_set_bytes
Memory usage bytes (total)
An estimate for the active memory in use in bytes, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_memory_working_set_bytes{name=~"^zoekt-indexserver.*"})
zoekt: memory_rss
Memory (RSS)
The total anonymous memory in use by the application, which includes Go stack and heap. This memory is is non-reclaimable, and high usage may trigger OOM kills. Note: the metric is named RSS because to match the cadvisor name, but anonymous is more accurate."
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_rss{name=~"^zoekt-indexserver.*"} / container_spec_memory_limit_bytes{name=~"^zoekt-indexserver.*"}) by (name) * 100.0
zoekt: memory_total_active_file
Memory usage (active file)
This metric shows the total active file-backed memory currently in use by the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_total_active_file_bytes{name=~"^zoekt-indexserver.*"} / container_spec_memory_limit_bytes{name=~"^zoekt-indexserver.*"}) by (name) * 100.0
zoekt: memory_kernel_usage
Memory usage (kernel)
The kernel usage metric shows the amount of memory used by the kernel on behalf of the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100112 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_kernel_usage{name=~"^zoekt-indexserver.*"} / container_spec_memory_limit_bytes{name=~"^zoekt-indexserver.*"}) by (name) * 100.0
Zoekt: Zoekt-webserver (CPU, Memory)
zoekt: cpu_usage_percentage
CPU usage
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^zoekt-webserver.*"}
zoekt: memory_usage_percentage
Memory usage percentage (total)
An estimate for the active memory in use, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^zoekt-webserver.*"}
zoekt: memory_working_set_bytes
Memory usage bytes (total)
An estimate for the active memory in use in bytes, which includes anonymous memory, file memory, and kernel memory. Some of this memory is reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_memory_working_set_bytes{name=~"^zoekt-webserver.*"})
zoekt: memory_rss
Memory (RSS)
The total anonymous memory in use by the application, which includes Go stack and heap. This memory is is non-reclaimable, and high usage may trigger OOM kills. Note: the metric is named RSS because to match the cadvisor name, but anonymous is more accurate."
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_rss{name=~"^zoekt-webserver.*"} / container_spec_memory_limit_bytes{name=~"^zoekt-webserver.*"}) by (name) * 100.0
zoekt: memory_total_active_file
Memory usage (active file)
This metric shows the total active file-backed memory currently in use by the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_total_active_file_bytes{name=~"^zoekt-webserver.*"} / container_spec_memory_limit_bytes{name=~"^zoekt-webserver.*"}) by (name) * 100.0
zoekt: memory_kernel_usage
Memory usage (kernel)
The kernel usage metric shows the amount of memory used by the kernel on behalf of the application. Some of it may be reclaimable, so high usage does not necessarily indicate memory pressure.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(container_memory_kernel_usage{name=~"^zoekt-webserver.*"} / container_spec_memory_limit_bytes{name=~"^zoekt-webserver.*"}) by (name) * 100.0
Zoekt: Memory mapping metrics
zoekt: memory_map_areas_percentage_used
Process memory map areas percentage used (per instance)
Processes have a limited about of memory map areas that they can use. In Zoekt, memory map areas are mainly used for loading shards into memory for queries (via mmap). However, memory map areas are also used for loading shared libraries, etc.
See https://en.wikipedia.org/wiki/Memory-mapped_file and the related articles for more information about memory maps.
Once the memory map limit is reached, the Linux kernel will prevent the process from creating any additional memory map areas. This could cause the process to crash.
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELL(proc_metrics_memory_map_current_count{instance=~`${instance:regex}`} / proc_metrics_memory_map_max_limit{instance=~`${instance:regex}`}) * 100
zoekt: memory_major_page_faults
Webserver page faults
The number of major page faults in a 5 minute window for Zoekt webservers. If this number increases significantly, it indicates that more searches need to load data from disk. There may not be enough memory to efficiently support amount of repo data being searched.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLrate(container_memory_failures_total{failure_type="pgmajfault", name=~"^zoekt-webserver.*"}[5m])
Zoekt: Search requests
zoekt: indexed_search_request_duration_p99_aggregate
99th percentile indexed search duration over 1m (aggregate)
This dashboard shows the 99th percentile of search request durations over the last minute (aggregated across all instances).
Large duration spikes can be an indicator of saturation and / or a performance regression.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, name)(rate(zoekt_search_duration_seconds_bucket[1m])))
zoekt: indexed_search_request_duration_p90_aggregate
90th percentile indexed search duration over 1m (aggregate)
This dashboard shows the 90th percentile of search request durations over the last minute (aggregated across all instances).
Large duration spikes can be an indicator of saturation and / or a performance regression.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name)(rate(zoekt_search_duration_seconds_bucket[1m])))
zoekt: indexed_search_request_duration_p75_aggregate
75th percentile indexed search duration over 1m (aggregate)
This dashboard shows the 75th percentile of search request durations over the last minute (aggregated across all instances).
Large duration spikes can be an indicator of saturation and / or a performance regression.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100402 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name)(rate(zoekt_search_duration_seconds_bucket[1m])))
zoekt: indexed_search_request_duration_p99_by_instance
99th percentile indexed search duration over 1m (per instance)
This dashboard shows the 99th percentile of search request durations over the last minute (broken out per instance).
Large duration spikes can be an indicator of saturation and / or a performance regression.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100410 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, instance)(rate(zoekt_search_duration_seconds_bucket{instance=~`${instance:regex}`}[1m])))
zoekt: indexed_search_request_duration_p90_by_instance
90th percentile indexed search duration over 1m (per instance)
This dashboard shows the 90th percentile of search request durations over the last minute (broken out per instance).
Large duration spikes can be an indicator of saturation and / or a performance regression.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100411 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, instance)(rate(zoekt_search_duration_seconds_bucket{instance=~`${instance:regex}`}[1m])))
zoekt: indexed_search_request_duration_p75_by_instance
75th percentile indexed search duration over 1m (per instance)
This dashboard shows the 75th percentile of search request durations over the last minute (broken out per instance).
Large duration spikes can be an indicator of saturation and / or a performance regression.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100412 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, instance)(rate(zoekt_search_duration_seconds_bucket{instance=~`${instance:regex}`}[1m])))
zoekt: indexed_search_num_concurrent_requests_aggregate
Amount of in-flight indexed search requests (aggregate)
This dashboard shows the current number of indexed search requests that are in-flight, aggregated across all instances.
In-flight search requests include both running and queued requests.
The number of in-flight requests can serve as a proxy for the general load that webserver instances are under.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100420 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name) (zoekt_search_running)
zoekt: indexed_search_num_concurrent_requests_by_instance
Amount of in-flight indexed search requests (per instance)
This dashboard shows the current number of indexed search requests that are-flight, broken out per instance.
In-flight search requests include both running and queued requests.
The number of in-flight requests can serve as a proxy for the general load that webserver instances are under.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100421 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance, name) (zoekt_search_running{instance=~`${instance:regex}`})
zoekt: indexed_search_concurrent_request_growth_rate_1m_aggregate
Rate of growth of in-flight indexed search requests over 1m (aggregate)
This dashboard shows the rate of growth of in-flight requests, aggregated across all instances.
In-flight search requests include both running and queued requests.
This metric gives a notion of how quickly the indexed-search backend is working through its request load (taking into account the request arrival rate and processing time). A sustained high rate of growth can indicate that the indexed-search backend is saturated.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100430 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name) (deriv(zoekt_search_running[1m]))
zoekt: indexed_search_concurrent_request_growth_rate_1m_per_instance
Rate of growth of in-flight indexed search requests over 1m (per instance)
This dashboard shows the rate of growth of in-flight requests, broken out per instance.
In-flight search requests include both running and queued requests.
This metric gives a notion of how quickly the indexed-search backend is working through its request load (taking into account the request arrival rate and processing time). A sustained high rate of growth can indicate that the indexed-search backend is saturated.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100431 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance) (deriv(zoekt_search_running[1m]))
zoekt: indexed_search_request_errors
Indexed search request errors every 5m by code
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100440 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (code)(increase(src_zoekt_request_duration_seconds_count{code!~"2.."}[5m])) / ignoring(code) group_left sum(increase(src_zoekt_request_duration_seconds_count[5m])) * 100
zoekt: zoekt_shards_sched
Current number of zoekt scheduler processes in a state
Each ongoing search request starts its life as an interactive query. If it takes too long it becomes a batch query. Between state transitions it can be queued.
If you have a high number of batch queries it is a sign there is a large load of slow queries. Alternatively your systems are underprovisioned and normal search queries are taking too long.
For a full explanation of the states see https://github.com/sourcegraph/zoekt/blob/930cd1c28917e64c87f0ce354a0fd040877cbba1/shards/sched.go#L311-L340
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100450 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (type, state) (zoekt_shards_sched)
zoekt: zoekt_shards_sched_total
Rate of zoekt scheduler process state transitions in the last 5m
Each ongoing search request starts its life as an interactive query. If it takes too long it becomes a batch query. Between state transitions it can be queued.
If you have a high number of batch queries it is a sign there is a large load of slow queries. Alternatively your systems are underprovisioned and normal search queries are taking too long.
For a full explanation of the states see https://github.com/sourcegraph/zoekt/blob/930cd1c28917e64c87f0ce354a0fd040877cbba1/shards/sched.go#L311-L340
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100451 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (type, state) (rate(zoekt_shards_sched[5m]))
Zoekt: Git fetch durations
zoekt: 90th_percentile_successful_git_fetch_durations_5m
90th percentile successful git fetch durations over 5m
Long git fetch times can be a leading indicator of saturation.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name)(rate(index_fetch_seconds_bucket{success="true"}[5m])))
zoekt: 90th_percentile_failed_git_fetch_durations_5m
90th percentile failed git fetch durations over 5m
Long git fetch times can be a leading indicator of saturation.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100501 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name)(rate(index_fetch_seconds_bucket{success="false"}[5m])))
Zoekt: Indexing results
zoekt: repo_index_state_aggregate
Index results state count over 5m (aggregate)
This dashboard shows the outcomes of recently completed indexing jobs across all index-server instances.
A persistent failing state indicates some repositories cannot be indexed, perhaps due to size and timeouts.
Legend:
- fail -> the indexing jobs failed
- success -> the indexing job succeeded and the index was updated
- success_meta -> the indexing job succeeded, but only metadata was updated
- noop -> the indexing job succeed, but we didn`t need to update anything
- empty -> the indexing job succeeded, but the index was empty (i.e. the repository is empty)
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (state) (increase(index_repo_seconds_count[5m]))
zoekt: repo_index_state_per_instance
Index results state count over 5m (per instance)
This dashboard shows the outcomes of recently completed indexing jobs, split out across each index-server instance.
(You can use the "instance" filter at the top of the page to select a particular instance.)
A persistent failing state indicates some repositories cannot be indexed, perhaps due to size and timeouts.
Legend:
- fail -> the indexing jobs failed
- success -> the indexing job succeeded and the index was updated
- success_meta -> the indexing job succeeded, but only metadata was updated
- noop -> the indexing job succeed, but we didn`t need to update anything
- empty -> the indexing job succeeded, but the index was empty (i.e. the repository is empty)
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100601 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance, state) (increase(index_repo_seconds_count{instance=~`${instance:regex}`}[5m]))
zoekt: repo_index_success_speed_heatmap
Successful indexing durations
Latency increases can indicate bottlenecks in the indexserver.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100610 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le, state) (increase(index_repo_seconds_bucket{state="success"}[$__rate_interval]))
zoekt: repo_index_fail_speed_heatmap
Failed indexing durations
Failures happening after a long time indicates timeouts.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100611 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le, state) (increase(index_repo_seconds_bucket{state="fail"}[$__rate_interval]))
zoekt: repo_index_success_speed_p99
99th percentile successful indexing durations over 5m (aggregate)
This dashboard shows the p99 duration of successful indexing jobs aggregated across all Zoekt instances.
Latency increases can indicate bottlenecks in the indexserver.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100620 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, name)(rate(index_repo_seconds_bucket{state="success"}[5m])))
zoekt: repo_index_success_speed_p90
90th percentile successful indexing durations over 5m (aggregate)
This dashboard shows the p90 duration of successful indexing jobs aggregated across all Zoekt instances.
Latency increases can indicate bottlenecks in the indexserver.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100621 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name)(rate(index_repo_seconds_bucket{state="success"}[5m])))
zoekt: repo_index_success_speed_p75
75th percentile successful indexing durations over 5m (aggregate)
This dashboard shows the p75 duration of successful indexing jobs aggregated across all Zoekt instances.
Latency increases can indicate bottlenecks in the indexserver.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100622 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name)(rate(index_repo_seconds_bucket{state="success"}[5m])))
zoekt: repo_index_success_speed_p99_per_instance
99th percentile successful indexing durations over 5m (per instance)
This dashboard shows the p99 duration of successful indexing jobs broken out per Zoekt instance.
Latency increases can indicate bottlenecks in the indexserver.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100630 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, instance)(rate(index_repo_seconds_bucket{state="success",instance=~`${instance:regex}`}[5m])))
zoekt: repo_index_success_speed_p90_per_instance
90th percentile successful indexing durations over 5m (per instance)
This dashboard shows the p90 duration of successful indexing jobs broken out per Zoekt instance.
Latency increases can indicate bottlenecks in the indexserver.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100631 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, instance)(rate(index_repo_seconds_bucket{state="success",instance=~`${instance:regex}`}[5m])))
zoekt: repo_index_success_speed_p75_per_instance
75th percentile successful indexing durations over 5m (per instance)
This dashboard shows the p75 duration of successful indexing jobs broken out per Zoekt instance.
Latency increases can indicate bottlenecks in the indexserver.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100632 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, instance)(rate(index_repo_seconds_bucket{state="success",instance=~`${instance:regex}`}[5m])))
zoekt: repo_index_failed_speed_p99
99th percentile failed indexing durations over 5m (aggregate)
This dashboard shows the p99 duration of failed indexing jobs aggregated across all Zoekt instances.
Failures happening after a long time indicates timeouts.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100640 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, name)(rate(index_repo_seconds_bucket{state="fail"}[5m])))
zoekt: repo_index_failed_speed_p90
90th percentile failed indexing durations over 5m (aggregate)
This dashboard shows the p90 duration of failed indexing jobs aggregated across all Zoekt instances.
Failures happening after a long time indicates timeouts.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100641 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name)(rate(index_repo_seconds_bucket{state="fail"}[5m])))
zoekt: repo_index_failed_speed_p75
75th percentile failed indexing durations over 5m (aggregate)
This dashboard shows the p75 duration of failed indexing jobs aggregated across all Zoekt instances.
Failures happening after a long time indicates timeouts.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100642 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name)(rate(index_repo_seconds_bucket{state="fail"}[5m])))
zoekt: repo_index_failed_speed_p99_per_instance
99th percentile failed indexing durations over 5m (per instance)
This dashboard shows the p99 duration of failed indexing jobs broken out per Zoekt instance.
Failures happening after a long time indicates timeouts.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100650 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, instance)(rate(index_repo_seconds_bucket{state="fail",instance=~`${instance:regex}`}[5m])))
zoekt: repo_index_failed_speed_p90_per_instance
90th percentile failed indexing durations over 5m (per instance)
This dashboard shows the p90 duration of failed indexing jobs broken out per Zoekt instance.
Failures happening after a long time indicates timeouts.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100651 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, instance)(rate(index_repo_seconds_bucket{state="fail",instance=~`${instance:regex}`}[5m])))
zoekt: repo_index_failed_speed_p75_per_instance
75th percentile failed indexing durations over 5m (per instance)
This dashboard shows the p75 duration of failed indexing jobs broken out per Zoekt instance.
Failures happening after a long time indicates timeouts.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100652 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, instance)(rate(index_repo_seconds_bucket{state="fail",instance=~`${instance:regex}`}[5m])))
Zoekt: Indexing queue statistics
zoekt: indexed_num_scheduled_jobs_aggregate
# scheduled index jobs (aggregate)
A queue that is constantly growing could be a leading indicator of a bottleneck or under-provisioning
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100700 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(index_queue_len)
zoekt: indexed_num_scheduled_jobs_per_instance
# scheduled index jobs (per instance)
A queue that is constantly growing could be a leading indicator of a bottleneck or under-provisioning
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100701 on your Sourcegraph instance.
Technical details
Query:
SHELLindex_queue_len{instance=~`${instance:regex}`}
zoekt: indexed_indexing_delay_heatmap
Repo indexing delay heatmap
The indexing delay represents the amount of time between when Zoekt received a repo indexing job, to when the repo was indexed. It includes the time the repo spent in the indexing queue, as well as the time it took to actually index the repo. This metric only includes successfully indexed repos.
Large indexing delays can be an indicator of:
- resource saturation
- each Zoekt replica has too many jobs for it to be able to process all of them promptly. In this scenario, consider adding additional Zoekt replicas to distribute the work better .
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100710 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le) (increase(index_indexing_delay_seconds_bucket{state=~"success|success_meta"}[$__rate_interval]))
zoekt: indexed_indexing_delay_p90_aggregate
90th percentile indexing delay over 5m (aggregate)
This dashboard shows the p90 indexing delay aggregated across all Zoekt instances.
The indexing delay represents the amount of time between when Zoekt received a repo indexing job, to when the repo was indexed. It includes the time the repo spent in the indexing queue, as well as the time it took to actually index the repo. This metric only includes successfully indexed repos.
Large indexing delays can be an indicator of:
- resource saturation
- each Zoekt replica has too many jobs for it to be able to process all of them promptly. In this scenario, consider adding additional Zoekt replicas to distribute the work better.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100720 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name)(rate(index_indexing_delay_seconds_bucket{state=~"success|success_meta"}[5m])))
zoekt: indexed_indexing_delay_p50_aggregate
50th percentile indexing delay over 5m (aggregate)
This dashboard shows the p50 indexing delay aggregated across all Zoekt instances.
The indexing delay represents the amount of time between when Zoekt received a repo indexing job, to when the repo was indexed. It includes the time the repo spent in the indexing queue, as well as the time it took to actually index the repo. This metric only includes successfully indexed repos.
Large indexing delays can be an indicator of:
- resource saturation
- each Zoekt replica has too many jobs for it to be able to process all of them promptly. In this scenario, consider adding additional Zoekt replicas to distribute the work better.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100721 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum by (le, name)(rate(index_indexing_delay_seconds_bucket{state=~"success|success_meta"}[5m])))
zoekt: indexed_indexing_delay_p90_per_instance
90th percentile indexing delay over 5m (per instance)
This dashboard shows the p90 indexing delay, broken out per Zoekt instance.
The indexing delay represents the amount of time between when Zoekt received a repo indexing job, to when the repo was indexed. It includes the time the repo spent in the indexing queue, as well as the time it took to actually index the repo.
Large indexing delays can be an indicator of:
- resource saturation
- each Zoekt replica has too many jobs for it to be able to process all of them promptly. In this scenario, consider adding additional Zoekt replicas to distribute the work better.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100730 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, instance)(rate(index_indexing_delay_seconds{instance=~`${instance:regex}`}[5m])))
zoekt: indexed_indexing_delay_p50_per_instance
50th percentile indexing delay over 5m (per instance)
This dashboard shows the p50 indexing delay, broken out per Zoekt instance.
The indexing delay represents the amount of time between when Zoekt received a repo indexing job, to when the repo was indexed. It includes the time the repo spent in the indexing queue, as well as the time it took to actually index the repo.
Large indexing delays can be an indicator of:
- resource saturation
- each Zoekt replica has too many jobs for it to be able to process all of them promptly. In this scenario, consider adding additional Zoekt replicas to distribute the work better.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100731 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum by (le, instance)(rate(index_indexing_delay_seconds{instance=~`${instance:regex}`}[5m])))
Zoekt: Compound shards
zoekt: compound_shards_aggregate
# of compound shards (aggregate)
The total number of compound shards aggregated over all instances.
This number should be consistent if the number of indexed repositories doesn`t change.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100800 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(index_number_compound_shards) by (app)
zoekt: compound_shards_per_instance
# of compound shards (per instance)
The total number of compound shards per instance.
This number should be consistent if the number of indexed repositories doesn`t change.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100801 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(index_number_compound_shards{instance=~`${instance:regex}`}) by (instance)
zoekt: average_shard_merging_duration_success
Average successful shard merging duration over 1 hour
Average duration of a successful merge over the last hour.
The duration depends on the target compound shard size. The larger the compound shard the longer a merge will take. Since the target compound shard size is set on start of zoekt-indexserver, the average duration should be consistent.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100810 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(index_shard_merging_duration_seconds_sum{error="false"}[1h])) / sum(rate(index_shard_merging_duration_seconds_count{error="false"}[1h]))
zoekt: average_shard_merging_duration_error
Average failed shard merging duration over 1 hour
Average duration of a failed merge over the last hour.
This curve should be flat. Any deviation should be investigated.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100811 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(index_shard_merging_duration_seconds_sum{error="true"}[1h])) / sum(rate(index_shard_merging_duration_seconds_count{error="true"}[1h]))
zoekt: shard_merging_errors_aggregate
Number of errors during shard merging (aggregate)
Number of errors during shard merging aggregated over all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100820 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(index_shard_merging_duration_seconds_count{error="true"}) by (app)
zoekt: shard_merging_errors_per_instance
Number of errors during shard merging (per instance)
Number of errors during shard merging per instance.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100821 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(index_shard_merging_duration_seconds_count{instance=~`${instance:regex}`, error="true"}) by (instance)
zoekt: shard_merging_merge_running_per_instance
If shard merging is running (per instance)
Set to 1 if shard merging is running.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100830 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (instance) (index_shard_merging_running{instance=~`${instance:regex}`})
zoekt: shard_merging_vacuum_running_per_instance
If vacuum is running (per instance)
Set to 1 if vacuum is running.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100831 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (instance) (index_vacuum_running{instance=~`${instance:regex}`})
Zoekt: Network I/O pod metrics (only available on Kubernetes)
zoekt: network_sent_bytes_aggregate
Transmission rate over 5m (aggregate)
The rate of bytes sent over the network across all pods
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100900 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(container_network_transmit_bytes_total{container_label_io_kubernetes_pod_name=~`.*indexed-search.*`}[5m]))
zoekt: network_received_packets_per_instance
Transmission rate over 5m (per instance)
The amount of bytes sent over the network by individual pods
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100901 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_network_transmit_bytes_total{container_label_io_kubernetes_pod_name=~`${instance:regex}`}[5m]))
zoekt: network_received_bytes_aggregate
Receive rate over 5m (aggregate)
The amount of bytes received from the network across pods
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100910 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(container_network_receive_bytes_total{container_label_io_kubernetes_pod_name=~`.*indexed-search.*`}[5m]))
zoekt: network_received_bytes_per_instance
Receive rate over 5m (per instance)
The amount of bytes received from the network by individual pods
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100911 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_network_receive_bytes_total{container_label_io_kubernetes_pod_name=~`${instance:regex}`}[5m]))
zoekt: network_transmitted_packets_dropped_by_instance
Transmit packet drop rate over 5m (by instance)
An increase in dropped packets could be a leading indicator of network saturation.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100920 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_network_transmit_packets_dropped_total{container_label_io_kubernetes_pod_name=~`${instance:regex}`}[5m]))
zoekt: network_transmitted_packets_errors_per_instance
Errors encountered while transmitting over 5m (per instance)
An increase in transmission errors could indicate a networking issue
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100921 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_network_transmit_errors_total{container_label_io_kubernetes_pod_name=~`${instance:regex}`}[5m]))
zoekt: network_received_packets_dropped_by_instance
Receive packet drop rate over 5m (by instance)
An increase in dropped packets could be a leading indicator of network saturation.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100922 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_network_receive_packets_dropped_total{container_label_io_kubernetes_pod_name=~`${instance:regex}`}[5m]))
zoekt: network_transmitted_packets_errors_by_instance
Errors encountered while receiving over 5m (per instance)
An increase in errors while receiving could indicate a networking issue.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=100923 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_network_receive_errors_total{container_label_io_kubernetes_pod_name=~`${instance:regex}`}[5m]))
Zoekt: Zoekt Webserver GRPC server metrics
zoekt: zoekt_webserver_grpc_request_rate_all_methods
Request rate across all methods over 2m
The number of gRPC requests received per second across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m]))
zoekt: zoekt_webserver_grpc_request_rate_per_method
Request rate per-method over 2m
The number of gRPC requests received per second broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{grpc_method=~`${zoekt_webserver_method:regex}`,instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])) by (grpc_method)
zoekt: zoekt_webserver_error_percentage_all_methods
Error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101010 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_code!="OK",instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m]))) / (sum(rate(grpc_server_handled_total{instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m]))) ))
zoekt: zoekt_webserver_grpc_error_percentage_per_method
Error percentage per-method over 2m
The percentage of gRPC requests that fail per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101011 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_method=~`${zoekt_webserver_method:regex}`,grpc_code!="OK",instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])) by (grpc_method)) / (sum(rate(grpc_server_handled_total{grpc_method=~`${zoekt_webserver_method:regex}`,instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])) by (grpc_method)) ))
zoekt: zoekt_webserver_p99_response_time_per_method
99th percentile response time per method over 2m
The 99th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101020 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${zoekt_webserver_method:regex}`,instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])))
zoekt: zoekt_webserver_p90_response_time_per_method
90th percentile response time per method over 2m
The 90th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101021 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${zoekt_webserver_method:regex}`,instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])))
zoekt: zoekt_webserver_p75_response_time_per_method
75th percentile response time per method over 2m
The 75th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101022 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${zoekt_webserver_method:regex}`,instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])))
zoekt: zoekt_webserver_p99_9_response_size_per_method
99.9th percentile total response size per method over 2m
The 99.9th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101030 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${zoekt_webserver_method:regex}`,instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])))
zoekt: zoekt_webserver_p90_response_size_per_method
90th percentile total response size per method over 2m
The 90th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101031 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${zoekt_webserver_method:regex}`,instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])))
zoekt: zoekt_webserver_p75_response_size_per_method
75th percentile total response size per method over 2m
The 75th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101032 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${zoekt_webserver_method:regex}`,instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])))
zoekt: zoekt_webserver_p99_9_invididual_sent_message_size_per_method
99.9th percentile individual sent message size per method over 2m
The 99.9th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101040 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${zoekt_webserver_method:regex}`,instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])))
zoekt: zoekt_webserver_p90_invididual_sent_message_size_per_method
90th percentile individual sent message size per method over 2m
The 90th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101041 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${zoekt_webserver_method:regex}`,instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])))
zoekt: zoekt_webserver_p75_invididual_sent_message_size_per_method
75th percentile individual sent message size per method over 2m
The 75th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101042 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${zoekt_webserver_method:regex}`,instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])))
zoekt: zoekt_webserver_grpc_response_stream_message_count_per_method
Average streaming response message count per-method over 2m
The average number of response messages sent during a streaming RPC method, broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101050 on your Sourcegraph instance.
Technical details
Query:
SHELL((sum(rate(grpc_server_msg_sent_total{grpc_type="server_stream",instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])) by (grpc_method))/(sum(rate(grpc_server_started_total{grpc_type="server_stream",instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])) by (grpc_method)))
zoekt: zoekt_webserver_grpc_all_codes_per_method
Response codes rate per-method over 2m
The rate of all generated gRPC response codes per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101060 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_handled_total{grpc_method=~`${zoekt_webserver_method:regex}`,instance=~`${webserver_instance:regex}`,grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])) by (grpc_method, grpc_code)
Zoekt: Zoekt Webserver GRPC "internal error" metrics
zoekt: zoekt_webserver_grpc_clients_error_percentage_all_methods
Client baseline error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods (regardless of whether or not there was an internal error), aggregated across all "zoekt_webserver" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101100 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"zoekt.webserver.v1.WebserverService",grpc_code!="OK"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])))))))
zoekt: zoekt_webserver_grpc_clients_error_percentage_per_method
Client baseline error percentage per-method over 2m
The percentage of gRPC requests that fail per method (regardless of whether or not there was an internal error), aggregated across all "zoekt_webserver" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101101 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"zoekt.webserver.v1.WebserverService",grpc_method=~"${zoekt_webserver_method:regex}",grpc_code!="OK"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"zoekt.webserver.v1.WebserverService",grpc_method=~"${zoekt_webserver_method:regex}"}[2m])) by (grpc_method))))))
zoekt: zoekt_webserver_grpc_clients_all_codes_per_method
Client baseline response codes rate per-method over 2m
The rate of all generated gRPC response codes per method (regardless of whether or not there was an internal error), aggregated across all "zoekt_webserver" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101102 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"zoekt.webserver.v1.WebserverService",grpc_method=~"${zoekt_webserver_method:regex}"}[2m])) by (grpc_method, grpc_code))
zoekt: zoekt_webserver_grpc_clients_internal_error_percentage_all_methods
Client-observed gRPC internal error percentage across all methods over 2m
The percentage of gRPC requests that appear to fail due to gRPC internal errors across all methods, aggregated across all "zoekt_webserver" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "zoekt_webserver" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101110 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"zoekt.webserver.v1.WebserverService",grpc_code!="OK",is_internal_error="true"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])))))))
zoekt: zoekt_webserver_grpc_clients_internal_error_percentage_per_method
Client-observed gRPC internal error percentage per-method over 2m
The percentage of gRPC requests that appear to fail to due to gRPC internal errors per method, aggregated across all "zoekt_webserver" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "zoekt_webserver" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101111 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"zoekt.webserver.v1.WebserverService",grpc_method=~"${zoekt_webserver_method:regex}",grpc_code!="OK",is_internal_error="true"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"zoekt.webserver.v1.WebserverService",grpc_method=~"${zoekt_webserver_method:regex}"}[2m])) by (grpc_method))))))
zoekt: zoekt_webserver_grpc_clients_internal_error_all_codes_per_method
Client-observed gRPC internal error response code rate per-method over 2m
The rate of gRPC internal-error response codes per method, aggregated across all "zoekt_webserver" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "zoekt_webserver" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error) as opposed to normal application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101112 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"zoekt.webserver.v1.WebserverService",is_internal_error="true",grpc_method=~"${zoekt_webserver_method:regex}"}[2m])) by (grpc_method, grpc_code))
Zoekt: Zoekt Webserver GRPC retry metrics
zoekt: zoekt_webserver_grpc_clients_retry_percentage_across_all_methods
Client retry percentage across all methods over 2m
The percentage of gRPC requests that were retried across all methods, aggregated across all "zoekt_webserver" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101200 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"zoekt.webserver.v1.WebserverService",is_retried="true"}[2m])))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"zoekt.webserver.v1.WebserverService"}[2m])))))))
zoekt: zoekt_webserver_grpc_clients_retry_percentage_per_method
Client retry percentage per-method over 2m
The percentage of gRPC requests that were retried aggregated across all "zoekt_webserver" clients, broken out per method.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101201 on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"zoekt.webserver.v1.WebserverService",is_retried="true",grpc_method=~"${zoekt_webserver_method:regex}"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"zoekt.webserver.v1.WebserverService",grpc_method=~"${zoekt_webserver_method:regex}"}[2m])) by (grpc_method))))))
zoekt: zoekt_webserver_grpc_clients_retry_count_per_method
Client retry count per-method over 2m
The count of gRPC requests that were retried aggregated across all "zoekt_webserver" clients, broken out per method
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101202 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"zoekt.webserver.v1.WebserverService",grpc_method=~"${zoekt_webserver_method:regex}",is_retried="true"}[2m])) by (grpc_method))
Zoekt: Data disk I/O metrics
zoekt: data_disk_reads_sec
Read request rate over 1m (per instance)
The number of read requests that were issued to the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), zoekt could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device zoekt is using, not the load zoekt is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101300 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_completed_total{instance=~`node-exporter.*`}[1m])))))
zoekt: data_disk_writes_sec
Write request rate over 1m (per instance)
The number of write requests that were issued to the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), zoekt could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device zoekt is using, not the load zoekt is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101301 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_completed_total{instance=~`node-exporter.*`}[1m])))))
zoekt: data_disk_read_throughput
Read throughput over 1m (per instance)
The amount of data that was read from the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), zoekt could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device zoekt is using, not the load zoekt is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101310 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_read_bytes_total{instance=~`node-exporter.*`}[1m])))))
zoekt: data_disk_write_throughput
Write throughput over 1m (per instance)
The amount of data that was written to the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), zoekt could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device zoekt is using, not the load zoekt is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101311 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_written_bytes_total{instance=~`node-exporter.*`}[1m])))))
zoekt: data_disk_read_duration
Average read duration over 1m (per instance)
The average time for read requests issued to the device to be served. This includes the time spent by the requests in queue and the time spent servicing them.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), zoekt could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device zoekt is using, not the load zoekt is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101320 on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_read_time_seconds_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_completed_total{instance=~`node-exporter.*`}[1m])))))))
zoekt: data_disk_write_duration
Average write duration over 1m (per instance)
The average time for write requests issued to the device to be served. This includes the time spent by the requests in queue and the time spent servicing them.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), zoekt could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device zoekt is using, not the load zoekt is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101321 on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_write_time_seconds_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_completed_total{instance=~`node-exporter.*`}[1m])))))))
zoekt: data_disk_read_request_size
Average read request size over 1m (per instance)
The average size of read requests that were issued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), zoekt could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device zoekt is using, not the load zoekt is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101330 on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_read_bytes_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_completed_total{instance=~`node-exporter.*`}[1m])))))))
zoekt: data_disk_write_request_size)
Average write request size over 1m (per instance)
The average size of write requests that were issued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), zoekt could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device zoekt is using, not the load zoekt is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101331 on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_written_bytes_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_completed_total{instance=~`node-exporter.*`}[1m])))))))
zoekt: data_disk_reads_merged_sec
Merged read request rate over 1m (per instance)
The number of read requests merged per second that were queued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), zoekt could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device zoekt is using, not the load zoekt is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101340 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_merged_total{instance=~`node-exporter.*`}[1m])))))
zoekt: data_disk_writes_merged_sec
Merged writes request rate over 1m (per instance)
The number of write requests merged per second that were queued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), zoekt could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device zoekt is using, not the load zoekt is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101341 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_merged_total{instance=~`node-exporter.*`}[1m])))))
zoekt: data_disk_average_queue_size
Average queue size over 1m (per instance)
The number of I/O operations that were being queued or being serviced. See https://blog.actorsfit.com/a?ID=00200-428fa2ac-e338-4540-848c-af9a3eb1ebd2 for background (avgqu-sz).
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), zoekt could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device zoekt is using, not the load zoekt is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101350 on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (zoekt_indexserver_mount_point_info{mount_name="indexDir",instance=~`${instance:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_io_time_weighted_seconds_total{instance=~`node-exporter.*`}[1m])))))
Zoekt: [indexed-search-indexer] Golang runtime monitoring
zoekt: go_goroutines
Maximum active goroutines
A high value here indicates a possible goroutine leak.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101400 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_goroutines{job=~".*indexed-search-indexer"})
zoekt: go_gc_duration_seconds
Maximum go garbage collection duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101401 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_gc_duration_seconds{job=~".*indexed-search-indexer"})
Zoekt: [indexed-search] Golang runtime monitoring
zoekt: go_goroutines
Maximum active goroutines
A high value here indicates a possible goroutine leak.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101500 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_goroutines{job=~".*indexed-search"})
zoekt: go_gc_duration_seconds
Maximum go garbage collection duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101501 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_gc_duration_seconds{job=~".*indexed-search"})
Zoekt: Kubernetes monitoring (only available on Kubernetes)
zoekt: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/zoekt/zoekt?viewPanel=101600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*indexed-search"}) / count by (app) (up{app=~".*indexed-search"}) * 100
Prometheus
Sourcegraph's all-in-one Prometheus and Alertmanager service.
To see this dashboard, visit /-/debug/grafana/d/prometheus/prometheus on your Sourcegraph instance.
Prometheus: Metrics
prometheus: metrics_cardinality
Metrics with highest cardinalities
The 10 highest-cardinality metrics collected by this Prometheus instance.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLtopk(10, count by (__name__, job)({__name__!=""}))
prometheus: samples_scraped
Samples scraped by job
The number of samples scraped after metric relabeling was applied by this Prometheus instance.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(job) (scrape_samples_post_metric_relabeling{job!=""})
prometheus: prometheus_rule_eval_duration
Average prometheus rule group evaluation duration over 10m by rule group
A high value here indicates Prometheus rule evaluation is taking longer than expected. It might indicate that certain rule groups are taking too long to evaluate, or Prometheus is underprovisioned.
Rules that Sourcegraph ships with are grouped under /sg_config_prometheus. Custom rules are grouped under /sg_prometheus_addons.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(rule_group) (avg_over_time(prometheus_rule_group_last_duration_seconds[10m]))
prometheus: prometheus_rule_eval_failures
Failed prometheus rule evaluations over 5m by rule group
Rules that Sourcegraph ships with are grouped under /sg_config_prometheus. Custom rules are grouped under /sg_prometheus_addons.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(rule_group) (rate(prometheus_rule_evaluation_failures_total[5m]))
Prometheus: Alerts
prometheus: alertmanager_notification_latency
Alertmanager notification latency over 1m by integration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(integration) (rate(alertmanager_notification_latency_seconds_sum[1m]))
prometheus: alertmanager_notification_failures
Failed alertmanager notifications over 1m by integration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(integration) (rate(alertmanager_notifications_failed_total[1m]))
Prometheus: Internals
prometheus: prometheus_config_status
Prometheus configuration reload status
A 1 indicates Prometheus reloaded its configuration successfully.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLprometheus_config_last_reload_successful
prometheus: alertmanager_config_status
Alertmanager configuration reload status
A 1 indicates Alertmanager reloaded its configuration successfully.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLalertmanager_config_last_reload_successful
prometheus: prometheus_tsdb_op_failure
Prometheus tsdb failures by operation over 1m by operation
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLincrease(label_replace({__name__=~"prometheus_tsdb_(.*)_failed_total"}, "operation", "$1", "__name__", "(.+)s_failed_total")[5m:1m])
prometheus: prometheus_target_sample_exceeded
Prometheus scrapes that exceed the sample limit over 10m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLincrease(prometheus_target_scrapes_exceeded_sample_limit_total[10m])
prometheus: prometheus_target_sample_duplicate
Prometheus scrapes rejected due to duplicate timestamps over 10m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLincrease(prometheus_target_scrapes_sample_duplicate_timestamp_total[10m])
Prometheus: Container monitoring (not available on server)
prometheus: container_missing
Container missing
This value is the number of times a container has not been seen for more than one minute. If you observe this value change independent of deployment events (such as an upgrade), it could indicate pods are being OOM killed or terminated for some other reasons.
- Kubernetes:
- Determine if the pod was OOM killed using
kubectl describe pod prometheus(look forOOMKilled: true) and, if so, consider increasing the memory limit in the relevantDeployment.yaml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingkubectl logs -p prometheus.
- Determine if the pod was OOM killed using
- Docker Compose:
- Determine if the pod was OOM killed using
docker inspect -f '\{\{json .State\}\}' prometheus(look for"OOMKilled":true) and, if so, consider increasing the memory limit of the prometheus container indocker-compose.yml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingdocker logs prometheus(note this will include logs from the previous and currently running container).
- Determine if the pod was OOM killed using
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLcount by(name) ((time() - container_last_seen{name=~"^prometheus.*"}) > 60)
prometheus: container_cpu_usage
Container cpu usage total (1m average) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^prometheus.*"}
prometheus: container_memory_usage
Container memory usage by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100302 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^prometheus.*"}
prometheus: fs_io_operations
Filesystem reads and writes rate by instance over 1h
This value indicates the number of filesystem read and write operations by containers of this service. When extremely high, this can indicate a resource usage problem, or can cause problems with the service itself, especially if high values or spikes correlate with {{CONTAINER_NAME}} issues.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100303 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(name) (rate(container_fs_reads_total{name=~"^prometheus.*"}[1h]) + rate(container_fs_writes_total{name=~"^prometheus.*"}[1h]))
Prometheus: Provisioning indicators (not available on server)
prometheus: provisioning_container_cpu_usage_long_term
Container cpu usage total (90th percentile over 1d) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLquantile_over_time(0.9, cadvisor_container_cpu_usage_percentage_total{name=~"^prometheus.*"}[1d])
prometheus: provisioning_container_memory_usage_long_term
Container memory usage (1d maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^prometheus.*"}[1d])
prometheus: provisioning_container_cpu_usage_short_term
Container cpu usage total (5m maximum) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100410 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^prometheus.*"}[5m])
prometheus: provisioning_container_memory_usage_short_term
Container memory usage (5m maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100411 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^prometheus.*"}[5m])
prometheus: container_oomkill_events_total
Container OOMKILL events total by instance
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100412 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^prometheus.*"})
Prometheus: Kubernetes monitoring (only available on Kubernetes)
prometheus: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/prometheus/prometheus?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*prometheus"}) / count by (app) (up{app=~".*prometheus"}) * 100
Executor
Executes jobs in an isolated environment.
To see this dashboard, visit /-/debug/grafana/d/executor/executor on your Sourcegraph instance.
Executor: Executor: Executor jobs
executor: multiqueue_executor_dequeue_cache_size
Unprocessed executor job dequeue cache size for multiqueue executors
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLmultiqueue_executor_dequeue_cache_size{queue=~"$queue",job=~"^(executor|sourcegraph-code-intel-indexers|executor-batches|frontend|sourcegraph-frontend|worker|sourcegraph-executors).*"}
Executor: Executor: Executor jobs
executor: executor_handlers
Executor active handlers
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_executor_processor_handlers{queue=~"${queue:regex}",sg_job=~"^sourcegraph-executors.*"})
executor: executor_processor_total
Executor operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_executor_processor_total{queue=~"${queue:regex}",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: executor_processor_99th_percentile_duration
Aggregate successful executor operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_executor_processor_duration_seconds_bucket{queue=~"${queue:regex}",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: executor_processor_errors_total
Executor operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100112 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_executor_processor_errors_total{queue=~"${queue:regex}",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: executor_processor_error_rate
Executor operation error rate over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100113 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_executor_processor_errors_total{queue=~"${queue:regex}",sg_job=~"^sourcegraph-executors.*"}[5m])) / (sum(increase(src_executor_processor_total{queue=~"${queue:regex}",sg_job=~"^sourcegraph-executors.*"}[5m])) + sum(increase(src_executor_processor_errors_total{queue=~"${queue:regex}",sg_job=~"^sourcegraph-executors.*"}[5m]))) * 100
Executor: Executor: Queue API client
executor: apiworker_apiclient_queue_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_apiclient_queue_total{sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_apiclient_queue_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_apiworker_apiclient_queue_duration_seconds_bucket{sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_apiclient_queue_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_apiclient_queue_errors_total{sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_apiclient_queue_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_apiclient_queue_errors_total{sg_job=~"^sourcegraph-executors.*"}[5m])) / (sum(increase(src_apiworker_apiclient_queue_total{sg_job=~"^sourcegraph-executors.*"}[5m])) + sum(increase(src_apiworker_apiclient_queue_errors_total{sg_job=~"^sourcegraph-executors.*"}[5m]))) * 100
executor: apiworker_apiclient_queue_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_apiclient_queue_total{sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_apiclient_queue_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_apiworker_apiclient_queue_duration_seconds_bucket{sg_job=~"^sourcegraph-executors.*"}[5m])))
executor: apiworker_apiclient_queue_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_apiclient_queue_errors_total{sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_apiclient_queue_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100213 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_apiclient_queue_errors_total{sg_job=~"^sourcegraph-executors.*"}[5m])) / (sum by (op)(increase(src_apiworker_apiclient_queue_total{sg_job=~"^sourcegraph-executors.*"}[5m])) + sum by (op)(increase(src_apiworker_apiclient_queue_errors_total{sg_job=~"^sourcegraph-executors.*"}[5m]))) * 100
Executor: Executor: Files API client
executor: apiworker_apiclient_files_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_apiclient_files_total{sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_apiclient_files_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_apiworker_apiclient_files_duration_seconds_bucket{sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_apiclient_files_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100302 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_apiclient_files_errors_total{sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_apiclient_files_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100303 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_apiclient_files_errors_total{sg_job=~"^sourcegraph-executors.*"}[5m])) / (sum(increase(src_apiworker_apiclient_files_total{sg_job=~"^sourcegraph-executors.*"}[5m])) + sum(increase(src_apiworker_apiclient_files_errors_total{sg_job=~"^sourcegraph-executors.*"}[5m]))) * 100
executor: apiworker_apiclient_files_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100310 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_apiclient_files_total{sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_apiclient_files_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100311 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_apiworker_apiclient_files_duration_seconds_bucket{sg_job=~"^sourcegraph-executors.*"}[5m])))
executor: apiworker_apiclient_files_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100312 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_apiclient_files_errors_total{sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_apiclient_files_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100313 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_apiclient_files_errors_total{sg_job=~"^sourcegraph-executors.*"}[5m])) / (sum by (op)(increase(src_apiworker_apiclient_files_total{sg_job=~"^sourcegraph-executors.*"}[5m])) + sum by (op)(increase(src_apiworker_apiclient_files_errors_total{sg_job=~"^sourcegraph-executors.*"}[5m]))) * 100
Executor: Executor: Job setup
executor: apiworker_command_total
Aggregate command operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_command_total{op=~"setup.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_99th_percentile_duration
Aggregate successful command operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_apiworker_command_duration_seconds_bucket{op=~"setup.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_errors_total
Aggregate command operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100402 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_command_errors_total{op=~"setup.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_error_rate
Aggregate command operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100403 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_command_errors_total{op=~"setup.*",sg_job=~"^sourcegraph-executors.*"}[5m])) / (sum(increase(src_apiworker_command_total{op=~"setup.*",sg_job=~"^sourcegraph-executors.*"}[5m])) + sum(increase(src_apiworker_command_errors_total{op=~"setup.*",sg_job=~"^sourcegraph-executors.*"}[5m]))) * 100
executor: apiworker_command_total
Command operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100410 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_command_total{op=~"setup.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_99th_percentile_duration
99th percentile successful command operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100411 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_apiworker_command_duration_seconds_bucket{op=~"setup.*",sg_job=~"^sourcegraph-executors.*"}[5m])))
executor: apiworker_command_errors_total
Command operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100412 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_command_errors_total{op=~"setup.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_error_rate
Command operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100413 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_command_errors_total{op=~"setup.*",sg_job=~"^sourcegraph-executors.*"}[5m])) / (sum by (op)(increase(src_apiworker_command_total{op=~"setup.*",sg_job=~"^sourcegraph-executors.*"}[5m])) + sum by (op)(increase(src_apiworker_command_errors_total{op=~"setup.*",sg_job=~"^sourcegraph-executors.*"}[5m]))) * 100
Executor: Executor: Job execution
executor: apiworker_command_total
Aggregate command operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_command_total{op=~"exec.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_99th_percentile_duration
Aggregate successful command operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100501 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_apiworker_command_duration_seconds_bucket{op=~"exec.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_errors_total
Aggregate command operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100502 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_command_errors_total{op=~"exec.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_error_rate
Aggregate command operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100503 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_command_errors_total{op=~"exec.*",sg_job=~"^sourcegraph-executors.*"}[5m])) / (sum(increase(src_apiworker_command_total{op=~"exec.*",sg_job=~"^sourcegraph-executors.*"}[5m])) + sum(increase(src_apiworker_command_errors_total{op=~"exec.*",sg_job=~"^sourcegraph-executors.*"}[5m]))) * 100
executor: apiworker_command_total
Command operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100510 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_command_total{op=~"exec.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_99th_percentile_duration
99th percentile successful command operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100511 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_apiworker_command_duration_seconds_bucket{op=~"exec.*",sg_job=~"^sourcegraph-executors.*"}[5m])))
executor: apiworker_command_errors_total
Command operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100512 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_command_errors_total{op=~"exec.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_error_rate
Command operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100513 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_command_errors_total{op=~"exec.*",sg_job=~"^sourcegraph-executors.*"}[5m])) / (sum by (op)(increase(src_apiworker_command_total{op=~"exec.*",sg_job=~"^sourcegraph-executors.*"}[5m])) + sum by (op)(increase(src_apiworker_command_errors_total{op=~"exec.*",sg_job=~"^sourcegraph-executors.*"}[5m]))) * 100
Executor: Executor: Job teardown
executor: apiworker_command_total
Aggregate command operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_command_total{op=~"teardown.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_99th_percentile_duration
Aggregate successful command operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100601 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_apiworker_command_duration_seconds_bucket{op=~"teardown.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_errors_total
Aggregate command operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100602 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_command_errors_total{op=~"teardown.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_error_rate
Aggregate command operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100603 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_apiworker_command_errors_total{op=~"teardown.*",sg_job=~"^sourcegraph-executors.*"}[5m])) / (sum(increase(src_apiworker_command_total{op=~"teardown.*",sg_job=~"^sourcegraph-executors.*"}[5m])) + sum(increase(src_apiworker_command_errors_total{op=~"teardown.*",sg_job=~"^sourcegraph-executors.*"}[5m]))) * 100
executor: apiworker_command_total
Command operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100610 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_command_total{op=~"teardown.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_99th_percentile_duration
99th percentile successful command operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100611 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_apiworker_command_duration_seconds_bucket{op=~"teardown.*",sg_job=~"^sourcegraph-executors.*"}[5m])))
executor: apiworker_command_errors_total
Command operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100612 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_command_errors_total{op=~"teardown.*",sg_job=~"^sourcegraph-executors.*"}[5m]))
executor: apiworker_command_error_rate
Command operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100613 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_apiworker_command_errors_total{op=~"teardown.*",sg_job=~"^sourcegraph-executors.*"}[5m])) / (sum by (op)(increase(src_apiworker_command_total{op=~"teardown.*",sg_job=~"^sourcegraph-executors.*"}[5m])) + sum by (op)(increase(src_apiworker_command_errors_total{op=~"teardown.*",sg_job=~"^sourcegraph-executors.*"}[5m]))) * 100
Executor: Executor: Compute instance metrics
executor: node_cpu_utilization
CPU utilization (minus idle/iowait)
Indicates the amount of CPU time excluding idle and iowait time, divided by the number of cores, as a percentage.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100700 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_cpu_seconds_total{sg_job=~"sourcegraph-executors",mode!~"(idle|iowait)",sg_instance=~"$instance"}[$__rate_interval])) by(sg_instance) / count(node_cpu_seconds_total{sg_job=~"sourcegraph-executors",mode="system",sg_instance=~"$instance"}) by (sg_instance) * 100
executor: node_cpu_saturation_cpu_wait
CPU saturation (time waiting)
Indicates the average summed time a number of (but strictly not all) non-idle processes spent waiting for CPU time. If this is higher than normal, then the CPU is underpowered for the workload and more powerful machines should be provisioned. This only represents a "less-than-all processes" time, because for processes to be waiting for CPU time there must be other process(es) consuming CPU time.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100701 on your Sourcegraph instance.
Technical details
Query:
SHELLrate(node_pressure_cpu_waiting_seconds_total{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval])
executor: node_memory_utilization
Memory utilization
Indicates the amount of available memory (including cache and buffers) as a percentage. Consistently high numbers are generally fine so long memory saturation figures are within acceptable ranges, these figures may be more useful for informing executor provisioning decisions, such as increasing worker parallelism, down-sizing machines etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100710 on your Sourcegraph instance.
Technical details
Query:
SHELL(1 - sum(node_memory_MemAvailable_bytes{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}) by (sg_instance) / sum(node_memory_MemTotal_bytes{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}) by (sg_instance)) * 100
executor: node_memory_saturation_vmeff
Memory saturation (vmem efficiency)
Indicates the efficiency of page reclaim, calculated as pgsteal/pgscan. Optimal figures are short spikes of near 100% and above, indicating that a high ratio of scanned pages are actually being freed, or exactly 0%, indicating that pages arent being scanned as there is no memory pressure. Sustained numbers >~100% may be sign of imminent memory exhaustion, while sustained 0% < x < ~100% figures are very serious.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100711 on your Sourcegraph instance.
Technical details
Query:
SHELL(rate(node_vmstat_pgsteal_anon{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval]) + rate(node_vmstat_pgsteal_direct{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval]) + rate(node_vmstat_pgsteal_file{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval]) + rate(node_vmstat_pgsteal_kswapd{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval])) / (rate(node_vmstat_pgscan_anon{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval]) + rate(node_vmstat_pgscan_direct{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval]) + rate(node_vmstat_pgscan_file{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval]) + rate(node_vmstat_pgscan_kswapd{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval])) * 100
executor: node_memory_saturation_pressure_stalled
Memory saturation (fully stalled)
Indicates the amount of time all non-idle processes were stalled waiting on memory operations to complete. This is often correlated with vmem efficiency ratio when pressure on available memory is high. If they`re not correlated, this could indicate issues with the machine hardware and/or configuration.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100712 on your Sourcegraph instance.
Technical details
Query:
SHELLrate(node_pressure_memory_stalled_seconds_total{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval])
executor: node_io_disk_utilization
Disk IO utilization (percentage time spent in IO)
Indicates the percentage of time a disk was busy. If this is less than 100%, then the disk has spare utilization capacity. However, a value of 100% does not necesarily indicate the disk is at max capacity. For single, serial request-serving devices, 100% may indicate maximum saturation, but for SSDs and RAID arrays this is less likely to be the case, as they are capable of serving multiple requests in parallel, other metrics such as throughput and request queue size should be factored in.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100720 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(label_replace(label_replace(rate(node_disk_io_time_seconds_total{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval]), "disk", "$1", "device", "^([^d].+)"), "disk", "ignite", "device", "dm-.*")) by(sg_instance,disk) * 100
executor: node_io_disk_saturation
Disk IO saturation (avg IO queue size)
Indicates the number of outstanding/queued IO requests. High but short-lived queue sizes may not present an issue, but if theyre consistently/often high and/or monotonically increasing, the disk may be failing or simply too slow for the amount of activity required. Consider replacing the drive(s) with SSDs if they are not already and/or replacing the faulty drive(s), if any.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100721 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(label_replace(label_replace(rate(node_disk_io_time_weighted_seconds_total{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval]), "disk", "$1", "device", "^([^d].+)"), "disk", "ignite", "device", "dm-.*")) by(sg_instance,disk)
executor: node_io_disk_saturation_pressure_full
Disk IO saturation (avg time of all processes stalled)
Indicates the averaged amount of time for which all non-idle processes were stalled waiting for IO to complete simultaneously aka where no processes could make progress.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100722 on your Sourcegraph instance.
Technical details
Query:
SHELLrate(node_pressure_io_stalled_seconds_total{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval])
executor: node_io_network_utilization
Network IO utilization (Rx)
Indicates the average summed receiving throughput of all network interfaces. This is often predominantly composed of the WAN/internet-connected interface, and knowing normal/good figures depends on knowing the bandwidth of the underlying hardware and the workloads.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100730 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_network_receive_bytes_total{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval])) by(sg_instance) * 8
executor: node_io_network_saturation
Network IO saturation (Rx packets dropped)
Number of dropped received packets. This can happen if the receive queues/buffers become full due to slow packet processing throughput. The queues/buffers could be configured to be larger as a stop-gap but the processing application should be investigated as soon as possible. https://www.kernel.org/doc/html/latest/networking/statistics.html#:~:text=not%20otherwise%20counted.-,rx_dropped,-Number%20of%20packets
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100731 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_network_receive_drop_total{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval])) by(sg_instance)
executor: node_io_network_saturation
Network IO errors (Rx)
Number of bad/malformed packets received. https://www.kernel.org/doc/html/latest/networking/statistics.html#:~:text=excluding%20the%20FCS.-,rx_errors,-Total%20number%20of
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100732 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_network_receive_errs_total{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval])) by(sg_instance)
executor: node_io_network_utilization
Network IO utilization (Tx)
Indicates the average summed transmitted throughput of all network interfaces. This is often predominantly composed of the WAN/internet-connected interface, and knowing normal/good figures depends on knowing the bandwidth of the underlying hardware and the workloads.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100740 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_network_transmit_bytes_total{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval])) by(sg_instance) * 8
executor: node_io_network_saturation
Network IO saturation (Tx packets dropped)
Number of dropped transmitted packets. This can happen if the receiving side`s receive queues/buffers become full due to slow packet processing throughput, the network link is congested etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100741 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_network_transmit_drop_total{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval])) by(sg_instance)
executor: node_io_network_saturation
Network IO errors (Tx)
Number of packet transmission errors. This is distinct from tx packet dropping, and can indicate a failing NIC, improperly configured network options anywhere along the line, signal noise etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100742 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_network_transmit_errs_total{sg_job=~"sourcegraph-executors",sg_instance=~"$instance"}[$__rate_interval])) by(sg_instance)
Executor: Executor: Docker Registry Mirror instance metrics
executor: node_cpu_utilization
CPU utilization (minus idle/iowait)
Indicates the amount of CPU time excluding idle and iowait time, divided by the number of cores, as a percentage.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100800 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_cpu_seconds_total{sg_job=~"sourcegraph-executors-registry",mode!~"(idle|iowait)",sg_instance=~"docker-registry"}[$__rate_interval])) by(sg_instance) / count(node_cpu_seconds_total{sg_job=~"sourcegraph-executors-registry",mode="system",sg_instance=~"docker-registry"}) by (sg_instance) * 100
executor: node_cpu_saturation_cpu_wait
CPU saturation (time waiting)
Indicates the average summed time a number of (but strictly not all) non-idle processes spent waiting for CPU time. If this is higher than normal, then the CPU is underpowered for the workload and more powerful machines should be provisioned. This only represents a "less-than-all processes" time, because for processes to be waiting for CPU time there must be other process(es) consuming CPU time.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100801 on your Sourcegraph instance.
Technical details
Query:
SHELLrate(node_pressure_cpu_waiting_seconds_total{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval])
executor: node_memory_utilization
Memory utilization
Indicates the amount of available memory (including cache and buffers) as a percentage. Consistently high numbers are generally fine so long memory saturation figures are within acceptable ranges, these figures may be more useful for informing executor provisioning decisions, such as increasing worker parallelism, down-sizing machines etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100810 on your Sourcegraph instance.
Technical details
Query:
SHELL(1 - sum(node_memory_MemAvailable_bytes{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}) by (sg_instance) / sum(node_memory_MemTotal_bytes{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}) by (sg_instance)) * 100
executor: node_memory_saturation_vmeff
Memory saturation (vmem efficiency)
Indicates the efficiency of page reclaim, calculated as pgsteal/pgscan. Optimal figures are short spikes of near 100% and above, indicating that a high ratio of scanned pages are actually being freed, or exactly 0%, indicating that pages arent being scanned as there is no memory pressure. Sustained numbers >~100% may be sign of imminent memory exhaustion, while sustained 0% < x < ~100% figures are very serious.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100811 on your Sourcegraph instance.
Technical details
Query:
SHELL(rate(node_vmstat_pgsteal_anon{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval]) + rate(node_vmstat_pgsteal_direct{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval]) + rate(node_vmstat_pgsteal_file{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval]) + rate(node_vmstat_pgsteal_kswapd{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval])) / (rate(node_vmstat_pgscan_anon{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval]) + rate(node_vmstat_pgscan_direct{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval]) + rate(node_vmstat_pgscan_file{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval]) + rate(node_vmstat_pgscan_kswapd{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval])) * 100
executor: node_memory_saturation_pressure_stalled
Memory saturation (fully stalled)
Indicates the amount of time all non-idle processes were stalled waiting on memory operations to complete. This is often correlated with vmem efficiency ratio when pressure on available memory is high. If they`re not correlated, this could indicate issues with the machine hardware and/or configuration.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100812 on your Sourcegraph instance.
Technical details
Query:
SHELLrate(node_pressure_memory_stalled_seconds_total{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval])
executor: node_io_disk_utilization
Disk IO utilization (percentage time spent in IO)
Indicates the percentage of time a disk was busy. If this is less than 100%, then the disk has spare utilization capacity. However, a value of 100% does not necesarily indicate the disk is at max capacity. For single, serial request-serving devices, 100% may indicate maximum saturation, but for SSDs and RAID arrays this is less likely to be the case, as they are capable of serving multiple requests in parallel, other metrics such as throughput and request queue size should be factored in.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100820 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(label_replace(label_replace(rate(node_disk_io_time_seconds_total{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval]), "disk", "$1", "device", "^([^d].+)"), "disk", "ignite", "device", "dm-.*")) by(sg_instance,disk) * 100
executor: node_io_disk_saturation
Disk IO saturation (avg IO queue size)
Indicates the number of outstanding/queued IO requests. High but short-lived queue sizes may not present an issue, but if theyre consistently/often high and/or monotonically increasing, the disk may be failing or simply too slow for the amount of activity required. Consider replacing the drive(s) with SSDs if they are not already and/or replacing the faulty drive(s), if any.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100821 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(label_replace(label_replace(rate(node_disk_io_time_weighted_seconds_total{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval]), "disk", "$1", "device", "^([^d].+)"), "disk", "ignite", "device", "dm-.*")) by(sg_instance,disk)
executor: node_io_disk_saturation_pressure_full
Disk IO saturation (avg time of all processes stalled)
Indicates the averaged amount of time for which all non-idle processes were stalled waiting for IO to complete simultaneously aka where no processes could make progress.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100822 on your Sourcegraph instance.
Technical details
Query:
SHELLrate(node_pressure_io_stalled_seconds_total{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval])
executor: node_io_network_utilization
Network IO utilization (Rx)
Indicates the average summed receiving throughput of all network interfaces. This is often predominantly composed of the WAN/internet-connected interface, and knowing normal/good figures depends on knowing the bandwidth of the underlying hardware and the workloads.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100830 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_network_receive_bytes_total{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval])) by(sg_instance) * 8
executor: node_io_network_saturation
Network IO saturation (Rx packets dropped)
Number of dropped received packets. This can happen if the receive queues/buffers become full due to slow packet processing throughput. The queues/buffers could be configured to be larger as a stop-gap but the processing application should be investigated as soon as possible. https://www.kernel.org/doc/html/latest/networking/statistics.html#:~:text=not%20otherwise%20counted.-,rx_dropped,-Number%20of%20packets
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100831 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_network_receive_drop_total{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval])) by(sg_instance)
executor: node_io_network_saturation
Network IO errors (Rx)
Number of bad/malformed packets received. https://www.kernel.org/doc/html/latest/networking/statistics.html#:~:text=excluding%20the%20FCS.-,rx_errors,-Total%20number%20of
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100832 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_network_receive_errs_total{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval])) by(sg_instance)
executor: node_io_network_utilization
Network IO utilization (Tx)
Indicates the average summed transmitted throughput of all network interfaces. This is often predominantly composed of the WAN/internet-connected interface, and knowing normal/good figures depends on knowing the bandwidth of the underlying hardware and the workloads.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100840 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_network_transmit_bytes_total{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval])) by(sg_instance) * 8
executor: node_io_network_saturation
Network IO saturation (Tx packets dropped)
Number of dropped transmitted packets. This can happen if the receiving side`s receive queues/buffers become full due to slow packet processing throughput, the network link is congested etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100841 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_network_transmit_drop_total{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval])) by(sg_instance)
executor: node_io_network_saturation
Network IO errors (Tx)
Number of packet transmission errors. This is distinct from tx packet dropping, and can indicate a failing NIC, improperly configured network options anywhere along the line, signal noise etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100842 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(node_network_transmit_errs_total{sg_job=~"sourcegraph-executors-registry",sg_instance=~"docker-registry"}[$__rate_interval])) by(sg_instance)
Executor: Golang runtime monitoring
executor: go_goroutines
Maximum active goroutines
A high value here indicates a possible goroutine leak.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100900 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(sg_instance) (go_goroutines{sg_job=~".*sourcegraph-executors"})
executor: go_gc_duration_seconds
Maximum go garbage collection duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/executor/executor?viewPanel=100901 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(sg_instance) (go_gc_duration_seconds{sg_job=~".*sourcegraph-executors"})
Global Containers Resource Usage
Container usage and provisioning indicators of all services.
To see this dashboard, visit /-/debug/grafana/d/containers/containers on your Sourcegraph instance.
Global Containers Resource Usage: Containers (not available on server)
containers: container_memory_usage
Container memory usage of all services
This value indicates the memory usage of all containers.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/containers/containers?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend|gitserver|pgsql|codeintel-db|codeinsights|precise-code-intel-worker|prometheus|redis-cache|redis-store|redis-exporter|searcher|syntect-server|worker|zoekt-indexserver|zoekt-webserver|indexed-search|grafana|blobstore|jaeger).*"}
containers: container_cpu_usage
Container cpu usage total (1m average) across all cores by instance
This value indicates the CPU usage of all containers.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/containers/containers?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend|gitserver|pgsql|codeintel-db|codeinsights|precise-code-intel-worker|prometheus|redis-cache|redis-store|redis-exporter|searcher|syntect-server|worker|zoekt-indexserver|zoekt-webserver|indexed-search|grafana|blobstore|jaeger).*"}
Global Containers Resource Usage: Containers: Provisioning Indicators (not available on server)
containers: container_memory_usage_provisioning
Container memory usage (5m maximum) of services that exceed 80% memory limit
Containers that exceed 80% memory limit. The value indicates potential underprovisioned resources.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/containers/containers?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend|gitserver|pgsql|codeintel-db|codeinsights|precise-code-intel-worker|prometheus|redis-cache|redis-store|redis-exporter|searcher|syntect-server|worker|zoekt-indexserver|zoekt-webserver|indexed-search|grafana|blobstore|jaeger).*"}[5m]) >= 80
containers: container_cpu_usage_provisioning
Container cpu usage total (5m maximum) across all cores of services that exceed 80% cpu limit
Containers that exceed 80% CPU limit. The value indicates potential underprovisioned resources.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/containers/containers?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend|gitserver|pgsql|codeintel-db|codeinsights|precise-code-intel-worker|prometheus|redis-cache|redis-store|redis-exporter|searcher|syntect-server|worker|zoekt-indexserver|zoekt-webserver|indexed-search|grafana|blobstore|jaeger).*"}[5m]) >= 80
containers: container_oomkill_events_total
Container OOMKILL events total
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/containers/containers?viewPanel=100120 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^(frontend|sourcegraph-frontend|gitserver|pgsql|codeintel-db|codeinsights|precise-code-intel-worker|prometheus|redis-cache|redis-store|redis-exporter|searcher|syntect-server|worker|zoekt-indexserver|zoekt-webserver|indexed-search|grafana|blobstore|jaeger).*"}) >= 1
containers: container_missing
Container missing
This value is the number of times a container has not been seen for more than one minute. If you observe this value change independent of deployment events (such as an upgrade), it could indicate pods are being OOM killed or terminated for some other reasons.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/containers/containers?viewPanel=100130 on your Sourcegraph instance.
Technical details
Query:
SHELLcount by(name) ((time() - container_last_seen{name=~"^(frontend|sourcegraph-frontend|gitserver|pgsql|codeintel-db|codeinsights|precise-code-intel-worker|prometheus|redis-cache|redis-store|redis-exporter|searcher|syntect-server|worker|zoekt-indexserver|zoekt-webserver|indexed-search|grafana|blobstore|jaeger).*"}) > 60)
Code Intelligence > Autoindexing
The service at internal/codeintel/autoindexing.
To see this dashboard, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing on your Sourcegraph instance.
Code Intelligence > Autoindexing: Codeintel: Autoindexing > Summary
####codeintel-autoindexing:
Auto-index jobs inserted over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_dbstore_indexes_inserted[5m]))
codeintel-autoindexing: codeintel_autoindexing_error_rate
Auto-indexing job scheduler operation error rate over 10m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_errors_total{op='HandleIndexSchedule',job=~"^${source:regex}.*"}[10m])) / (sum(increase(src_codeintel_autoindexing_total{op='HandleIndexSchedule',job=~"^${source:regex}.*"}[10m])) + sum(increase(src_codeintel_autoindexing_errors_total{op='HandleIndexSchedule',job=~"^${source:regex}.*"}[10m]))) * 100
Code Intelligence > Autoindexing: Codeintel: Autoindexing > Service
codeintel-autoindexing: codeintel_autoindexing_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_autoindexing_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_autoindexing_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_autoindexing_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-autoindexing: codeintel_autoindexing_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_autoindexing_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-autoindexing: codeintel_autoindexing_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100112 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100113 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_autoindexing_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_autoindexing_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Autoindexing: Codeintel: Autoindexing > GQL transport
codeintel-autoindexing: codeintel_autoindexing_transport_graphql_total
Aggregate resolver operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_transport_graphql_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_transport_graphql_99th_percentile_duration
Aggregate successful resolver operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_autoindexing_transport_graphql_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_transport_graphql_errors_total
Aggregate resolver operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_transport_graphql_error_rate
Aggregate resolver operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_autoindexing_transport_graphql_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_autoindexing_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-autoindexing: codeintel_autoindexing_transport_graphql_total
Resolver operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_transport_graphql_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_transport_graphql_99th_percentile_duration
99th percentile successful resolver operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_autoindexing_transport_graphql_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-autoindexing: codeintel_autoindexing_transport_graphql_errors_total
Resolver operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_transport_graphql_error_rate
Resolver operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100213 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_autoindexing_transport_graphql_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_autoindexing_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Autoindexing: Codeintel: Autoindexing > Store (internal)
codeintel-autoindexing: codeintel_autoindexing_store_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_store_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_autoindexing_store_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_store_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100302 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_store_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_store_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100303 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_store_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_autoindexing_store_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_autoindexing_store_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-autoindexing: codeintel_autoindexing_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100310 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_store_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_store_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100311 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_autoindexing_store_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-autoindexing: codeintel_autoindexing_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100312 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_store_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100313 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_store_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_autoindexing_store_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_autoindexing_store_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Autoindexing: Codeintel: Autoindexing > Background jobs (internal)
codeintel-autoindexing: codeintel_autoindexing_background_total
Aggregate background operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_background_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_background_99th_percentile_duration
Aggregate successful background operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_autoindexing_background_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_background_errors_total
Aggregate background operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100402 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_background_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_background_error_rate
Aggregate background operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100403 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_background_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_autoindexing_background_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_autoindexing_background_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-autoindexing: codeintel_autoindexing_background_total
Background operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100410 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_background_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_background_99th_percentile_duration
99th percentile successful background operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100411 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_autoindexing_background_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-autoindexing: codeintel_autoindexing_background_errors_total
Background operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100412 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_background_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_background_error_rate
Background operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100413 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_background_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_autoindexing_background_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_autoindexing_background_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Autoindexing: Codeintel: Autoindexing > Inference service (internal)
codeintel-autoindexing: codeintel_autoindexing_inference_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_inference_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_inference_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100501 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_autoindexing_inference_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_inference_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100502 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_inference_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_inference_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100503 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_inference_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_autoindexing_inference_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_autoindexing_inference_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-autoindexing: codeintel_autoindexing_inference_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100510 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_inference_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_inference_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100511 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_autoindexing_inference_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-autoindexing: codeintel_autoindexing_inference_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100512 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_inference_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_inference_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100513 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_inference_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_autoindexing_inference_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_autoindexing_inference_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Autoindexing: Codeintel: Luasandbox service
codeintel-autoindexing: luasandbox_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_luasandbox_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: luasandbox_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100601 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_luasandbox_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: luasandbox_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100602 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_luasandbox_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: luasandbox_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100603 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_luasandbox_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_luasandbox_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_luasandbox_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-autoindexing: luasandbox_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100610 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_luasandbox_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: luasandbox_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100611 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_luasandbox_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-autoindexing: luasandbox_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100612 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_luasandbox_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: luasandbox_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100613 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_luasandbox_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_luasandbox_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_luasandbox_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Autoindexing: Codeintel: Autoindexing > Janitor task > Codeintel autoindexing janitor unknown repository
codeintel-autoindexing: codeintel_autoindexing_janitor_unknown_repository_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100700 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_janitor_unknown_repository_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_janitor_unknown_repository_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100701 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_janitor_unknown_repository_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_janitor_unknown_repository_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100710 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_janitor_unknown_repository_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_janitor_unknown_repository_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100711 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_autoindexing_janitor_unknown_repository_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-autoindexing: codeintel_autoindexing_janitor_unknown_repository_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100712 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_janitor_unknown_repository_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_janitor_unknown_repository_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100713 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_janitor_unknown_repository_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_autoindexing_janitor_unknown_repository_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_autoindexing_janitor_unknown_repository_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Autoindexing: Codeintel: Autoindexing > Janitor task > Codeintel autoindexing janitor unknown commit
codeintel-autoindexing: codeintel_autoindexing_janitor_unknown_commit_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100800 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_janitor_unknown_commit_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_janitor_unknown_commit_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100801 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_janitor_unknown_commit_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_janitor_unknown_commit_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100810 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_janitor_unknown_commit_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_janitor_unknown_commit_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100811 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_autoindexing_janitor_unknown_commit_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-autoindexing: codeintel_autoindexing_janitor_unknown_commit_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100812 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_janitor_unknown_commit_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_janitor_unknown_commit_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100813 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_janitor_unknown_commit_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_autoindexing_janitor_unknown_commit_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_autoindexing_janitor_unknown_commit_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Autoindexing: Codeintel: Autoindexing > Janitor task > Codeintel autoindexing janitor expired
codeintel-autoindexing: codeintel_autoindexing_janitor_expired_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100900 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_janitor_expired_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_janitor_expired_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100901 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindexing_janitor_expired_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_janitor_expired_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100910 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_janitor_expired_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_janitor_expired_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100911 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_autoindexing_janitor_expired_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-autoindexing: codeintel_autoindexing_janitor_expired_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100912 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_janitor_expired_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-autoindexing: codeintel_autoindexing_janitor_expired_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-autoindexing/codeintel-autoindexing?viewPanel=100913 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindexing_janitor_expired_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_autoindexing_janitor_expired_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_autoindexing_janitor_expired_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Code Nav
The service at internal/codeintel/codenav`.
To see this dashboard, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav on your Sourcegraph instance.
Code Intelligence > Code Nav: Codeintel: CodeNav > Service
codeintel-codenav: codeintel_codenav_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_codenav_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_codenav_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100002 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_codenav_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100003 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_codenav_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_codenav_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_codenav_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-codenav: codeintel_codenav_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_codenav_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_codenav_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-codenav: codeintel_codenav_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100012 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_codenav_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100013 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_codenav_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_codenav_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_codenav_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Code Nav: Codeintel: CodeNav > LSIF store
codeintel-codenav: codeintel_codenav_lsifstore_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_codenav_lsifstore_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_lsifstore_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_codenav_lsifstore_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_lsifstore_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_codenav_lsifstore_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_lsifstore_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_codenav_lsifstore_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_codenav_lsifstore_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_codenav_lsifstore_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-codenav: codeintel_codenav_lsifstore_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_codenav_lsifstore_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_lsifstore_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_codenav_lsifstore_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-codenav: codeintel_codenav_lsifstore_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100112 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_codenav_lsifstore_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_lsifstore_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100113 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_codenav_lsifstore_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_codenav_lsifstore_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_codenav_lsifstore_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Code Nav: Codeintel: CodeNav > GQL Transport
codeintel-codenav: codeintel_codenav_transport_graphql_total
Aggregate resolver operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_codenav_transport_graphql_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_transport_graphql_99th_percentile_duration
Aggregate successful resolver operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_codenav_transport_graphql_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_transport_graphql_errors_total
Aggregate resolver operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_codenav_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_transport_graphql_error_rate
Aggregate resolver operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_codenav_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_codenav_transport_graphql_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_codenav_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-codenav: codeintel_codenav_transport_graphql_total
Resolver operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_codenav_transport_graphql_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_transport_graphql_99th_percentile_duration
99th percentile successful resolver operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_codenav_transport_graphql_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-codenav: codeintel_codenav_transport_graphql_errors_total
Resolver operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_codenav_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_transport_graphql_error_rate
Resolver operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100213 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_codenav_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_codenav_transport_graphql_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_codenav_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Code Nav: Codeintel: CodeNav > Store
codeintel-codenav: codeintel_codenav_store_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_codenav_store_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_codenav_store_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_store_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100302 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_codenav_store_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_store_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100303 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_codenav_store_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_codenav_store_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_codenav_store_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-codenav: codeintel_codenav_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100310 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_codenav_store_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_store_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100311 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_codenav_store_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-codenav: codeintel_codenav_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100312 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_codenav_store_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-codenav: codeintel_codenav_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-codenav/codeintel-codenav?viewPanel=100313 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_codenav_store_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_codenav_store_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_codenav_store_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Policies
The service at internal/codeintel/policies.
To see this dashboard, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies on your Sourcegraph instance.
Code Intelligence > Policies: Codeintel: Policies > Service
codeintel-policies: codeintel_policies_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_policies_total{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_policies_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100002 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_policies_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100003 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_policies_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_policies_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_policies_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-policies: codeintel_policies_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_policies_total{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_policies_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-policies: codeintel_policies_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100012 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_policies_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100013 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_policies_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_policies_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_policies_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Policies: Codeintel: Policies > Store
codeintel-policies: codeintel_policies_store_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_policies_store_total{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_policies_store_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_store_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_policies_store_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_store_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_policies_store_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_policies_store_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_policies_store_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-policies: codeintel_policies_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_policies_store_total{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_store_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_policies_store_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-policies: codeintel_policies_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100112 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_policies_store_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100113 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_policies_store_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_policies_store_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_policies_store_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Policies: Codeintel: Policies > GQL Transport
codeintel-policies: codeintel_policies_transport_graphql_total
Aggregate resolver operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_policies_transport_graphql_total{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_transport_graphql_99th_percentile_duration
Aggregate successful resolver operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_policies_transport_graphql_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_transport_graphql_errors_total
Aggregate resolver operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_policies_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_transport_graphql_error_rate
Aggregate resolver operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_policies_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_policies_transport_graphql_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_policies_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-policies: codeintel_policies_transport_graphql_total
Resolver operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_policies_transport_graphql_total{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_transport_graphql_99th_percentile_duration
99th percentile successful resolver operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_policies_transport_graphql_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-policies: codeintel_policies_transport_graphql_errors_total
Resolver operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_policies_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-policies: codeintel_policies_transport_graphql_error_rate
Resolver operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100213 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_policies_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_policies_transport_graphql_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_policies_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Policies: Codeintel: Policies > Repository Pattern Matcher task
codeintel-policies: codeintel_background_policies_updated_total_total
Lsif repository pattern matcher repositories pattern matcher every 5m
Number of configuration policies whose repository membership list was updated
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-policies/codeintel-policies?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_background_policies_updated_total_total{job=~"^${source:regex}.*"}[5m]))
Code Intelligence > Uploads
The service at internal/codeintel/uploads.
To see this dashboard, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads on your Sourcegraph instance.
Code Intelligence > Uploads: Codeintel: Uploads > Service
codeintel-uploads: codeintel_uploads_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploads_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100002 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100003 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_uploads_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_uploads_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-uploads: codeintel_uploads_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100012 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100013 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > Store (internal)
codeintel-uploads: codeintel_uploads_store_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploads_store_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_store_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_store_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_uploads_store_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_uploads_store_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-uploads: codeintel_uploads_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_store_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_store_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100112 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100113 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_store_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > GQL Transport
codeintel-uploads: codeintel_uploads_transport_graphql_total
Aggregate resolver operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_transport_graphql_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_transport_graphql_99th_percentile_duration
Aggregate successful resolver operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploads_transport_graphql_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_transport_graphql_errors_total
Aggregate resolver operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_transport_graphql_error_rate
Aggregate resolver operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_uploads_transport_graphql_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_uploads_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-uploads: codeintel_uploads_transport_graphql_total
Resolver operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_transport_graphql_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_transport_graphql_99th_percentile_duration
99th percentile successful resolver operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_transport_graphql_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_transport_graphql_errors_total
Resolver operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_transport_graphql_error_rate
Resolver operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100213 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_transport_graphql_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_transport_graphql_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > HTTP Transport
codeintel-uploads: codeintel_uploads_transport_http_total
Aggregate http handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_transport_http_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_transport_http_99th_percentile_duration
Aggregate successful http handler operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploads_transport_http_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_transport_http_errors_total
Aggregate http handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100302 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_transport_http_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_transport_http_error_rate
Aggregate http handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100303 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_transport_http_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum(increase(src_codeintel_uploads_transport_http_total{job=~"^${source:regex}.*"}[5m])) + sum(increase(src_codeintel_uploads_transport_http_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
codeintel-uploads: codeintel_uploads_transport_http_total
Http handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100310 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_transport_http_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_transport_http_99th_percentile_duration
99th percentile successful http handler operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100311 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_transport_http_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_transport_http_errors_total
Http handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100312 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_transport_http_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_transport_http_error_rate
Http handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100313 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_transport_http_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_transport_http_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_transport_http_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > Expiration task
codeintel-uploads: codeintel_background_repositories_scanned_total
Lsif upload repository scan repositories scanned every 5m
Number of repositories scanned for data retention
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_background_repositories_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_background_upload_records_scanned_total
Lsif upload records scan records scanned every 5m
Number of codeintel upload records scanned for data retention
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_background_upload_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_background_commits_scanned_total
Lsif upload commits scanned commits scanned every 5m
Number of commits reachable from a codeintel upload record scanned for data retention
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100402 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_background_commits_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_background_upload_records_expired_total
Lsif upload records expired uploads scanned every 5m
Number of codeintel upload records marked as expired
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100403 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_background_upload_records_expired_total{job=~"^${source:regex}.*"}[5m]))
Code Intelligence > Uploads: Codeintel: Uploads > Janitor task > Codeintel uploads janitor unknown repository
codeintel-uploads: codeintel_uploads_janitor_unknown_repository_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_janitor_unknown_repository_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_unknown_repository_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100501 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_janitor_unknown_repository_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_unknown_repository_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100510 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_unknown_repository_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_unknown_repository_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100511 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_janitor_unknown_repository_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_janitor_unknown_repository_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100512 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_unknown_repository_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_unknown_repository_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100513 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_unknown_repository_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_janitor_unknown_repository_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_janitor_unknown_repository_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > Janitor task > Codeintel uploads janitor unknown commit
codeintel-uploads: codeintel_uploads_janitor_unknown_commit_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_janitor_unknown_commit_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_unknown_commit_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100601 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_janitor_unknown_commit_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_unknown_commit_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100610 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_unknown_commit_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_unknown_commit_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100611 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_janitor_unknown_commit_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_janitor_unknown_commit_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100612 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_unknown_commit_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_unknown_commit_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100613 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_unknown_commit_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_janitor_unknown_commit_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_janitor_unknown_commit_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > Janitor task > Codeintel uploads janitor abandoned
codeintel-uploads: codeintel_uploads_janitor_abandoned_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100700 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_janitor_abandoned_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_abandoned_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100701 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_janitor_abandoned_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_abandoned_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100710 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_abandoned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_abandoned_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100711 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_janitor_abandoned_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_janitor_abandoned_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100712 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_abandoned_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_abandoned_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100713 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_abandoned_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_janitor_abandoned_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_janitor_abandoned_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > Janitor task > Codeintel uploads expirer unreferenced
codeintel-uploads: codeintel_uploads_expirer_unreferenced_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100800 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_expirer_unreferenced_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_expirer_unreferenced_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100801 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_expirer_unreferenced_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_expirer_unreferenced_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100810 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_expirer_unreferenced_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_expirer_unreferenced_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100811 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_expirer_unreferenced_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_expirer_unreferenced_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100812 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_expirer_unreferenced_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_expirer_unreferenced_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100813 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_expirer_unreferenced_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_expirer_unreferenced_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_expirer_unreferenced_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > Janitor task > Codeintel uploads expirer unreferenced graph
codeintel-uploads: codeintel_uploads_expirer_unreferenced_graph_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100900 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_expirer_unreferenced_graph_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_expirer_unreferenced_graph_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100901 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_expirer_unreferenced_graph_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_expirer_unreferenced_graph_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100910 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_expirer_unreferenced_graph_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_expirer_unreferenced_graph_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100911 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_expirer_unreferenced_graph_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_expirer_unreferenced_graph_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100912 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_expirer_unreferenced_graph_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_expirer_unreferenced_graph_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=100913 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_expirer_unreferenced_graph_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_expirer_unreferenced_graph_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_expirer_unreferenced_graph_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > Janitor task > Codeintel uploads hard deleter
codeintel-uploads: codeintel_uploads_hard_deleter_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_hard_deleter_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_hard_deleter_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_hard_deleter_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_hard_deleter_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101010 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_hard_deleter_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_hard_deleter_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101011 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_hard_deleter_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_hard_deleter_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101012 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_hard_deleter_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_hard_deleter_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101013 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_hard_deleter_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_hard_deleter_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_hard_deleter_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > Janitor task > Codeintel uploads janitor audit logs
codeintel-uploads: codeintel_uploads_janitor_audit_logs_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_janitor_audit_logs_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_audit_logs_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_janitor_audit_logs_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_audit_logs_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101110 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_audit_logs_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_audit_logs_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_janitor_audit_logs_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_janitor_audit_logs_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101112 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_audit_logs_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_audit_logs_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101113 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_audit_logs_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_janitor_audit_logs_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_janitor_audit_logs_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > Janitor task > Codeintel uploads janitor scip documents
codeintel-uploads: codeintel_uploads_janitor_scip_documents_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_janitor_scip_documents_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_scip_documents_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_janitor_scip_documents_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_scip_documents_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_scip_documents_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_scip_documents_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_janitor_scip_documents_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_janitor_scip_documents_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_scip_documents_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_janitor_scip_documents_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101213 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_janitor_scip_documents_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_janitor_scip_documents_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_janitor_scip_documents_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > Reconciler task > Codeintel uploads reconciler scip metadata
codeintel-uploads: codeintel_uploads_reconciler_scip_metadata_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_reconciler_scip_metadata_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_reconciler_scip_metadata_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101301 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_reconciler_scip_metadata_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_reconciler_scip_metadata_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101310 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_reconciler_scip_metadata_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_reconciler_scip_metadata_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101311 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_reconciler_scip_metadata_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_reconciler_scip_metadata_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101312 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_reconciler_scip_metadata_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_reconciler_scip_metadata_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101313 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_reconciler_scip_metadata_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_reconciler_scip_metadata_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_reconciler_scip_metadata_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Code Intelligence > Uploads: Codeintel: Uploads > Reconciler task > Codeintel uploads reconciler scip data
codeintel-uploads: codeintel_uploads_reconciler_scip_data_records_scanned_total
Records scanned every 5m
The number of candidate records considered for cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_reconciler_scip_data_records_scanned_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_reconciler_scip_data_records_altered_total
Records altered every 5m
The number of candidate records altered as part of cleanup.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_reconciler_scip_data_records_altered_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_reconciler_scip_data_total
Job invocation operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101410 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_reconciler_scip_data_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_reconciler_scip_data_99th_percentile_duration
99th percentile successful job invocation operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101411 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_reconciler_scip_data_duration_seconds_bucket{job=~"^${source:regex}.*"}[5m])))
codeintel-uploads: codeintel_uploads_reconciler_scip_data_errors_total
Job invocation operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101412 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_reconciler_scip_data_errors_total{job=~"^${source:regex}.*"}[5m]))
codeintel-uploads: codeintel_uploads_reconciler_scip_data_error_rate
Job invocation operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/codeintel-uploads/codeintel-uploads?viewPanel=101413 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_reconciler_scip_data_errors_total{job=~"^${source:regex}.*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_reconciler_scip_data_total{job=~"^${source:regex}.*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_reconciler_scip_data_errors_total{job=~"^${source:regex}.*"}[5m]))) * 100
Telemetry
Monitoring telemetry services in Sourcegraph.
To see this dashboard, visit /-/debug/grafana/d/telemetry/telemetry on your Sourcegraph instance.
Telemetry: Telemetry Gateway Exporter: Events export and queue metrics
telemetry: telemetry_gateway_exporter_queue_size
Telemetry event payloads pending export
The number of events queued to be exported.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_telemetrygatewayexporter_queue_size)
telemetry: telemetry_gateway_exporter_queue_growth
Rate of growth of events export queue over 30m
A positive value indicates the queue is growing.
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(deriv(src_telemetrygatewayexporter_queue_size[30m]))
telemetry: src_telemetrygatewayexporter_exported_events
Events exported from queue per hour
The number of events being exported.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(increase(src_telemetrygatewayexporter_exported_events[1h]))
telemetry: telemetry_gateway_exporter_batch_size
Number of events exported per batch over 30m
The number of events exported in each batch. The largest bucket is the maximum number of events exported per batch.
If the distribution trends to the maximum bucket, then events export throughput is at or approaching saturation - try increasing TELEMETRY_GATEWAY_EXPORTER_EXPORT_BATCH_SIZE or decreasing TELEMETRY_GATEWAY_EXPORTER_EXPORT_INTERVAL.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le) (rate(src_telemetrygatewayexporter_batch_size_bucket[30m]))
Telemetry: Telemetry Gateway Exporter: Events export job operations
telemetry: telemetrygatewayexporter_exporter_total
Events exporter operations every 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_telemetrygatewayexporter_exporter_total{job=~"^worker.*"}[30m]))
telemetry: telemetrygatewayexporter_exporter_99th_percentile_duration
Aggregate successful events exporter operation duration distribution over 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_telemetrygatewayexporter_exporter_duration_seconds_bucket{job=~"^worker.*"}[30m]))
telemetry: telemetrygatewayexporter_exporter_errors_total
Events exporter operation errors every 30m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_telemetrygatewayexporter_exporter_errors_total{job=~"^worker.*"}[30m]))
telemetry: telemetrygatewayexporter_exporter_error_rate
Events exporter operation error rate over 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100103 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_telemetrygatewayexporter_exporter_errors_total{job=~"^worker.*"}[30m])) / (sum(increase(src_telemetrygatewayexporter_exporter_total{job=~"^worker.*"}[30m])) + sum(increase(src_telemetrygatewayexporter_exporter_errors_total{job=~"^worker.*"}[30m]))) * 100
Telemetry: Telemetry Gateway Exporter: Events export queue cleanup job operations
telemetry: telemetrygatewayexporter_queue_cleanup_total
Events export queue cleanup operations every 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_telemetrygatewayexporter_queue_cleanup_total{job=~"^worker.*"}[30m]))
telemetry: telemetrygatewayexporter_queue_cleanup_99th_percentile_duration
Aggregate successful events export queue cleanup operation duration distribution over 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_telemetrygatewayexporter_queue_cleanup_duration_seconds_bucket{job=~"^worker.*"}[30m]))
telemetry: telemetrygatewayexporter_queue_cleanup_errors_total
Events export queue cleanup operation errors every 30m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_telemetrygatewayexporter_queue_cleanup_errors_total{job=~"^worker.*"}[30m]))
telemetry: telemetrygatewayexporter_queue_cleanup_error_rate
Events export queue cleanup operation error rate over 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100203 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_telemetrygatewayexporter_queue_cleanup_errors_total{job=~"^worker.*"}[30m])) / (sum(increase(src_telemetrygatewayexporter_queue_cleanup_total{job=~"^worker.*"}[30m])) + sum(increase(src_telemetrygatewayexporter_queue_cleanup_errors_total{job=~"^worker.*"}[30m]))) * 100
Telemetry: Telemetry Gateway Exporter: Events export queue metrics reporting job operations
telemetry: telemetrygatewayexporter_queue_metrics_reporter_total
Events export backlog metrics reporting operations every 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_telemetrygatewayexporter_queue_metrics_reporter_total{job=~"^worker.*"}[30m]))
telemetry: telemetrygatewayexporter_queue_metrics_reporter_99th_percentile_duration
Aggregate successful events export backlog metrics reporting operation duration distribution over 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_telemetrygatewayexporter_queue_metrics_reporter_duration_seconds_bucket{job=~"^worker.*"}[30m]))
telemetry: telemetrygatewayexporter_queue_metrics_reporter_errors_total
Events export backlog metrics reporting operation errors every 30m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100302 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_telemetrygatewayexporter_queue_metrics_reporter_errors_total{job=~"^worker.*"}[30m]))
telemetry: telemetrygatewayexporter_queue_metrics_reporter_error_rate
Events export backlog metrics reporting operation error rate over 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100303 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_telemetrygatewayexporter_queue_metrics_reporter_errors_total{job=~"^worker.*"}[30m])) / (sum(increase(src_telemetrygatewayexporter_queue_metrics_reporter_total{job=~"^worker.*"}[30m])) + sum(increase(src_telemetrygatewayexporter_queue_metrics_reporter_errors_total{job=~"^worker.*"}[30m]))) * 100
Telemetry: Telemetry persistence
telemetry: telemetry_v2_export_queue_write_failures
Failed writes to events export queue over 5m
Telemetry V2 writes send events into the telemetry_events_export_queue for the exporter to periodically export.
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(increase(src_telemetry_export_store_queued_events{failed="true"}[5m])) / sum(increase(src_telemetry_export_store_queued_events[5m]))) * 100
telemetry: telemetry_v2_event_logs_write_failures
Failed write V2 events to V1 'event_logs' over 5m
Telemetry V2 writes also attempt to tee events into the legacy V1 events format in the event_logs database table for long-term local persistence.
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(increase(src_telemetry_teestore_v1_events{failed="true"}[5m])) / sum(increase(src_telemetry_teestore_v1_events[5m]))) * 100
Telemetry: Telemetry Gateway Exporter: (off by default) User metadata export job operations
telemetry: telemetrygatewayexporter_usermetadata_exporter_total
(off by default) user metadata exporter operations every 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_telemetrygatewayexporter_usermetadata_exporter_total{job=~"^worker.*"}[30m]))
telemetry: telemetrygatewayexporter_usermetadata_exporter_99th_percentile_duration
Aggregate successful (off by default) user metadata exporter operation duration distribution over 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100501 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_telemetrygatewayexporter_usermetadata_exporter_duration_seconds_bucket{job=~"^worker.*"}[30m]))
telemetry: telemetrygatewayexporter_usermetadata_exporter_errors_total
(off by default) user metadata exporter operation errors every 30m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100502 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_telemetrygatewayexporter_usermetadata_exporter_errors_total{job=~"^worker.*"}[30m]))
telemetry: telemetrygatewayexporter_usermetadata_exporter_error_rate
(off by default) user metadata exporter operation error rate over 30m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/telemetry/telemetry?viewPanel=100503 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_telemetrygatewayexporter_usermetadata_exporter_errors_total{job=~"^worker.*"}[30m])) / (sum(increase(src_telemetrygatewayexporter_usermetadata_exporter_total{job=~"^worker.*"}[30m])) + sum(increase(src_telemetrygatewayexporter_usermetadata_exporter_errors_total{job=~"^worker.*"}[30m]))) * 100
OpenTelemetry Collector
The OpenTelemetry collector ingests OpenTelemetry data from Sourcegraph and exports it to the configured backends.
To see this dashboard, visit /-/debug/grafana/d/otel-collector/otel-collector on your Sourcegraph instance.
OpenTelemetry Collector: Receivers
otel-collector: otel_span_receive_rate
Spans received per receiver per minute
Shows the rate of spans accepted by the configured reveiver
A Trace is a collection of spans and a span represents a unit of work or operation. Spans are the building blocks of Traces. The spans have only been accepted by the receiver, which means they still have to move through the configured pipeline to be exported. For more information on tracing and configuration of a OpenTelemetry receiver see https://opentelemetry.io/docs/collector/configuration/#receivers.
See the Exporters section see spans that have made it through the pipeline and are exported.
Depending the configured processors, received spans might be dropped and not exported. For more information on configuring processors see https://opentelemetry.io/docs/collector/configuration/#processors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (receiver) (rate(otelcol_receiver_accepted_spans[1m]))
otel-collector: otel_span_refused
Spans refused per receiver
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (receiver) (rate(otelcol_receiver_refused_spans[1m]))
OpenTelemetry Collector: Exporters
otel-collector: otel_span_export_rate
Spans exported per exporter per minute
Shows the rate of spans being sent by the exporter
A Trace is a collection of spans. A Span represents a unit of work or operation. Spans are the building blocks of Traces. The rate of spans here indicates spans that have made it through the configured pipeline and have been sent to the configured export destination.
For more information on configuring a exporter for the OpenTelemetry collector see https://opentelemetry.io/docs/collector/configuration/#exporters.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (exporter) (rate(otelcol_exporter_sent_spans[1m]))
otel-collector: otel_span_export_failures
Span export failures by exporter
Shows the rate of spans failed to be sent by the configured reveiver. A number higher than 0 for a long period can indicate a problem with the exporter configuration or with the service that is being exported too
For more information on configuring a exporter for the OpenTelemetry collector see https://opentelemetry.io/docs/collector/configuration/#exporters.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (exporter) (rate(otelcol_exporter_send_failed_spans[1m]))
OpenTelemetry Collector: Queue Length
otel-collector: otelcol_exporter_queue_capacity
Exporter queue capacity
Shows the the capacity of the retry queue (in batches).
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (exporter) (rate(otelcol_exporter_queue_capacity{job=~"^.*"}[1m]))
otel-collector: otelcol_exporter_queue_size
Exporter queue size
Shows the current size of retry queue
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (exporter) (rate(otelcol_exporter_queue_size{job=~"^.*"}[1m]))
otel-collector: otelcol_exporter_enqueue_failed_spans
Exporter enqueue failed spans
Shows the rate of spans failed to be enqueued by the configured exporter. A number higher than 0 for a long period can indicate a problem with the exporter configuration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (exporter) (rate(otelcol_exporter_enqueue_failed_spans{job=~"^.*"}[1m]))
OpenTelemetry Collector: Processors
otel-collector: otelcol_processor_dropped_spans
Spans dropped per processor per minute
Shows the rate of spans dropped by the configured processor
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (processor) (rate(otelcol_processor_dropped_spans[1m]))
OpenTelemetry Collector: Collector resource usage
otel-collector: otel_cpu_usage
Cpu usage of the collector
Shows CPU usage as reported by the OpenTelemetry collector.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100400 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (job) (rate(otelcol_process_cpu_seconds{job=~"^.*"}[1m]))
otel-collector: otel_memory_resident_set_size
Memory allocated to the otel collector
Shows the allocated memory Resident Set Size (RSS) as reported by the OpenTelemetry collector.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100401 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (job) (rate(otelcol_process_memory_rss{job=~"^.*"}[1m]))
otel-collector: otel_memory_usage
Memory used by the collector
Shows how much memory is being used by the otel collector.
- High memory usage might indicate thad the configured pipeline is keeping a lot of spans in memory for processing
- Spans failing to be sent and the exporter is configured to retry
- A high batch count by using a batch processor
For more information on configuring processors for the OpenTelemetry collector see https://opentelemetry.io/docs/collector/configuration/#processors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100402 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (job) (rate(otelcol_process_runtime_total_alloc_bytes{job=~"^.*"}[1m]))
OpenTelemetry Collector: Container monitoring (not available on server)
otel-collector: container_missing
Container missing
This value is the number of times a container has not been seen for more than one minute. If you observe this value change independent of deployment events (such as an upgrade), it could indicate pods are being OOM killed or terminated for some other reasons.
- Kubernetes:
- Determine if the pod was OOM killed using
kubectl describe pod otel-collector(look forOOMKilled: true) and, if so, consider increasing the memory limit in the relevantDeployment.yaml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingkubectl logs -p otel-collector.
- Determine if the pod was OOM killed using
- Docker Compose:
- Determine if the pod was OOM killed using
docker inspect -f '\{\{json .State\}\}' otel-collector(look for"OOMKilled":true) and, if so, consider increasing the memory limit of the otel-collector container indocker-compose.yml. - Check the logs before the container restarted to see if there are
panic:messages or similar usingdocker logs otel-collector(note this will include logs from the previous and currently running container).
- Determine if the pod was OOM killed using
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100500 on your Sourcegraph instance.
Technical details
Query:
SHELLcount by(name) ((time() - container_last_seen{name=~"^otel-collector.*"}) > 60)
otel-collector: container_cpu_usage
Container cpu usage total (1m average) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100501 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^otel-collector.*"}
otel-collector: container_memory_usage
Container memory usage by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100502 on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^otel-collector.*"}
otel-collector: fs_io_operations
Filesystem reads and writes rate by instance over 1h
This value indicates the number of filesystem read and write operations by containers of this service. When extremely high, this can indicate a resource usage problem, or can cause problems with the service itself, especially if high values or spikes correlate with {{CONTAINER_NAME}} issues.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100503 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(name) (rate(container_fs_reads_total{name=~"^otel-collector.*"}[1h]) + rate(container_fs_writes_total{name=~"^otel-collector.*"}[1h]))
OpenTelemetry Collector: Kubernetes monitoring (only available on Kubernetes)
otel-collector: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/otel-collector/otel-collector?viewPanel=100600 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*otel-collector"}) / count by (app) (up{app=~".*otel-collector"}) * 100
Completions
Cody chat and code completions.
To see this dashboard, visit /-/debug/grafana/d/completions/completions on your Sourcegraph instance.
Completions: Completions requests
completions: api_request_rate
Rate of completions API requests
Rate (QPS) of requests to cody chat and code completion endpoints.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (code)(irate(src_http_request_duration_seconds_count{route=~"^cody.completions.*"}[5m]))
Completions: Chat completions
completions: chat_completions_p99_stream_duration
Stream: total time (p99)
Time spent on the Stream() invocation, i.e. how long results take to connect, stream results, and finish streaming.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_stream_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: chat_completions_p95_stream_duration
Stream: total time (p95)
Time spent on the Stream() invocation, i.e. how long results take to connect, stream results, and finish streaming.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_stream_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: chat_completions_p75_stream_duration
Stream: total time (p75)
Time spent on the Stream() invocation, i.e. how long results take to connect, stream results, and finish streaming.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_stream_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: chat_completions_p50_stream_duration
Stream: total time (p50)
Time spent on the Stream() invocation, i.e. how long results take to connect, stream results, and finish streaming.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100103 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_stream_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: chat_completions_p99_non_stream_overhead_duration
Non-stream overhead (p99)
Time between Go HTTP handler invocation and Stream() invocation, overhead of e.g. request validation, routing to gateway/other, model resolution, error reporting/tracing, guardrails, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_handler_overhead_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le,model))
completions: chat_completions_p95_non_stream_overhead_duration
Non-stream overhead (p95)
Time between Go HTTP handler invocation and Stream() invocation, overhead of e.g. request validation, routing to gateway/other, model resolution, error reporting/tracing, guardrails, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_handler_overhead_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le,model))
completions: chat_completions_p75_non_stream_overhead_duration
Non-stream overhead (p75)
Time between Go HTTP handler invocation and Stream() invocation, overhead of e.g. request validation, routing to gateway/other, model resolution, error reporting/tracing, guardrails, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100112 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_handler_overhead_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le,model))
completions: chat_completions_p50_non_stream_overhead_duration
Non-stream overhead (p50)
Time between Go HTTP handler invocation and Stream() invocation, overhead of e.g. request validation, routing to gateway/other, model resolution, error reporting/tracing, guardrails, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100113 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_handler_overhead_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le,model))
completions: chat_completions_p99_stream_first_event_duration
Stream: time to first event (p99)
Time between calling Stream(), the client connecting to the server etc. and actually getting the first streaming event back.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100120 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_stream_first_event_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: chat_completions_p95_stream_first_event_duration
Stream: time to first event (p95)
Time between calling Stream(), the client connecting to the server etc. and actually getting the first streaming event back.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100121 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_stream_first_event_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: chat_completions_p75_stream_first_event_duration
Stream: time to first event (p75)
Time between calling Stream(), the client connecting to the server etc. and actually getting the first streaming event back.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100122 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_stream_first_event_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: chat_completions_p50_stream_first_event_duration
Stream: time to first event (p50)
Time between calling Stream(), the client connecting to the server etc. and actually getting the first streaming event back.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100123 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_stream_first_event_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: chat_completions_p99_upstream_roundtrip_duration
Stream: first byte sent -> received (p99)
Time between sending the first byte to the upstream, and then getting the first byte back from the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100130 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_upstream_roundtrip_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le, provider))
completions: chat_completions_p95_upstream_roundtrip_duration
Stream: first byte sent -> received (p95)
Time between sending the first byte to the upstream, and then getting the first byte back from the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100131 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_upstream_roundtrip_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le, provider))
completions: chat_completions_p75_upstream_roundtrip_duration
Stream: first byte sent -> received (p75)
Time between sending the first byte to the upstream, and then getting the first byte back from the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100132 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_upstream_roundtrip_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le, provider))
completions: chat_completions_p50_upstream_roundtrip_duration
Stream: first byte sent -> received (p50)
Time between sending the first byte to the upstream, and then getting the first byte back from the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100133 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_upstream_roundtrip_duration_seconds_bucket{feature="chat_completions",model=~'${model}'}[$sampling_duration])) by (le, provider))
completions: chat_completions_p99_http_connect_total
Stream: HTTP connect: total (p99)
Time spent acquiring an HTTP connection to the upstream, either from an existing pool OR by performing DNS resolution, TCP connection, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100140 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_upstream_connection_total_duration_seconds_bucket[$sampling_duration])) by (le, connection_type, provider))
completions: chat_completions_p95_http_connect_total
Stream: HTTP connect: total (p95)
Time spent acquiring an HTTP connection to the upstream, either from an existing pool OR by performing DNS resolution, TCP connection, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100141 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_upstream_connection_total_duration_seconds_bucket[$sampling_duration])) by (le, connection_type, provider))
completions: chat_completions_p75_http_connect_total
Stream: HTTP connect: total (p75)
Time spent acquiring an HTTP connection to the upstream, either from an existing pool OR by performing DNS resolution, TCP connection, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100142 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_upstream_connection_total_duration_seconds_bucket[$sampling_duration])) by (le, connection_type, provider))
completions: chat_completions_p50_http_connect_total
Stream: HTTP connect: total (p50)
Time spent acquiring an HTTP connection to the upstream, either from an existing pool OR by performing DNS resolution, TCP connection, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100143 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_upstream_connection_total_duration_seconds_bucket[$sampling_duration])) by (le, connection_type, provider))
completions: chat_completions_p99_http_connect_dns
Stream: HTTP connect: dns (p99)
Portion of time spent on DNS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100150 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_upstream_connection_dns_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: chat_completions_p95_http_connect_dns
Stream: HTTP connect: dns (p95)
Portion of time spent on DNS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100151 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_upstream_connection_dns_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: chat_completions_p75_http_connect_dns
Stream: HTTP connect: dns (p75)
Portion of time spent on DNS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100152 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_upstream_connection_dns_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: chat_completions_p50_http_connect_dns
Stream: HTTP connect: dns (p50)
Portion of time spent on DNS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100153 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_upstream_connection_dns_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: chat_completions_p99_http_connect_tls
Stream: HTTP connect: tls (p99)
Portion of time spent on TLS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100160 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_upstream_connection_tls_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: chat_completions_p95_http_connect_tls
Stream: HTTP connect: tls (p95)
Portion of time spent on TLS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100161 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_upstream_connection_tls_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: chat_completions_p75_http_connect_tls
Stream: HTTP connect: tls (p75)
Portion of time spent on TLS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100162 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_upstream_connection_tls_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: chat_completions_p50_http_connect_tls
Stream: HTTP connect: tls (p50)
Portion of time spent on TLS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100163 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_upstream_connection_tls_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: chat_completions_p99_http_connect_dial
Stream: HTTP connect: dial (p99)
Portion of time spent on golang Dial() when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100170 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_upstream_connection_dial_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: chat_completions_p95_http_connect_dial
Stream: HTTP connect: dial (p95)
Portion of time spent on golang Dial() when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100171 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_upstream_connection_dial_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: chat_completions_p75_http_connect_dial
Stream: HTTP connect: dial (p75)
Portion of time spent on golang Dial() when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100172 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_upstream_connection_dial_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: chat_completions_p50_http_connect_dial
Stream: HTTP connect: dial (p50)
Portion of time spent on golang Dial() when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100173 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_upstream_connection_dial_duration_seconds_bucket[$sampling_duration])) by (le, provider))
Completions: Code completions
completions: code_completions_p99_stream_duration
Stream: total time (p99)
Time spent on the Stream() invocation, i.e. how long results take to connect, stream results, and finish streaming.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_stream_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: code_completions_p95_stream_duration
Stream: total time (p95)
Time spent on the Stream() invocation, i.e. how long results take to connect, stream results, and finish streaming.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_stream_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: code_completions_p75_stream_duration
Stream: total time (p75)
Time spent on the Stream() invocation, i.e. how long results take to connect, stream results, and finish streaming.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_stream_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: code_completions_p50_stream_duration
Stream: total time (p50)
Time spent on the Stream() invocation, i.e. how long results take to connect, stream results, and finish streaming.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100203 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_stream_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: code_completions_p99_non_stream_overhead_duration
Non-stream overhead (p99)
Time between Go HTTP handler invocation and Stream() invocation, overhead of e.g. request validation, routing to gateway/other, model resolution, error reporting/tracing, guardrails, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_handler_overhead_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le,model))
completions: code_completions_p95_non_stream_overhead_duration
Non-stream overhead (p95)
Time between Go HTTP handler invocation and Stream() invocation, overhead of e.g. request validation, routing to gateway/other, model resolution, error reporting/tracing, guardrails, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_handler_overhead_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le,model))
completions: code_completions_p75_non_stream_overhead_duration
Non-stream overhead (p75)
Time between Go HTTP handler invocation and Stream() invocation, overhead of e.g. request validation, routing to gateway/other, model resolution, error reporting/tracing, guardrails, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_handler_overhead_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le,model))
completions: code_completions_p50_non_stream_overhead_duration
Non-stream overhead (p50)
Time between Go HTTP handler invocation and Stream() invocation, overhead of e.g. request validation, routing to gateway/other, model resolution, error reporting/tracing, guardrails, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100213 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_handler_overhead_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le,model))
completions: code_completions_p99_stream_first_event_duration
Stream: time to first event (p99)
Time between calling Stream(), the client connecting to the server etc. and actually getting the first streaming event back.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100220 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_stream_first_event_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: code_completions_p95_stream_first_event_duration
Stream: time to first event (p95)
Time between calling Stream(), the client connecting to the server etc. and actually getting the first streaming event back.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100221 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_stream_first_event_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: code_completions_p75_stream_first_event_duration
Stream: time to first event (p75)
Time between calling Stream(), the client connecting to the server etc. and actually getting the first streaming event back.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100222 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_stream_first_event_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: code_completions_p50_stream_first_event_duration
Stream: time to first event (p50)
Time between calling Stream(), the client connecting to the server etc. and actually getting the first streaming event back.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100223 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_stream_first_event_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le, model))
completions: code_completions_p99_upstream_roundtrip_duration
Stream: first byte sent -> received (p99)
Time between sending the first byte to the upstream, and then getting the first byte back from the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100230 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_upstream_roundtrip_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le, provider))
completions: code_completions_p95_upstream_roundtrip_duration
Stream: first byte sent -> received (p95)
Time between sending the first byte to the upstream, and then getting the first byte back from the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100231 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_upstream_roundtrip_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le, provider))
completions: code_completions_p75_upstream_roundtrip_duration
Stream: first byte sent -> received (p75)
Time between sending the first byte to the upstream, and then getting the first byte back from the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100232 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_upstream_roundtrip_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le, provider))
completions: code_completions_p50_upstream_roundtrip_duration
Stream: first byte sent -> received (p50)
Time between sending the first byte to the upstream, and then getting the first byte back from the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100233 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_upstream_roundtrip_duration_seconds_bucket{feature="code_completions",model=~'${model}'}[$sampling_duration])) by (le, provider))
completions: code_completions_p99_http_connect_total
Stream: HTTP connect: total (p99)
Time spent acquiring an HTTP connection to the upstream, either from an existing pool OR by performing DNS resolution, TCP connection, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100240 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_upstream_connection_total_duration_seconds_bucket[$sampling_duration])) by (le, connection_type, provider))
completions: code_completions_p95_http_connect_total
Stream: HTTP connect: total (p95)
Time spent acquiring an HTTP connection to the upstream, either from an existing pool OR by performing DNS resolution, TCP connection, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100241 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_upstream_connection_total_duration_seconds_bucket[$sampling_duration])) by (le, connection_type, provider))
completions: code_completions_p75_http_connect_total
Stream: HTTP connect: total (p75)
Time spent acquiring an HTTP connection to the upstream, either from an existing pool OR by performing DNS resolution, TCP connection, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100242 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_upstream_connection_total_duration_seconds_bucket[$sampling_duration])) by (le, connection_type, provider))
completions: code_completions_p50_http_connect_total
Stream: HTTP connect: total (p50)
Time spent acquiring an HTTP connection to the upstream, either from an existing pool OR by performing DNS resolution, TCP connection, etc.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100243 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_upstream_connection_total_duration_seconds_bucket[$sampling_duration])) by (le, connection_type, provider))
completions: code_completions_p99_http_connect_dns
Stream: HTTP connect: dns (p99)
Portion of time spent on DNS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100250 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_upstream_connection_dns_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: code_completions_p95_http_connect_dns
Stream: HTTP connect: dns (p95)
Portion of time spent on DNS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100251 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_upstream_connection_dns_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: code_completions_p75_http_connect_dns
Stream: HTTP connect: dns (p75)
Portion of time spent on DNS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100252 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_upstream_connection_dns_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: code_completions_p50_http_connect_dns
Stream: HTTP connect: dns (p50)
Portion of time spent on DNS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100253 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_upstream_connection_dns_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: code_completions_p99_http_connect_tls
Stream: HTTP connect: tls (p99)
Portion of time spent on TLS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100260 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_upstream_connection_tls_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: code_completions_p95_http_connect_tls
Stream: HTTP connect: tls (p95)
Portion of time spent on TLS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100261 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_upstream_connection_tls_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: code_completions_p75_http_connect_tls
Stream: HTTP connect: tls (p75)
Portion of time spent on TLS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100262 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_upstream_connection_tls_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: code_completions_p50_http_connect_tls
Stream: HTTP connect: tls (p50)
Portion of time spent on TLS when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100263 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_upstream_connection_tls_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: code_completions_p99_http_connect_dial
Stream: HTTP connect: dial (p99)
Portion of time spent on golang Dial() when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100270 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_completions_upstream_connection_dial_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: code_completions_p95_http_connect_dial
Stream: HTTP connect: dial (p95)
Portion of time spent on golang Dial() when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100271 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completions_upstream_connection_dial_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: code_completions_p75_http_connect_dial
Stream: HTTP connect: dial (p75)
Portion of time spent on golang Dial() when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100272 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_completions_upstream_connection_dial_duration_seconds_bucket[$sampling_duration])) by (le, provider))
completions: code_completions_p50_http_connect_dial
Stream: HTTP connect: dial (p50)
Portion of time spent on golang Dial() when acquiring an HTTP connection to the upstream.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100273 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.50, sum(rate(src_completions_upstream_connection_dial_duration_seconds_bucket[$sampling_duration])) by (le, provider))
Completions: Completion credits entitlements
completions: completion_credits_check_entitlement_duration_p95
95th percentile completion credits entitlement check duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100300 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completion_credits_check_entitlement_duration_ms_bucket[5m])) by (le))
completions: completion_credits_consume_credits_duration_p95
95th percentile completion credits consume duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100301 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_completion_credits_consume_duration_ms_bucket[5m])) by (le))
completions: completion_credits_check_entitlement_durations
Completion credits entitlement check duration over 5m
- This metric tracks pre-completion-request latency for checking if completion credits entitlement has been exceeded.
- If this value is high, this latency may be noticeable to users.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/completions/completions?viewPanel=100310 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le) (rate(src_completion_credits_check_entitlement_duration_ms_bucket[5m]))
Periodic Goroutines
Overview of all periodic background routines across Sourcegraph services.
To see this dashboard, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines on your Sourcegraph instance.
Periodic Goroutines: Periodic Goroutines Overview
periodic-goroutines: total_running_goroutines
Total number of running periodic goroutines across all services
The total number of running periodic goroutines across all services. This provides a high-level overview of system activity.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_periodic_goroutine_running)
periodic-goroutines: goroutines_by_service
Number of running periodic goroutines by service
The number of running periodic goroutines broken down by service. This helps identify which services are running the most background routines.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (job) (src_periodic_goroutine_running)
periodic-goroutines: top_error_producers
Top 10 periodic goroutines by error rate
The top 10 periodic goroutines with the highest error rates. These routines may require immediate attention or investigation.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLtopk(10, sum by (name, job) (rate(src_periodic_goroutine_errors_total[5m])))
periodic-goroutines: top_time_consumers
Top 10 slowest periodic goroutines
The top 10 periodic goroutines with the longest average execution time. These routines may be candidates for optimization or load distribution.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLtopk(10, max by (name, job) (rate(src_periodic_goroutine_duration_seconds_sum[5m]) / rate(src_periodic_goroutine_duration_seconds_count[5m])))
Periodic Goroutines: Drill down
periodic-goroutines: filtered_success_rate
Success rate for selected goroutines
The rate of successful executions for the filtered periodic goroutines.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job) (rate(src_periodic_goroutine_total{name=~'${routineName:regex}', job=~'${serviceName:regex}'}[5m]))
periodic-goroutines: filtered_error_rate
Error rate for selected goroutines
The rate of errors for the filtered periodic goroutines.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job) (rate(src_periodic_goroutine_errors_total{name=~'${routineName:regex}', job=~'${serviceName:regex}'}[5m]))
periodic-goroutines: filtered_duration
95th percentile execution time for selected goroutines
The 95th percentile execution time for the filtered periodic goroutines.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job, le) (rate(src_periodic_goroutine_duration_seconds_bucket{name=~'${routineName:regex}', job=~'${serviceName:regex}'}[5m])))
periodic-goroutines: filtered_loop_time
95th percentile loop time for selected goroutines
The 95th percentile loop time for the filtered periodic goroutines.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job, le) (rate(src_periodic_goroutine_loop_duration_seconds_bucket{name=~'${routineName:regex}', job=~'${serviceName:regex}'}[5m])))
periodic-goroutines: filtered_tenant_count
Number of tenants processed by selected goroutines
Number of tenants processed by each selected periodic goroutine.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines?viewPanel=100120 on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name, job) (src_periodic_goroutine_tenant_count{name=~'${routineName:regex}', job=~'${serviceName:regex}'})
periodic-goroutines: filtered_tenant_duration
95th percentile tenant processing time for selected goroutines
The 95th percentile processing time for individual tenants.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines?viewPanel=100121 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by (name, job, le) (rate(src_periodic_goroutine_tenant_duration_seconds_bucket{name=~'${routineName:regex}', job=~'${serviceName:regex}'}[5m])))
periodic-goroutines: filtered_tenant_success_rate
Tenant success rate for selected goroutines
The rate of successful tenant processing operations.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines?viewPanel=100130 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job) (rate(src_periodic_goroutine_tenant_success_total{name=~'${routineName:regex}', job=~'${serviceName:regex}'}[5m]))
periodic-goroutines: filtered_tenant_error_rate
Tenant error rate for selected goroutines
The rate of tenant processing operations resulting in errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/periodic-goroutines/periodic-goroutines?viewPanel=100131 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (name, job) (rate(src_periodic_goroutine_tenant_errors_total{name=~'${routineName:regex}', job=~'${serviceName:regex}'}[5m]))
Background Jobs Dashboard
Overview of all background jobs in the system.
To see this dashboard, visit /-/debug/grafana/d/background-jobs/background-jobs on your Sourcegraph instance.
Background Jobs Dashboard: DBWorker Store Operations
background-jobs: operation_rates_by_method
Rate of operations by method (5m)
shows the rate of different dbworker store operations
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100000 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op) (rate(src_workerutil_dbworker_store_total{domain=~"$dbworker_domain"}[5m]))
background-jobs: error_rates
Rate of errors by method (5m)
Rate of errors by operation type. Check specific operations with high error rates.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100001 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op) (rate(src_workerutil_dbworker_store_errors_total{domain=~"$dbworker_domain"}[5m]))
background-jobs: p90_duration_by_method
90th percentile duration by method
90th percentile latency for dbworker store operations.
Investigate database query performance and indexing for the affected operations. Look for slow queries in database logs.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100010 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.9, sum by(le, op) (rate(src_workerutil_dbworker_store_duration_seconds_bucket{domain=~"$dbworker_domain"}[5m])))
background-jobs: p50_duration_by_method
Median duration by method
median latency for dbworker store operations
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100011 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.5, sum by(le, op) (rate(src_workerutil_dbworker_store_duration_seconds_bucket{domain=~"$dbworker_domain"}[5m])))
background-jobs: p90_duration_by_domain
90th percentile duration by domain
90th percentile latency for dbworker store operations.
Investigate database performance for the specific domain. May indicate issues with specific database tables or query patterns.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100012 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.9, sum by(le, domain) (rate(src_workerutil_dbworker_store_duration_seconds_bucket{domain=~"$dbworker_domain"}[5m])))
background-jobs: p50_duration_by_method
Median operation duration by method
median latency for dbworker store operations by method
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100013 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.5, sum by(le, op) (rate(src_workerutil_dbworker_store_duration_seconds_bucket{domain=~"$dbworker_domain"}[5m])))
background-jobs: dequeue_performance
Dequeue operation metrics
rate of dequeue operations by domain - critical for worker performance
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100020 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (domain) (rate(src_workerutil_dbworker_store_total{op="Dequeue", domain=~"$dbworker_domain"}[5m]))
background-jobs: error_percentage_by_method
Percentage of operations resulting in error by method
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100021 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum by (op) (rate(src_workerutil_dbworker_store_errors_total{domain=~"$dbworker_domain"}[5m])) / sum by (op) (rate(src_workerutil_dbworker_store_total{domain=~"$dbworker_domain"}[5m]))) * 100
background-jobs: error_percentage_by_domain
Percentage of operations resulting in error by domain
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100022 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum by (domain) (rate(src_workerutil_dbworker_store_errors_total{domain=~"$dbworker_domain"}[5m])) / sum by (domain) (rate(src_workerutil_dbworker_store_total{domain=~"$dbworker_domain"}[5m]))) * 100
background-jobs: operation_latency_heatmap
Distribution of operation durations
Distribution of operation durations - shows the spread of latencies across all operations
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100023 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le) (rate(src_workerutil_dbworker_store_duration_seconds_bucket{domain=~"$dbworker_domain"}[5m]))
Background Jobs Dashboard: DBWorker Resetter
background-jobs: resetter_duration
Time spent running the resetter
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100100 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum by(le, domain) (rate(src_dbworker_resetter_duration_seconds_bucket{domain=~"$resetter_domain"}[5m])))
background-jobs: resetter_runs
Number of times the resetter ran
the number of times the resetter ran in the last 5 minutes
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100101 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (domain) (increase(src_dbworker_resetter_total{domain=~"$resetter_domain"}[5m]))
background-jobs: resetter_failures
Number of times the resetter failed to run
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100102 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (domain) (increase(src_dbworker_resetter_errors_total{domain=~"$resetter_domain"}[5m]))
background-jobs: reset_records
Number of stalled records reset back to 'queued' state
the number of stalled records that were reset back to the queued state in the last 5 minutes
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100110 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (domain) (increase(src_dbworker_resetter_record_resets_total{domain=~"$resetter_domain"}[5m]))
background-jobs: failed_records
Number of stalled records marked as 'failed'
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100111 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (domain) (increase(src_dbworker_resetter_record_reset_failures_total{domain=~"$resetter_domain"}[5m]))
background-jobs: stall_duration
Duration jobs were stalled before being reset
median time a job was stalled before being reset
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100120 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le) (rate(src_dbworker_resetter_stall_duration_seconds_bucket{domain=~"$resetter_domain"}[5m]))
background-jobs: stall_duration_p90
90th percentile of stall duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100121 on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.9, sum by(le, domain) (rate(src_dbworker_resetter_stall_duration_seconds_bucket{domain=~"$resetter_domain"}[5m])))
background-jobs: reset_vs_failure_ratio
Ratio of jobs reset to queued versus marked as failed
ratio of reset jobs to failed jobs - higher values indicate healthier job processing
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100122 on your Sourcegraph instance.
Technical details
Query:
SHELL(sum by (domain) (increase(src_dbworker_resetter_record_resets_total{domain=~"$resetter_domain"}[1h]))) / on(domain) (sum by (domain) (increase(src_dbworker_resetter_record_reset_failures_total{domain=~"$resetter_domain"}[1h]) > 0) or on(domain) sum by (domain) (increase(src_dbworker_resetter_record_resets_total{domain=~"$resetter_domain"}[1h]) * 0 + 1))
Background Jobs Dashboard: Worker Queue Metrics
background-jobs: aggregate_queue_size
Total number of jobs queued across all domains
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100200 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(max by (domain) (src_workerutil_queue_depth))
background-jobs: max_queue_duration
Maximum time a job has been in queue across all domains
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100201 on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_workerutil_queue_duration_seconds)
background-jobs: queue_growth_rate
Rate of queue growth/decrease
Rate at which queue is growing. Positive values indicate more jobs are being added than processed.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100202 on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_queue_depth[30m]))/1800
background-jobs: queue_depth_by_domain
Number of jobs in queue by domain
Number of queued jobs per domain. Large values may indicate workers are not keeping up with incoming jobs.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100210 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (domain) (max by (domain) (src_workerutil_queue_depth))
background-jobs: queue_duration_by_domain
Maximum queue time by domain
Maximum time a job has been waiting in queue per domain. Long durations indicate potential worker stalls.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100211 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (domain) (max by (domain) (src_workerutil_queue_duration_seconds))
background-jobs: queue_growth_by_domain
Rate of change in queue size by domain
Rate of change in queue size per domain. Consistently positive values indicate jobs are being queued faster than processed.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/background-jobs/background-jobs?viewPanel=100212 on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (domain) (idelta(src_workerutil_queue_depth[10m])) / 600