k8s.io/kubernetes@v1.29.3/test/instrumentation/documentation/documentation-list.yaml (about) 1 - name: version_info 2 namespace: etcd 3 help: Etcd server's binary version 4 type: Gauge 5 stabilityLevel: ALPHA 6 labels: 7 - binary_version 8 - name: certificate_manager_client_ttl_seconds 9 subsystem: kubelet 10 help: Gauge of the TTL (time-to-live) of the Kubelet's client certificate. The value 11 is in seconds until certificate expiry (negative if already expired). If client 12 certificate is invalid or unused, the value will be +INF. 13 type: Gauge 14 stabilityLevel: ALPHA 15 - name: addresses_skipped_per_sync 16 subsystem: endpoint_slice_mirroring_controller 17 help: Number of addresses skipped on each Endpoints sync due to being invalid or 18 exceeding MaxEndpointsPerSubset 19 type: Histogram 20 stabilityLevel: ALPHA 21 buckets: 22 - 2 23 - 4 24 - 8 25 - 16 26 - 32 27 - 64 28 - 128 29 - 256 30 - 512 31 - 1024 32 - 2048 33 - 4096 34 - 8192 35 - 16384 36 - 32768 37 - name: changes 38 subsystem: endpoint_slice_mirroring_controller 39 help: Number of EndpointSlice changes 40 type: Counter 41 stabilityLevel: ALPHA 42 labels: 43 - operation 44 - name: desired_endpoint_slices 45 subsystem: endpoint_slice_mirroring_controller 46 help: Number of EndpointSlices that would exist with perfect endpoint allocation 47 type: Gauge 48 stabilityLevel: ALPHA 49 - name: endpoints_added_per_sync 50 subsystem: endpoint_slice_mirroring_controller 51 help: Number of endpoints added on each Endpoints sync 52 type: Histogram 53 stabilityLevel: ALPHA 54 buckets: 55 - 2 56 - 4 57 - 8 58 - 16 59 - 32 60 - 64 61 - 128 62 - 256 63 - 512 64 - 1024 65 - 2048 66 - 4096 67 - 8192 68 - 16384 69 - 32768 70 - name: endpoints_desired 71 subsystem: endpoint_slice_mirroring_controller 72 help: Number of endpoints desired 73 type: Gauge 74 stabilityLevel: ALPHA 75 - name: endpoints_removed_per_sync 76 subsystem: endpoint_slice_mirroring_controller 77 help: Number of endpoints removed on each Endpoints sync 78 type: Histogram 79 stabilityLevel: ALPHA 80 buckets: 81 - 2 82 - 4 83 - 8 84 - 16 85 - 32 86 - 64 87 - 128 88 - 256 89 - 512 90 - 1024 91 - 2048 92 - 4096 93 - 8192 94 - 16384 95 - 32768 96 - name: endpoints_sync_duration 97 subsystem: endpoint_slice_mirroring_controller 98 help: Duration of syncEndpoints() in seconds 99 type: Histogram 100 stabilityLevel: ALPHA 101 buckets: 102 - 0.001 103 - 0.002 104 - 0.004 105 - 0.008 106 - 0.016 107 - 0.032 108 - 0.064 109 - 0.128 110 - 0.256 111 - 0.512 112 - 1.024 113 - 2.048 114 - 4.096 115 - 8.192 116 - 16.384 117 - name: endpoints_updated_per_sync 118 subsystem: endpoint_slice_mirroring_controller 119 help: Number of endpoints updated on each Endpoints sync 120 type: Histogram 121 stabilityLevel: ALPHA 122 buckets: 123 - 2 124 - 4 125 - 8 126 - 16 127 - 32 128 - 64 129 - 128 130 - 256 131 - 512 132 - 1024 133 - 2048 134 - 4096 135 - 8192 136 - 16384 137 - 32768 138 - name: num_endpoint_slices 139 subsystem: endpoint_slice_mirroring_controller 140 help: Number of EndpointSlices 141 type: Gauge 142 stabilityLevel: ALPHA 143 - name: sync_duration_seconds 144 subsystem: root_ca_cert_publisher 145 help: Number of namespace syncs happened in root ca cert publisher. 146 type: Histogram 147 stabilityLevel: ALPHA 148 labels: 149 - code 150 buckets: 151 - 0.001 152 - 0.002 153 - 0.004 154 - 0.008 155 - 0.016 156 - 0.032 157 - 0.064 158 - 0.128 159 - 0.256 160 - 0.512 161 - 1.024 162 - 2.048 163 - 4.096 164 - 8.192 165 - 16.384 166 - name: sync_total 167 subsystem: root_ca_cert_publisher 168 help: Number of namespace syncs happened in root ca cert publisher. 169 type: Counter 170 stabilityLevel: ALPHA 171 labels: 172 - code 173 - name: job_creation_skew_duration_seconds 174 subsystem: cronjob_controller 175 help: Time between when a cronjob is scheduled to be run, and when the corresponding 176 job is created 177 type: Histogram 178 stabilityLevel: STABLE 179 buckets: 180 - 1 181 - 2 182 - 4 183 - 8 184 - 16 185 - 32 186 - 64 187 - 128 188 - 256 189 - 512 190 - name: resources_sync_error_total 191 subsystem: garbagecollector_controller 192 help: Number of garbage collector resources sync errors 193 type: Counter 194 stabilityLevel: ALPHA 195 - name: metric_computation_duration_seconds 196 subsystem: horizontal_pod_autoscaler_controller 197 help: The time(seconds) that the HPA controller takes to calculate one metric. The 198 label 'action' should be either 'scale_down', 'scale_up', or 'none'. The label 199 'error' should be either 'spec', 'internal', or 'none'. The label 'metric_type' 200 corresponds to HPA.spec.metrics[*].type 201 type: Histogram 202 stabilityLevel: ALPHA 203 labels: 204 - action 205 - error 206 - metric_type 207 buckets: 208 - 0.001 209 - 0.002 210 - 0.004 211 - 0.008 212 - 0.016 213 - 0.032 214 - 0.064 215 - 0.128 216 - 0.256 217 - 0.512 218 - 1.024 219 - 2.048 220 - 4.096 221 - 8.192 222 - 16.384 223 - name: metric_computation_total 224 subsystem: horizontal_pod_autoscaler_controller 225 help: Number of metric computations. The label 'action' should be either 'scale_down', 226 'scale_up', or 'none'. Also, the label 'error' should be either 'spec', 'internal', 227 or 'none'. The label 'metric_type' corresponds to HPA.spec.metrics[*].type 228 type: Counter 229 stabilityLevel: ALPHA 230 labels: 231 - action 232 - error 233 - metric_type 234 - name: reconciliation_duration_seconds 235 subsystem: horizontal_pod_autoscaler_controller 236 help: The time(seconds) that the HPA controller takes to reconcile once. The label 237 'action' should be either 'scale_down', 'scale_up', or 'none'. Also, the label 238 'error' should be either 'spec', 'internal', or 'none'. Note that if both spec 239 and internal errors happen during a reconciliation, the first one to occur is 240 reported in `error` label. 241 type: Histogram 242 stabilityLevel: ALPHA 243 labels: 244 - action 245 - error 246 buckets: 247 - 0.001 248 - 0.002 249 - 0.004 250 - 0.008 251 - 0.016 252 - 0.032 253 - 0.064 254 - 0.128 255 - 0.256 256 - 0.512 257 - 1.024 258 - 2.048 259 - 4.096 260 - 8.192 261 - 16.384 262 - name: reconciliations_total 263 subsystem: horizontal_pod_autoscaler_controller 264 help: Number of reconciliations of HPA controller. The label 'action' should be 265 either 'scale_down', 'scale_up', or 'none'. Also, the label 'error' should be 266 either 'spec', 'internal', or 'none'. Note that if both spec and internal errors 267 happen during a reconciliation, the first one to occur is reported in `error` 268 label. 269 type: Counter 270 stabilityLevel: ALPHA 271 labels: 272 - action 273 - error 274 - name: pod_failures_handled_by_failure_policy_total 275 subsystem: job_controller 276 help: "`The number of failed Pods handled by failure policy with\n\t\t\trespect 277 to the failure policy action applied based on the matched\n\t\t\trule. Possible 278 values of the action label correspond to the\n\t\t\tpossible values for the failure 279 policy rule action, which are:\n\t\t\t\"FailJob\", \"Ignore\" and \"Count\".`" 280 type: Counter 281 stabilityLevel: ALPHA 282 labels: 283 - action 284 - name: terminated_pods_tracking_finalizer_total 285 subsystem: job_controller 286 help: |- 287 `The number of terminated pods (phase=Failed|Succeeded) 288 that have the finalizer batch.kubernetes.io/job-tracking 289 The event label can be "add" or "delete".` 290 type: Counter 291 stabilityLevel: ALPHA 292 labels: 293 - event 294 - name: unhealthy_nodes_in_zone 295 subsystem: node_collector 296 help: Gauge measuring number of not Ready Nodes per zones. 297 type: Gauge 298 stabilityLevel: ALPHA 299 labels: 300 - zone 301 - name: update_all_nodes_health_duration_seconds 302 subsystem: node_collector 303 help: Duration in seconds for NodeController to update the health of all nodes. 304 type: Histogram 305 stabilityLevel: ALPHA 306 buckets: 307 - 0.01 308 - 0.04 309 - 0.16 310 - 0.64 311 - 2.56 312 - 10.24 313 - 40.96 314 - 163.84 315 - name: update_node_health_duration_seconds 316 subsystem: node_collector 317 help: Duration in seconds for NodeController to update the health of a single node. 318 type: Histogram 319 stabilityLevel: ALPHA 320 buckets: 321 - 0.001 322 - 0.004 323 - 0.016 324 - 0.064 325 - 0.256 326 - 1.024 327 - 4.096 328 - 16.384 329 - name: zone_health 330 subsystem: node_collector 331 help: Gauge measuring percentage of healthy nodes per zone. 332 type: Gauge 333 stabilityLevel: ALPHA 334 labels: 335 - zone 336 - name: zone_size 337 subsystem: node_collector 338 help: Gauge measuring number of registered Nodes per zones. 339 type: Gauge 340 stabilityLevel: ALPHA 341 labels: 342 - zone 343 - name: cidrset_allocation_tries_per_request 344 subsystem: node_ipam_controller 345 help: Number of endpoints added on each Service sync 346 type: Histogram 347 stabilityLevel: ALPHA 348 labels: 349 - clusterCIDR 350 buckets: 351 - 1 352 - 5 353 - 25 354 - 125 355 - 625 356 - name: cidrset_cidrs_allocations_total 357 subsystem: node_ipam_controller 358 help: Counter measuring total number of CIDR allocations. 359 type: Counter 360 stabilityLevel: ALPHA 361 labels: 362 - clusterCIDR 363 - name: cidrset_cidrs_releases_total 364 subsystem: node_ipam_controller 365 help: Counter measuring total number of CIDR releases. 366 type: Counter 367 stabilityLevel: ALPHA 368 labels: 369 - clusterCIDR 370 - name: cidrset_usage_cidrs 371 subsystem: node_ipam_controller 372 help: Gauge measuring percentage of allocated CIDRs. 373 type: Gauge 374 stabilityLevel: ALPHA 375 labels: 376 - clusterCIDR 377 - name: cirdset_max_cidrs 378 subsystem: node_ipam_controller 379 help: Maximum number of CIDRs that can be allocated. 380 type: Gauge 381 stabilityLevel: ALPHA 382 labels: 383 - clusterCIDR 384 - name: multicidrset_allocation_tries_per_request 385 subsystem: node_ipam_controller 386 help: Histogram measuring CIDR allocation tries per request. 387 type: Histogram 388 stabilityLevel: ALPHA 389 labels: 390 - clusterCIDR 391 buckets: 392 - 1 393 - 5 394 - 25 395 - 125 396 - 625 397 - name: multicidrset_cidrs_allocations_total 398 subsystem: node_ipam_controller 399 help: Counter measuring total number of CIDR allocations. 400 type: Counter 401 stabilityLevel: ALPHA 402 labels: 403 - clusterCIDR 404 - name: multicidrset_cidrs_releases_total 405 subsystem: node_ipam_controller 406 help: Counter measuring total number of CIDR releases. 407 type: Counter 408 stabilityLevel: ALPHA 409 labels: 410 - clusterCIDR 411 - name: multicidrset_usage_cidrs 412 subsystem: node_ipam_controller 413 help: Gauge measuring percentage of allocated CIDRs. 414 type: Gauge 415 stabilityLevel: ALPHA 416 labels: 417 - clusterCIDR 418 - name: multicirdset_max_cidrs 419 subsystem: node_ipam_controller 420 help: Maximum number of CIDRs that can be allocated. 421 type: Gauge 422 stabilityLevel: ALPHA 423 labels: 424 - clusterCIDR 425 - name: force_delete_pod_errors_total 426 subsystem: pod_gc_collector 427 help: Number of errors encountered when forcefully deleting the pods since the Pod 428 GC Controller started. 429 type: Counter 430 stabilityLevel: ALPHA 431 labels: 432 - namespace 433 - reason 434 - name: force_delete_pods_total 435 subsystem: pod_gc_collector 436 help: Number of pods that are being forcefully deleted since the Pod GC Controller 437 started. 438 type: Counter 439 stabilityLevel: ALPHA 440 labels: 441 - namespace 442 - reason 443 - name: sorting_deletion_age_ratio 444 subsystem: replicaset_controller 445 help: The ratio of chosen deleted pod's ages to the current youngest pod's age (at 446 the time). Should be <2.The intent of this metric is to measure the rough efficacy 447 of the LogarithmicScaleDown feature gate's effect onthe sorting (and deletion) 448 of pods when a replicaset scales down. This only considers Ready pods when calculating 449 and reporting. 450 type: Histogram 451 stabilityLevel: ALPHA 452 buckets: 453 - 0.25 454 - 0.5 455 - 1 456 - 2 457 - 4 458 - 8 459 - name: create_attempts_total 460 subsystem: resourceclaim_controller 461 help: Number of ResourceClaims creation requests 462 type: Counter 463 stabilityLevel: ALPHA 464 - name: create_failures_total 465 subsystem: resourceclaim_controller 466 help: Number of ResourceClaims creation request failures 467 type: Counter 468 stabilityLevel: ALPHA 469 - name: job_deletion_duration_seconds 470 subsystem: ttl_after_finished_controller 471 help: The time it took to delete the job since it became eligible for deletion 472 type: Histogram 473 stabilityLevel: ALPHA 474 buckets: 475 - 0.1 476 - 0.2 477 - 0.4 478 - 0.8 479 - 1.6 480 - 3.2 481 - 6.4 482 - 12.8 483 - 25.6 484 - 51.2 485 - 102.4 486 - 204.8 487 - 409.6 488 - 819.2 489 - name: job_pods_finished_total 490 subsystem: job_controller 491 help: The number of finished Pods that are fully tracked 492 type: Counter 493 stabilityLevel: STABLE 494 labels: 495 - completion_mode 496 - result 497 - name: job_sync_duration_seconds 498 subsystem: job_controller 499 help: The time it took to sync a job 500 type: Histogram 501 stabilityLevel: STABLE 502 labels: 503 - action 504 - completion_mode 505 - result 506 buckets: 507 - 0.001 508 - 0.002 509 - 0.004 510 - 0.008 511 - 0.016 512 - 0.032 513 - 0.064 514 - 0.128 515 - 0.256 516 - 0.512 517 - 1.024 518 - 2.048 519 - 4.096 520 - 8.192 521 - 16.384 522 - name: job_syncs_total 523 subsystem: job_controller 524 help: The number of job syncs 525 type: Counter 526 stabilityLevel: STABLE 527 labels: 528 - action 529 - completion_mode 530 - result 531 - name: jobs_finished_total 532 subsystem: job_controller 533 help: The number of finished jobs 534 type: Counter 535 stabilityLevel: STABLE 536 labels: 537 - completion_mode 538 - reason 539 - result 540 - name: evictions_total 541 subsystem: node_collector 542 help: Number of Node evictions that happened since current instance of NodeController 543 started. 544 type: Counter 545 stabilityLevel: STABLE 546 labels: 547 - zone 548 - name: attachdetach_controller_forced_detaches 549 subsystem: attach_detach_controller 550 help: Number of times the A/D Controller performed a forced detach 551 type: Counter 552 stabilityLevel: ALPHA 553 labels: 554 - reason 555 - name: attachdetach_controller_total_volumes 556 help: Number of volumes in A/D Controller 557 type: Custom 558 stabilityLevel: ALPHA 559 labels: 560 - plugin_name 561 - state 562 - name: create_failures_total 563 subsystem: ephemeral_volume_controller 564 help: Number of PersistenVolumeClaims creation requests 565 type: Counter 566 stabilityLevel: ALPHA 567 - name: create_total 568 subsystem: ephemeral_volume_controller 569 help: Number of PersistenVolumeClaims creation requests 570 type: Counter 571 stabilityLevel: ALPHA 572 - name: client_expiration_renew_errors 573 subsystem: certificate_manager 574 namespace: kubelet 575 help: Counter of certificate renewal errors. 576 type: Counter 577 stabilityLevel: ALPHA 578 - name: certificate_manager_server_rotation_seconds 579 subsystem: kubelet 580 help: Histogram of the number of seconds the previous certificate lived before being 581 rotated. 582 type: Histogram 583 stabilityLevel: ALPHA 584 buckets: 585 - 60 586 - 3600 587 - 14400 588 - 86400 589 - 604800 590 - 2.592e+06 591 - 7.776e+06 592 - 1.5552e+07 593 - 3.1104e+07 594 - 1.24416e+08 595 - name: certificate_manager_server_ttl_seconds 596 subsystem: kubelet 597 help: Gauge of the shortest TTL (time-to-live) of the Kubelet's serving certificate. 598 The value is in seconds until certificate expiry (negative if already expired). 599 If serving certificate is invalid or unused, the value will be +INF. 600 type: Gauge 601 stabilityLevel: ALPHA 602 - name: credential_provider_plugin_duration 603 subsystem: kubelet 604 help: Duration of execution in seconds for credential provider plugin 605 type: Histogram 606 stabilityLevel: ALPHA 607 labels: 608 - plugin_name 609 buckets: 610 - 0.005 611 - 0.01 612 - 0.025 613 - 0.05 614 - 0.1 615 - 0.25 616 - 0.5 617 - 1 618 - 2.5 619 - 5 620 - 10 621 - name: credential_provider_plugin_errors 622 subsystem: kubelet 623 help: Number of errors from credential provider plugin 624 type: Counter 625 stabilityLevel: ALPHA 626 labels: 627 - plugin_name 628 - name: server_expiration_renew_errors 629 subsystem: kubelet 630 help: Counter of certificate renewal errors. 631 type: Counter 632 stabilityLevel: ALPHA 633 - name: pv_collector_bound_pv_count 634 help: Gauge measuring number of persistent volume currently bound 635 type: Custom 636 stabilityLevel: ALPHA 637 labels: 638 - storage_class 639 - name: pv_collector_bound_pvc_count 640 help: Gauge measuring number of persistent volume claim currently bound 641 type: Custom 642 stabilityLevel: ALPHA 643 labels: 644 - namespace 645 - name: pv_collector_total_pv_count 646 help: Gauge measuring total number of persistent volumes 647 type: Custom 648 stabilityLevel: ALPHA 649 labels: 650 - plugin_name 651 - volume_mode 652 - name: pv_collector_unbound_pv_count 653 help: Gauge measuring number of persistent volume currently unbound 654 type: Custom 655 stabilityLevel: ALPHA 656 labels: 657 - storage_class 658 - name: pv_collector_unbound_pvc_count 659 help: Gauge measuring number of persistent volume claim currently unbound 660 type: Custom 661 stabilityLevel: ALPHA 662 labels: 663 - namespace 664 - name: retroactive_storageclass_errors_total 665 help: Total number of failed retroactive StorageClass assignments to persistent 666 volume claim 667 type: Counter 668 stabilityLevel: ALPHA 669 - name: retroactive_storageclass_total 670 help: Total number of retroactive StorageClass assignments to persistent volume 671 claim 672 type: Counter 673 stabilityLevel: ALPHA 674 - name: storage_count_attachable_volumes_in_use 675 help: Measure number of volumes in use 676 type: Custom 677 stabilityLevel: ALPHA 678 labels: 679 - node 680 - volume_plugin 681 - name: volume_operation_total_errors 682 help: Total volume operation errors 683 type: Counter 684 stabilityLevel: ALPHA 685 labels: 686 - operation_name 687 - plugin_name 688 - name: container_swap_usage_bytes 689 help: Current amount of the container swap usage in bytes. Reported only on non-windows 690 systems 691 type: Custom 692 stabilityLevel: ALPHA 693 labels: 694 - container 695 - pod 696 - namespace 697 - name: active_pods 698 subsystem: kubelet 699 help: The number of pods the kubelet considers active and which are being considered 700 when admitting new pods. static is true if the pod is not from the apiserver. 701 type: Gauge 702 stabilityLevel: ALPHA 703 labels: 704 - static 705 - name: cgroup_manager_duration_seconds 706 subsystem: kubelet 707 help: Duration in seconds for cgroup manager operations. Broken down by method. 708 type: Histogram 709 stabilityLevel: ALPHA 710 labels: 711 - operation_type 712 buckets: 713 - 0.005 714 - 0.01 715 - 0.025 716 - 0.05 717 - 0.1 718 - 0.25 719 - 0.5 720 - 1 721 - 2.5 722 - 5 723 - 10 724 - name: kubelet_container_log_filesystem_used_bytes 725 help: Bytes used by the container's logs on the filesystem. 726 type: Custom 727 stabilityLevel: ALPHA 728 labels: 729 - uid 730 - namespace 731 - pod 732 - container 733 - name: containers_per_pod_count 734 subsystem: kubelet 735 help: The number of containers per pod. 736 type: Histogram 737 stabilityLevel: ALPHA 738 buckets: 739 - 1 740 - 2 741 - 4 742 - 8 743 - 16 744 - name: cpu_manager_pinning_errors_total 745 subsystem: kubelet 746 help: The number of cpu core allocations which required pinning failed. 747 type: Counter 748 stabilityLevel: ALPHA 749 - name: cpu_manager_pinning_requests_total 750 subsystem: kubelet 751 help: The number of cpu core allocations which required pinning. 752 type: Counter 753 stabilityLevel: ALPHA 754 - name: desired_pods 755 subsystem: kubelet 756 help: The number of pods the kubelet is being instructed to run. static is true 757 if the pod is not from the apiserver. 758 type: Gauge 759 stabilityLevel: ALPHA 760 labels: 761 - static 762 - name: device_plugin_alloc_duration_seconds 763 subsystem: kubelet 764 help: Duration in seconds to serve a device plugin Allocation request. Broken down 765 by resource name. 766 type: Histogram 767 stabilityLevel: ALPHA 768 labels: 769 - resource_name 770 buckets: 771 - 0.005 772 - 0.01 773 - 0.025 774 - 0.05 775 - 0.1 776 - 0.25 777 - 0.5 778 - 1 779 - 2.5 780 - 5 781 - 10 782 - name: device_plugin_registration_total 783 subsystem: kubelet 784 help: Cumulative number of device plugin registrations. Broken down by resource 785 name. 786 type: Counter 787 stabilityLevel: ALPHA 788 labels: 789 - resource_name 790 - name: evented_pleg_connection_error_count 791 subsystem: kubelet 792 help: The number of errors encountered during the establishment of streaming connection 793 with the CRI runtime. 794 type: Counter 795 stabilityLevel: ALPHA 796 - name: evented_pleg_connection_latency_seconds 797 subsystem: kubelet 798 help: The latency of streaming connection with the CRI runtime, measured in seconds. 799 type: Histogram 800 stabilityLevel: ALPHA 801 buckets: 802 - 0.005 803 - 0.01 804 - 0.025 805 - 0.05 806 - 0.1 807 - 0.25 808 - 0.5 809 - 1 810 - 2.5 811 - 5 812 - 10 813 - name: evented_pleg_connection_success_count 814 subsystem: kubelet 815 help: The number of times a streaming client was obtained to receive CRI Events. 816 type: Counter 817 stabilityLevel: ALPHA 818 - name: eviction_stats_age_seconds 819 subsystem: kubelet 820 help: Time between when stats are collected, and when pod is evicted based on those 821 stats by eviction signal 822 type: Histogram 823 stabilityLevel: ALPHA 824 labels: 825 - eviction_signal 826 buckets: 827 - 0.005 828 - 0.01 829 - 0.025 830 - 0.05 831 - 0.1 832 - 0.25 833 - 0.5 834 - 1 835 - 2.5 836 - 5 837 - 10 838 - name: evictions 839 subsystem: kubelet 840 help: Cumulative number of pod evictions by eviction signal 841 type: Counter 842 stabilityLevel: ALPHA 843 labels: 844 - eviction_signal 845 - name: graceful_shutdown_end_time_seconds 846 subsystem: kubelet 847 help: Last graceful shutdown start time since unix epoch in seconds 848 type: Gauge 849 stabilityLevel: ALPHA 850 - name: graceful_shutdown_start_time_seconds 851 subsystem: kubelet 852 help: Last graceful shutdown start time since unix epoch in seconds 853 type: Gauge 854 stabilityLevel: ALPHA 855 - name: lifecycle_handler_http_fallbacks_total 856 subsystem: kubelet 857 help: The number of times lifecycle handlers successfully fell back to http from 858 https. 859 type: Counter 860 stabilityLevel: ALPHA 861 - name: managed_ephemeral_containers 862 subsystem: kubelet 863 help: Current number of ephemeral containers in pods managed by this kubelet. 864 type: Gauge 865 stabilityLevel: ALPHA 866 - name: mirror_pods 867 subsystem: kubelet 868 help: The number of mirror pods the kubelet will try to create (one per admitted 869 static pod) 870 type: Gauge 871 stabilityLevel: ALPHA 872 - name: node_name 873 subsystem: kubelet 874 help: The node's name. The count is always 1. 875 type: Gauge 876 stabilityLevel: ALPHA 877 labels: 878 - node 879 - name: orphan_pod_cleaned_volumes 880 subsystem: kubelet 881 help: The total number of orphaned Pods whose volumes were cleaned in the last periodic 882 sweep. 883 type: Gauge 884 stabilityLevel: ALPHA 885 - name: orphan_pod_cleaned_volumes_errors 886 subsystem: kubelet 887 help: The number of orphaned Pods whose volumes failed to be cleaned in the last 888 periodic sweep. 889 type: Gauge 890 stabilityLevel: ALPHA 891 - name: orphaned_runtime_pods_total 892 subsystem: kubelet 893 help: Number of pods that have been detected in the container runtime without being 894 already known to the pod worker. This typically indicates the kubelet was restarted 895 while a pod was force deleted in the API or in the local configuration, which 896 is unusual. 897 type: Counter 898 stabilityLevel: ALPHA 899 - name: pleg_discard_events 900 subsystem: kubelet 901 help: The number of discard events in PLEG. 902 type: Counter 903 stabilityLevel: ALPHA 904 - name: pleg_last_seen_seconds 905 subsystem: kubelet 906 help: Timestamp in seconds when PLEG was last seen active. 907 type: Gauge 908 stabilityLevel: ALPHA 909 - name: pleg_relist_duration_seconds 910 subsystem: kubelet 911 help: Duration in seconds for relisting pods in PLEG. 912 type: Histogram 913 stabilityLevel: ALPHA 914 buckets: 915 - 0.005 916 - 0.01 917 - 0.025 918 - 0.05 919 - 0.1 920 - 0.25 921 - 0.5 922 - 1 923 - 2.5 924 - 5 925 - 10 926 - name: pleg_relist_interval_seconds 927 subsystem: kubelet 928 help: Interval in seconds between relisting in PLEG. 929 type: Histogram 930 stabilityLevel: ALPHA 931 buckets: 932 - 0.005 933 - 0.01 934 - 0.025 935 - 0.05 936 - 0.1 937 - 0.25 938 - 0.5 939 - 1 940 - 2.5 941 - 5 942 - 10 943 - name: pod_resources_endpoint_errors_get 944 subsystem: kubelet 945 help: Number of requests to the PodResource Get endpoint which returned error. Broken 946 down by server api version. 947 type: Counter 948 stabilityLevel: ALPHA 949 labels: 950 - server_api_version 951 - name: pod_resources_endpoint_errors_get_allocatable 952 subsystem: kubelet 953 help: Number of requests to the PodResource GetAllocatableResources endpoint which 954 returned error. Broken down by server api version. 955 type: Counter 956 stabilityLevel: ALPHA 957 labels: 958 - server_api_version 959 - name: pod_resources_endpoint_errors_list 960 subsystem: kubelet 961 help: Number of requests to the PodResource List endpoint which returned error. 962 Broken down by server api version. 963 type: Counter 964 stabilityLevel: ALPHA 965 labels: 966 - server_api_version 967 - name: pod_resources_endpoint_requests_get 968 subsystem: kubelet 969 help: Number of requests to the PodResource Get endpoint. Broken down by server 970 api version. 971 type: Counter 972 stabilityLevel: ALPHA 973 labels: 974 - server_api_version 975 - name: pod_resources_endpoint_requests_get_allocatable 976 subsystem: kubelet 977 help: Number of requests to the PodResource GetAllocatableResources endpoint. Broken 978 down by server api version. 979 type: Counter 980 stabilityLevel: ALPHA 981 labels: 982 - server_api_version 983 - name: pod_resources_endpoint_requests_list 984 subsystem: kubelet 985 help: Number of requests to the PodResource List endpoint. Broken down by server 986 api version. 987 type: Counter 988 stabilityLevel: ALPHA 989 labels: 990 - server_api_version 991 - name: pod_resources_endpoint_requests_total 992 subsystem: kubelet 993 help: Cumulative number of requests to the PodResource endpoint. Broken down by 994 server api version. 995 type: Counter 996 stabilityLevel: ALPHA 997 labels: 998 - server_api_version 999 - name: pod_start_duration_seconds 1000 subsystem: kubelet 1001 help: Duration in seconds from kubelet seeing a pod for the first time to the pod 1002 starting to run 1003 type: Histogram 1004 stabilityLevel: ALPHA 1005 buckets: 1006 - 0.005 1007 - 0.01 1008 - 0.025 1009 - 0.05 1010 - 0.1 1011 - 0.25 1012 - 0.5 1013 - 1 1014 - 2.5 1015 - 5 1016 - 10 1017 - name: pod_start_sli_duration_seconds 1018 subsystem: kubelet 1019 help: Duration in seconds to start a pod, excluding time to pull images and run 1020 init containers, measured from pod creation timestamp to when all its containers 1021 are reported as started and observed via watch 1022 type: Histogram 1023 stabilityLevel: ALPHA 1024 buckets: 1025 - 0.5 1026 - 1 1027 - 2 1028 - 3 1029 - 4 1030 - 5 1031 - 6 1032 - 8 1033 - 10 1034 - 20 1035 - 30 1036 - 45 1037 - 60 1038 - 120 1039 - 180 1040 - 240 1041 - 300 1042 - 360 1043 - 480 1044 - 600 1045 - 900 1046 - 1200 1047 - 1800 1048 - 2700 1049 - 3600 1050 - name: pod_status_sync_duration_seconds 1051 subsystem: kubelet 1052 help: Duration in seconds to sync a pod status update. Measures time from detection 1053 of a change to pod status until the API is successfully updated for that pod, 1054 even if multiple intevening changes to pod status occur. 1055 type: Histogram 1056 stabilityLevel: ALPHA 1057 buckets: 1058 - 0.01 1059 - 0.05 1060 - 0.1 1061 - 0.5 1062 - 1 1063 - 5 1064 - 10 1065 - 20 1066 - 30 1067 - 45 1068 - 60 1069 - name: pod_worker_duration_seconds 1070 subsystem: kubelet 1071 help: 'Duration in seconds to sync a single pod. Broken down by operation type: 1072 create, update, or sync' 1073 type: Histogram 1074 stabilityLevel: ALPHA 1075 labels: 1076 - operation_type 1077 buckets: 1078 - 0.005 1079 - 0.01 1080 - 0.025 1081 - 0.05 1082 - 0.1 1083 - 0.25 1084 - 0.5 1085 - 1 1086 - 2.5 1087 - 5 1088 - 10 1089 - name: pod_worker_start_duration_seconds 1090 subsystem: kubelet 1091 help: Duration in seconds from kubelet seeing a pod to starting a worker. 1092 type: Histogram 1093 stabilityLevel: ALPHA 1094 buckets: 1095 - 0.005 1096 - 0.01 1097 - 0.025 1098 - 0.05 1099 - 0.1 1100 - 0.25 1101 - 0.5 1102 - 1 1103 - 2.5 1104 - 5 1105 - 10 1106 - name: preemptions 1107 subsystem: kubelet 1108 help: Cumulative number of pod preemptions by preemption resource 1109 type: Counter 1110 stabilityLevel: ALPHA 1111 labels: 1112 - preemption_signal 1113 - name: restarted_pods_total 1114 subsystem: kubelet 1115 help: Number of pods that have been restarted because they were deleted and recreated 1116 with the same UID while the kubelet was watching them (common for static pods, 1117 extremely uncommon for API pods) 1118 type: Counter 1119 stabilityLevel: ALPHA 1120 labels: 1121 - static 1122 - name: run_podsandbox_duration_seconds 1123 subsystem: kubelet 1124 help: Duration in seconds of the run_podsandbox operations. Broken down by RuntimeClass.Handler. 1125 type: Histogram 1126 stabilityLevel: ALPHA 1127 labels: 1128 - runtime_handler 1129 buckets: 1130 - 0.005 1131 - 0.01 1132 - 0.025 1133 - 0.05 1134 - 0.1 1135 - 0.25 1136 - 0.5 1137 - 1 1138 - 2.5 1139 - 5 1140 - 10 1141 - name: run_podsandbox_errors_total 1142 subsystem: kubelet 1143 help: Cumulative number of the run_podsandbox operation errors by RuntimeClass.Handler. 1144 type: Counter 1145 stabilityLevel: ALPHA 1146 labels: 1147 - runtime_handler 1148 - name: running_containers 1149 subsystem: kubelet 1150 help: Number of containers currently running 1151 type: Gauge 1152 stabilityLevel: ALPHA 1153 labels: 1154 - container_state 1155 - name: running_pods 1156 subsystem: kubelet 1157 help: Number of pods that have a running pod sandbox 1158 type: Gauge 1159 stabilityLevel: ALPHA 1160 - name: runtime_operations_duration_seconds 1161 subsystem: kubelet 1162 help: Duration in seconds of runtime operations. Broken down by operation type. 1163 type: Histogram 1164 stabilityLevel: ALPHA 1165 labels: 1166 - operation_type 1167 buckets: 1168 - 0.005 1169 - 0.0125 1170 - 0.03125 1171 - 0.078125 1172 - 0.1953125 1173 - 0.48828125 1174 - 1.220703125 1175 - 3.0517578125 1176 - 7.62939453125 1177 - 19.073486328125 1178 - 47.6837158203125 1179 - 119.20928955078125 1180 - 298.0232238769531 1181 - 745.0580596923828 1182 - name: runtime_operations_errors_total 1183 subsystem: kubelet 1184 help: Cumulative number of runtime operation errors by operation type. 1185 type: Counter 1186 stabilityLevel: ALPHA 1187 labels: 1188 - operation_type 1189 - name: runtime_operations_total 1190 subsystem: kubelet 1191 help: Cumulative number of runtime operations by operation type. 1192 type: Counter 1193 stabilityLevel: ALPHA 1194 labels: 1195 - operation_type 1196 - name: started_containers_errors_total 1197 subsystem: kubelet 1198 help: Cumulative number of errors when starting containers 1199 type: Counter 1200 stabilityLevel: ALPHA 1201 labels: 1202 - code 1203 - container_type 1204 - name: started_containers_total 1205 subsystem: kubelet 1206 help: Cumulative number of containers started 1207 type: Counter 1208 stabilityLevel: ALPHA 1209 labels: 1210 - container_type 1211 - name: started_host_process_containers_errors_total 1212 subsystem: kubelet 1213 help: Cumulative number of errors when starting hostprocess containers. This metric 1214 will only be collected on Windows. 1215 type: Counter 1216 stabilityLevel: ALPHA 1217 labels: 1218 - code 1219 - container_type 1220 - name: started_host_process_containers_total 1221 subsystem: kubelet 1222 help: Cumulative number of hostprocess containers started. This metric will only 1223 be collected on Windows. 1224 type: Counter 1225 stabilityLevel: ALPHA 1226 labels: 1227 - container_type 1228 - name: started_pods_errors_total 1229 subsystem: kubelet 1230 help: Cumulative number of errors when starting pods 1231 type: Counter 1232 stabilityLevel: ALPHA 1233 - name: started_pods_total 1234 subsystem: kubelet 1235 help: Cumulative number of pods started 1236 type: Counter 1237 stabilityLevel: ALPHA 1238 - name: topology_manager_admission_duration_ms 1239 subsystem: kubelet 1240 help: Duration in milliseconds to serve a pod admission request. 1241 type: Histogram 1242 stabilityLevel: ALPHA 1243 buckets: 1244 - 0.05 1245 - 0.1 1246 - 0.2 1247 - 0.4 1248 - 0.8 1249 - 1.6 1250 - 3.2 1251 - 6.4 1252 - 12.8 1253 - 25.6 1254 - 51.2 1255 - 102.4 1256 - 204.8 1257 - 409.6 1258 - 819.2 1259 - name: topology_manager_admission_errors_total 1260 subsystem: kubelet 1261 help: The number of admission request failures where resources could not be aligned. 1262 type: Counter 1263 stabilityLevel: ALPHA 1264 - name: topology_manager_admission_requests_total 1265 subsystem: kubelet 1266 help: The number of admission requests where resources have to be aligned. 1267 type: Counter 1268 stabilityLevel: ALPHA 1269 - name: kubelet_volume_stats_available_bytes 1270 help: Number of available bytes in the volume 1271 type: Custom 1272 stabilityLevel: ALPHA 1273 labels: 1274 - namespace 1275 - persistentvolumeclaim 1276 - name: kubelet_volume_stats_capacity_bytes 1277 help: Capacity in bytes of the volume 1278 type: Custom 1279 stabilityLevel: ALPHA 1280 labels: 1281 - namespace 1282 - persistentvolumeclaim 1283 - name: kubelet_volume_stats_health_status_abnormal 1284 help: Abnormal volume health status. The count is either 1 or 0. 1 indicates the 1285 volume is unhealthy, 0 indicates volume is healthy 1286 type: Custom 1287 stabilityLevel: ALPHA 1288 labels: 1289 - namespace 1290 - persistentvolumeclaim 1291 - name: kubelet_volume_stats_inodes 1292 help: Maximum number of inodes in the volume 1293 type: Custom 1294 stabilityLevel: ALPHA 1295 labels: 1296 - namespace 1297 - persistentvolumeclaim 1298 - name: kubelet_volume_stats_inodes_free 1299 help: Number of free inodes in the volume 1300 type: Custom 1301 stabilityLevel: ALPHA 1302 labels: 1303 - namespace 1304 - persistentvolumeclaim 1305 - name: kubelet_volume_stats_inodes_used 1306 help: Number of used inodes in the volume 1307 type: Custom 1308 stabilityLevel: ALPHA 1309 labels: 1310 - namespace 1311 - persistentvolumeclaim 1312 - name: kubelet_volume_stats_used_bytes 1313 help: Number of used bytes in the volume 1314 type: Custom 1315 stabilityLevel: ALPHA 1316 labels: 1317 - namespace 1318 - persistentvolumeclaim 1319 - name: working_pods 1320 subsystem: kubelet 1321 help: Number of pods the kubelet is actually running, broken down by lifecycle phase, 1322 whether the pod is desired, orphaned, or runtime only (also orphaned), and whether 1323 the pod is static. An orphaned pod has been removed from local configuration or 1324 force deleted in the API and consumes resources that are not otherwise visible. 1325 type: Gauge 1326 stabilityLevel: ALPHA 1327 labels: 1328 - config 1329 - lifecycle 1330 - static 1331 - name: node_swap_usage_bytes 1332 help: Current swap usage of the node in bytes. Reported only on non-windows systems 1333 type: Custom 1334 stabilityLevel: ALPHA 1335 - name: pod_swap_usage_bytes 1336 help: Current amount of the pod swap usage in bytes. Reported only on non-windows 1337 systems 1338 type: Custom 1339 stabilityLevel: ALPHA 1340 labels: 1341 - pod 1342 - namespace 1343 - name: scrape_error 1344 help: 1 if there was an error while getting container metrics, 0 otherwise 1345 type: Custom 1346 deprecatedVersion: 1.29.0 1347 stabilityLevel: ALPHA 1348 - name: container_cpu_usage_seconds_total 1349 help: Cumulative cpu time consumed by the container in core-seconds 1350 type: Custom 1351 stabilityLevel: STABLE 1352 labels: 1353 - container 1354 - pod 1355 - namespace 1356 - name: container_memory_working_set_bytes 1357 help: Current working set of the container in bytes 1358 type: Custom 1359 stabilityLevel: STABLE 1360 labels: 1361 - container 1362 - pod 1363 - namespace 1364 - name: container_start_time_seconds 1365 help: Start time of the container since unix epoch in seconds 1366 type: Custom 1367 stabilityLevel: STABLE 1368 labels: 1369 - container 1370 - pod 1371 - namespace 1372 - name: node_cpu_usage_seconds_total 1373 help: Cumulative cpu time consumed by the node in core-seconds 1374 type: Custom 1375 stabilityLevel: STABLE 1376 - name: node_memory_working_set_bytes 1377 help: Current working set of the node in bytes 1378 type: Custom 1379 stabilityLevel: STABLE 1380 - name: pod_cpu_usage_seconds_total 1381 help: Cumulative cpu time consumed by the pod in core-seconds 1382 type: Custom 1383 stabilityLevel: STABLE 1384 labels: 1385 - pod 1386 - namespace 1387 - name: pod_memory_working_set_bytes 1388 help: Current working set of the pod in bytes 1389 type: Custom 1390 stabilityLevel: STABLE 1391 labels: 1392 - pod 1393 - namespace 1394 - name: resource_scrape_error 1395 help: 1 if there was an error while getting container metrics, 0 otherwise 1396 type: Custom 1397 stabilityLevel: STABLE 1398 - name: force_cleaned_failed_volume_operation_errors_total 1399 help: The number of volumes that failed force cleanup after their reconstruction 1400 failed during kubelet startup. 1401 type: Counter 1402 stabilityLevel: ALPHA 1403 - name: force_cleaned_failed_volume_operations_total 1404 help: The number of volumes that were force cleaned after their reconstruction failed 1405 during kubelet startup. This includes both successful and failed cleanups. 1406 type: Counter 1407 stabilityLevel: ALPHA 1408 - name: http_inflight_requests 1409 subsystem: kubelet 1410 help: Number of the inflight http requests 1411 type: Gauge 1412 stabilityLevel: ALPHA 1413 labels: 1414 - long_running 1415 - method 1416 - path 1417 - server_type 1418 - name: http_requests_duration_seconds 1419 subsystem: kubelet 1420 help: Duration in seconds to serve http requests 1421 type: Histogram 1422 stabilityLevel: ALPHA 1423 labels: 1424 - long_running 1425 - method 1426 - path 1427 - server_type 1428 buckets: 1429 - 0.005 1430 - 0.01 1431 - 0.025 1432 - 0.05 1433 - 0.1 1434 - 0.25 1435 - 0.5 1436 - 1 1437 - 2.5 1438 - 5 1439 - 10 1440 - name: http_requests_total 1441 subsystem: kubelet 1442 help: Number of the http requests received since the server started 1443 type: Counter 1444 stabilityLevel: ALPHA 1445 labels: 1446 - long_running 1447 - method 1448 - path 1449 - server_type 1450 - name: volume_metric_collection_duration_seconds 1451 subsystem: kubelet 1452 help: Duration in seconds to calculate volume stats 1453 type: Histogram 1454 stabilityLevel: ALPHA 1455 labels: 1456 - metric_source 1457 buckets: 1458 - 0.005 1459 - 0.01 1460 - 0.025 1461 - 0.05 1462 - 0.1 1463 - 0.25 1464 - 0.5 1465 - 1 1466 - 2.5 1467 - 5 1468 - 10 1469 - name: network_programming_duration_seconds 1470 subsystem: kubeproxy 1471 help: In Cluster Network Programming Latency in seconds 1472 type: Histogram 1473 stabilityLevel: ALPHA 1474 buckets: 1475 - 0.25 1476 - 0.5 1477 - 1 1478 - 2 1479 - 3 1480 - 4 1481 - 5 1482 - 6 1483 - 7 1484 - 8 1485 - 9 1486 - 10 1487 - 11 1488 - 12 1489 - 13 1490 - 14 1491 - 15 1492 - 16 1493 - 17 1494 - 18 1495 - 19 1496 - 20 1497 - 21 1498 - 22 1499 - 23 1500 - 24 1501 - 25 1502 - 26 1503 - 27 1504 - 28 1505 - 29 1506 - 30 1507 - 31 1508 - 32 1509 - 33 1510 - 34 1511 - 35 1512 - 36 1513 - 37 1514 - 38 1515 - 39 1516 - 40 1517 - 41 1518 - 42 1519 - 43 1520 - 44 1521 - 45 1522 - 46 1523 - 47 1524 - 48 1525 - 49 1526 - 50 1527 - 51 1528 - 52 1529 - 53 1530 - 54 1531 - 55 1532 - 56 1533 - 57 1534 - 58 1535 - 59 1536 - 60 1537 - 65 1538 - 70 1539 - 75 1540 - 80 1541 - 85 1542 - 90 1543 - 95 1544 - 100 1545 - 105 1546 - 110 1547 - 115 1548 - 120 1549 - 150 1550 - 180 1551 - 210 1552 - 240 1553 - 270 1554 - 300 1555 - name: proxy_healthz_total 1556 subsystem: kubeproxy 1557 help: Cumulative proxy healthz HTTP status 1558 type: Counter 1559 stabilityLevel: ALPHA 1560 labels: 1561 - code 1562 - name: proxy_livez_total 1563 subsystem: kubeproxy 1564 help: Cumulative proxy livez HTTP status 1565 type: Counter 1566 stabilityLevel: ALPHA 1567 labels: 1568 - code 1569 - name: sync_full_proxy_rules_duration_seconds 1570 subsystem: kubeproxy 1571 help: SyncProxyRules latency in seconds for full resyncs 1572 type: Histogram 1573 stabilityLevel: ALPHA 1574 buckets: 1575 - 0.001 1576 - 0.002 1577 - 0.004 1578 - 0.008 1579 - 0.016 1580 - 0.032 1581 - 0.064 1582 - 0.128 1583 - 0.256 1584 - 0.512 1585 - 1.024 1586 - 2.048 1587 - 4.096 1588 - 8.192 1589 - 16.384 1590 - name: sync_partial_proxy_rules_duration_seconds 1591 subsystem: kubeproxy 1592 help: SyncProxyRules latency in seconds for partial resyncs 1593 type: Histogram 1594 stabilityLevel: ALPHA 1595 buckets: 1596 - 0.001 1597 - 0.002 1598 - 0.004 1599 - 0.008 1600 - 0.016 1601 - 0.032 1602 - 0.064 1603 - 0.128 1604 - 0.256 1605 - 0.512 1606 - 1.024 1607 - 2.048 1608 - 4.096 1609 - 8.192 1610 - 16.384 1611 - name: sync_proxy_rules_duration_seconds 1612 subsystem: kubeproxy 1613 help: SyncProxyRules latency in seconds 1614 type: Histogram 1615 stabilityLevel: ALPHA 1616 buckets: 1617 - 0.001 1618 - 0.002 1619 - 0.004 1620 - 0.008 1621 - 0.016 1622 - 0.032 1623 - 0.064 1624 - 0.128 1625 - 0.256 1626 - 0.512 1627 - 1.024 1628 - 2.048 1629 - 4.096 1630 - 8.192 1631 - 16.384 1632 - name: sync_proxy_rules_endpoint_changes_pending 1633 subsystem: kubeproxy 1634 help: Pending proxy rules Endpoint changes 1635 type: Gauge 1636 stabilityLevel: ALPHA 1637 - name: sync_proxy_rules_endpoint_changes_total 1638 subsystem: kubeproxy 1639 help: Cumulative proxy rules Endpoint changes 1640 type: Counter 1641 stabilityLevel: ALPHA 1642 - name: sync_proxy_rules_iptables_last 1643 subsystem: kubeproxy 1644 help: Number of iptables rules written by kube-proxy in last sync 1645 type: Gauge 1646 stabilityLevel: ALPHA 1647 labels: 1648 - table 1649 - name: sync_proxy_rules_iptables_partial_restore_failures_total 1650 subsystem: kubeproxy 1651 help: Cumulative proxy iptables partial restore failures 1652 type: Counter 1653 stabilityLevel: ALPHA 1654 - name: sync_proxy_rules_iptables_restore_failures_total 1655 subsystem: kubeproxy 1656 help: Cumulative proxy iptables restore failures 1657 type: Counter 1658 stabilityLevel: ALPHA 1659 - name: sync_proxy_rules_iptables_total 1660 subsystem: kubeproxy 1661 help: Total number of iptables rules owned by kube-proxy 1662 type: Gauge 1663 stabilityLevel: ALPHA 1664 labels: 1665 - table 1666 - name: sync_proxy_rules_last_queued_timestamp_seconds 1667 subsystem: kubeproxy 1668 help: The last time a sync of proxy rules was queued 1669 type: Gauge 1670 stabilityLevel: ALPHA 1671 - name: sync_proxy_rules_last_timestamp_seconds 1672 subsystem: kubeproxy 1673 help: The last time proxy rules were successfully synced 1674 type: Gauge 1675 stabilityLevel: ALPHA 1676 - name: sync_proxy_rules_no_local_endpoints_total 1677 subsystem: kubeproxy 1678 help: Number of services with a Local traffic policy and no endpoints 1679 type: Gauge 1680 stabilityLevel: ALPHA 1681 labels: 1682 - traffic_policy 1683 - name: sync_proxy_rules_service_changes_pending 1684 subsystem: kubeproxy 1685 help: Pending proxy rules Service changes 1686 type: Gauge 1687 stabilityLevel: ALPHA 1688 - name: sync_proxy_rules_service_changes_total 1689 subsystem: kubeproxy 1690 help: Cumulative proxy rules Service changes 1691 type: Counter 1692 stabilityLevel: ALPHA 1693 - name: plugin_manager_total_plugins 1694 help: Number of plugins in Plugin Manager 1695 type: Custom 1696 stabilityLevel: ALPHA 1697 labels: 1698 - socket_path 1699 - state 1700 - name: probe_duration_seconds 1701 subsystem: prober 1702 help: Duration in seconds for a probe response. 1703 type: Histogram 1704 stabilityLevel: ALPHA 1705 labels: 1706 - container 1707 - namespace 1708 - pod 1709 - probe_type 1710 - name: probe_total 1711 subsystem: prober 1712 help: Cumulative number of a liveness, readiness or startup probe for a container 1713 by result. 1714 type: Counter 1715 stabilityLevel: ALPHA 1716 labels: 1717 - container 1718 - namespace 1719 - pod 1720 - pod_uid 1721 - probe_type 1722 - result 1723 - name: reconstruct_volume_operations_errors_total 1724 help: The number of volumes that failed reconstruction from the operating system 1725 during kubelet startup. 1726 type: Counter 1727 stabilityLevel: ALPHA 1728 - name: reconstruct_volume_operations_total 1729 help: The number of volumes that were attempted to be reconstructed from the operating 1730 system during kubelet startup. This includes both successful and failed reconstruction. 1731 type: Counter 1732 stabilityLevel: ALPHA 1733 - name: volume_manager_selinux_container_errors_total 1734 help: Number of errors when kubelet cannot compute SELinux context for a container. 1735 Kubelet can't start such a Pod then and it will retry, therefore value of this 1736 metric may not represent the actual nr. of containers. 1737 type: Gauge 1738 stabilityLevel: ALPHA 1739 - name: volume_manager_selinux_container_warnings_total 1740 help: Number of errors when kubelet cannot compute SELinux context for a container 1741 that are ignored. They will become real errors when SELinuxMountReadWriteOncePod 1742 feature is expanded to all volume access modes. 1743 type: Gauge 1744 stabilityLevel: ALPHA 1745 - name: volume_manager_selinux_pod_context_mismatch_errors_total 1746 help: Number of errors when a Pod defines different SELinux contexts for its containers 1747 that use the same volume. Kubelet can't start such a Pod then and it will retry, 1748 therefore value of this metric may not represent the actual nr. of Pods. 1749 type: Gauge 1750 stabilityLevel: ALPHA 1751 - name: volume_manager_selinux_pod_context_mismatch_warnings_total 1752 help: Number of errors when a Pod defines different SELinux contexts for its containers 1753 that use the same volume. They are not errors yet, but they will become real errors 1754 when SELinuxMountReadWriteOncePod feature is expanded to all volume access modes. 1755 type: Gauge 1756 stabilityLevel: ALPHA 1757 - name: volume_manager_selinux_volume_context_mismatch_errors_total 1758 help: Number of errors when a Pod uses a volume that is already mounted with a different 1759 SELinux context than the Pod needs. Kubelet can't start such a Pod then and it 1760 will retry, therefore value of this metric may not represent the actual nr. of 1761 Pods. 1762 type: Gauge 1763 stabilityLevel: ALPHA 1764 - name: volume_manager_selinux_volume_context_mismatch_warnings_total 1765 help: Number of errors when a Pod uses a volume that is already mounted with a different 1766 SELinux context than the Pod needs. They are not errors yet, but they will become 1767 real errors when SELinuxMountReadWriteOncePod feature is expanded to all volume 1768 access modes. 1769 type: Gauge 1770 stabilityLevel: ALPHA 1771 - name: volume_manager_selinux_volumes_admitted_total 1772 help: Number of volumes whose SELinux context was fine and will be mounted with 1773 mount -o context option. 1774 type: Gauge 1775 stabilityLevel: ALPHA 1776 - name: volume_manager_total_volumes 1777 help: Number of volumes in Volume Manager 1778 type: Custom 1779 stabilityLevel: ALPHA 1780 labels: 1781 - plugin_name 1782 - state 1783 - name: csr_honored_duration_total 1784 subsystem: certificates_registry 1785 namespace: apiserver 1786 help: Total number of issued CSRs with a requested duration that was honored, sliced 1787 by signer (only kubernetes.io signer names are specifically identified) 1788 type: Counter 1789 stabilityLevel: ALPHA 1790 labels: 1791 - signerName 1792 - name: csr_requested_duration_total 1793 subsystem: certificates_registry 1794 namespace: apiserver 1795 help: Total number of issued CSRs with a requested duration, sliced by signer (only 1796 kubernetes.io signer names are specifically identified) 1797 type: Counter 1798 stabilityLevel: ALPHA 1799 labels: 1800 - signerName 1801 - name: allocated_ips 1802 subsystem: clusterip_allocator 1803 namespace: kube_apiserver 1804 help: Gauge measuring the number of allocated IPs for Services 1805 type: Gauge 1806 stabilityLevel: ALPHA 1807 labels: 1808 - cidr 1809 - name: allocation_errors_total 1810 subsystem: clusterip_allocator 1811 namespace: kube_apiserver 1812 help: Number of errors trying to allocate Cluster IPs 1813 type: Counter 1814 stabilityLevel: ALPHA 1815 labels: 1816 - cidr 1817 - scope 1818 - name: allocation_total 1819 subsystem: clusterip_allocator 1820 namespace: kube_apiserver 1821 help: Number of Cluster IPs allocations 1822 type: Counter 1823 stabilityLevel: ALPHA 1824 labels: 1825 - cidr 1826 - scope 1827 - name: available_ips 1828 subsystem: clusterip_allocator 1829 namespace: kube_apiserver 1830 help: Gauge measuring the number of available IPs for Services 1831 type: Gauge 1832 stabilityLevel: ALPHA 1833 labels: 1834 - cidr 1835 - name: allocated_ports 1836 subsystem: nodeport_allocator 1837 namespace: kube_apiserver 1838 help: Gauge measuring the number of allocated NodePorts for Services 1839 type: Gauge 1840 stabilityLevel: ALPHA 1841 - name: available_ports 1842 subsystem: nodeport_allocator 1843 namespace: kube_apiserver 1844 help: Gauge measuring the number of available NodePorts for Services 1845 type: Gauge 1846 stabilityLevel: ALPHA 1847 - name: backend_tls_failure_total 1848 subsystem: pod_logs 1849 namespace: kube_apiserver 1850 help: Total number of requests for pods/logs that failed due to kubelet server TLS 1851 verification 1852 type: Counter 1853 stabilityLevel: ALPHA 1854 - name: insecure_backend_total 1855 subsystem: pod_logs 1856 namespace: kube_apiserver 1857 help: 'Total number of requests for pods/logs sliced by usage type: enforce_tls, 1858 skip_tls_allowed, skip_tls_denied' 1859 type: Counter 1860 stabilityLevel: ALPHA 1861 labels: 1862 - usage 1863 - name: pods_logs_backend_tls_failure_total 1864 subsystem: pod_logs 1865 namespace: kube_apiserver 1866 help: Total number of requests for pods/logs that failed due to kubelet server TLS 1867 verification 1868 type: Counter 1869 deprecatedVersion: 1.27.0 1870 stabilityLevel: ALPHA 1871 - name: pods_logs_insecure_backend_total 1872 subsystem: pod_logs 1873 namespace: kube_apiserver 1874 help: 'Total number of requests for pods/logs sliced by usage type: enforce_tls, 1875 skip_tls_allowed, skip_tls_denied' 1876 type: Counter 1877 deprecatedVersion: 1.27.0 1878 stabilityLevel: ALPHA 1879 labels: 1880 - usage 1881 - name: goroutines 1882 subsystem: scheduler 1883 help: Number of running goroutines split by the work they do such as binding. 1884 type: Gauge 1885 stabilityLevel: ALPHA 1886 labels: 1887 - operation 1888 - name: permit_wait_duration_seconds 1889 subsystem: scheduler 1890 help: Duration of waiting on permit. 1891 type: Histogram 1892 stabilityLevel: ALPHA 1893 labels: 1894 - result 1895 buckets: 1896 - 0.001 1897 - 0.002 1898 - 0.004 1899 - 0.008 1900 - 0.016 1901 - 0.032 1902 - 0.064 1903 - 0.128 1904 - 0.256 1905 - 0.512 1906 - 1.024 1907 - 2.048 1908 - 4.096 1909 - 8.192 1910 - 16.384 1911 - name: plugin_evaluation_total 1912 subsystem: scheduler 1913 help: Number of attempts to schedule pods by each plugin and the extension point 1914 (available only in PreFilter and Filter.). 1915 type: Counter 1916 stabilityLevel: ALPHA 1917 labels: 1918 - extension_point 1919 - plugin 1920 - profile 1921 - name: plugin_execution_duration_seconds 1922 subsystem: scheduler 1923 help: Duration for running a plugin at a specific extension point. 1924 type: Histogram 1925 stabilityLevel: ALPHA 1926 labels: 1927 - extension_point 1928 - plugin 1929 - status 1930 buckets: 1931 - 1e-05 1932 - 1.5000000000000002e-05 1933 - 2.2500000000000005e-05 1934 - 3.375000000000001e-05 1935 - 5.062500000000001e-05 1936 - 7.593750000000002e-05 1937 - 0.00011390625000000003 1938 - 0.00017085937500000006 1939 - 0.0002562890625000001 1940 - 0.00038443359375000017 1941 - 0.0005766503906250003 1942 - 0.0008649755859375004 1943 - 0.0012974633789062506 1944 - 0.0019461950683593758 1945 - 0.0029192926025390638 1946 - 0.004378938903808595 1947 - 0.006568408355712893 1948 - 0.009852612533569338 1949 - 0.014778918800354007 1950 - 0.02216837820053101 1951 - name: scheduler_cache_size 1952 subsystem: scheduler 1953 help: Number of nodes, pods, and assumed (bound) pods in the scheduler cache. 1954 type: Gauge 1955 stabilityLevel: ALPHA 1956 labels: 1957 - type 1958 - name: scheduling_algorithm_duration_seconds 1959 subsystem: scheduler 1960 help: Scheduling algorithm latency in seconds 1961 type: Histogram 1962 stabilityLevel: ALPHA 1963 buckets: 1964 - 0.001 1965 - 0.002 1966 - 0.004 1967 - 0.008 1968 - 0.016 1969 - 0.032 1970 - 0.064 1971 - 0.128 1972 - 0.256 1973 - 0.512 1974 - 1.024 1975 - 2.048 1976 - 4.096 1977 - 8.192 1978 - 16.384 1979 - name: unschedulable_pods 1980 subsystem: scheduler 1981 help: The number of unschedulable pods broken down by plugin name. A pod will increment 1982 the gauge for all plugins that caused it to not schedule and so this metric have 1983 meaning only when broken down by plugin. 1984 type: Gauge 1985 stabilityLevel: ALPHA 1986 labels: 1987 - plugin 1988 - profile 1989 - name: binder_cache_requests_total 1990 subsystem: scheduler_volume 1991 help: Total number for request volume binding cache 1992 type: Counter 1993 stabilityLevel: ALPHA 1994 labels: 1995 - operation 1996 - name: scheduling_stage_error_total 1997 subsystem: scheduler_volume 1998 help: Volume scheduling stage error count 1999 type: Counter 2000 stabilityLevel: ALPHA 2001 labels: 2002 - operation 2003 - name: legacy_auto_token_uses_total 2004 subsystem: serviceaccount 2005 help: Cumulative auto-generated legacy tokens used 2006 type: Counter 2007 stabilityLevel: ALPHA 2008 - name: legacy_manual_token_uses_total 2009 subsystem: serviceaccount 2010 help: Cumulative manually created legacy tokens used 2011 type: Counter 2012 stabilityLevel: ALPHA 2013 - name: legacy_tokens_total 2014 subsystem: serviceaccount 2015 help: Cumulative legacy service account tokens used 2016 type: Counter 2017 stabilityLevel: ALPHA 2018 - name: stale_tokens_total 2019 subsystem: serviceaccount 2020 help: Cumulative stale projected service account tokens used 2021 type: Counter 2022 stabilityLevel: ALPHA 2023 - name: valid_tokens_total 2024 subsystem: serviceaccount 2025 help: Cumulative valid projected service account tokens used 2026 type: Counter 2027 stabilityLevel: ALPHA 2028 - name: pod_scheduling_sli_duration_seconds 2029 subsystem: scheduler 2030 help: E2e latency for a pod being scheduled, from the time the pod enters the scheduling 2031 queue an d might involve multiple scheduling attempts. 2032 type: Histogram 2033 stabilityLevel: BETA 2034 labels: 2035 - attempts 2036 buckets: 2037 - 0.01 2038 - 0.02 2039 - 0.04 2040 - 0.08 2041 - 0.16 2042 - 0.32 2043 - 0.64 2044 - 1.28 2045 - 2.56 2046 - 5.12 2047 - 10.24 2048 - 20.48 2049 - 40.96 2050 - 81.92 2051 - 163.84 2052 - 327.68 2053 - 655.36 2054 - 1310.72 2055 - 2621.44 2056 - 5242.88 2057 - name: kube_pod_resource_limit 2058 help: Resources limit for workloads on the cluster, broken down by pod. This shows 2059 the resource usage the scheduler and kubelet expect per pod for resources along 2060 with the unit for the resource if any. 2061 type: Custom 2062 stabilityLevel: STABLE 2063 labels: 2064 - namespace 2065 - pod 2066 - node 2067 - scheduler 2068 - priority 2069 - resource 2070 - unit 2071 - name: kube_pod_resource_request 2072 help: Resources requested by workloads on the cluster, broken down by pod. This 2073 shows the resource usage the scheduler and kubelet expect per pod for resources 2074 along with the unit for the resource if any. 2075 type: Custom 2076 stabilityLevel: STABLE 2077 labels: 2078 - namespace 2079 - pod 2080 - node 2081 - scheduler 2082 - priority 2083 - resource 2084 - unit 2085 - name: framework_extension_point_duration_seconds 2086 subsystem: scheduler 2087 help: Latency for running all plugins of a specific extension point. 2088 type: Histogram 2089 stabilityLevel: STABLE 2090 labels: 2091 - extension_point 2092 - profile 2093 - status 2094 buckets: 2095 - 0.0001 2096 - 0.0002 2097 - 0.0004 2098 - 0.0008 2099 - 0.0016 2100 - 0.0032 2101 - 0.0064 2102 - 0.0128 2103 - 0.0256 2104 - 0.0512 2105 - 0.1024 2106 - 0.2048 2107 - name: pending_pods 2108 subsystem: scheduler 2109 help: Number of pending pods, by the queue type. 'active' means number of pods in 2110 activeQ; 'backoff' means number of pods in backoffQ; 'unschedulable' means number 2111 of pods in unschedulablePods that the scheduler attempted to schedule and failed; 2112 'gated' is the number of unschedulable pods that the scheduler never attempted 2113 to schedule because they are gated. 2114 type: Gauge 2115 stabilityLevel: STABLE 2116 labels: 2117 - queue 2118 - name: pod_scheduling_attempts 2119 subsystem: scheduler 2120 help: Number of attempts to successfully schedule a pod. 2121 type: Histogram 2122 stabilityLevel: STABLE 2123 buckets: 2124 - 1 2125 - 2 2126 - 4 2127 - 8 2128 - 16 2129 - name: pod_scheduling_duration_seconds 2130 subsystem: scheduler 2131 help: E2e latency for a pod being scheduled which may include multiple scheduling 2132 attempts. 2133 type: Histogram 2134 deprecatedVersion: 1.28.0 2135 stabilityLevel: STABLE 2136 labels: 2137 - attempts 2138 buckets: 2139 - 0.01 2140 - 0.02 2141 - 0.04 2142 - 0.08 2143 - 0.16 2144 - 0.32 2145 - 0.64 2146 - 1.28 2147 - 2.56 2148 - 5.12 2149 - 10.24 2150 - 20.48 2151 - 40.96 2152 - 81.92 2153 - 163.84 2154 - 327.68 2155 - 655.36 2156 - 1310.72 2157 - 2621.44 2158 - 5242.88 2159 - name: preemption_attempts_total 2160 subsystem: scheduler 2161 help: Total preemption attempts in the cluster till now 2162 type: Counter 2163 stabilityLevel: STABLE 2164 - name: preemption_victims 2165 subsystem: scheduler 2166 help: Number of selected preemption victims 2167 type: Histogram 2168 stabilityLevel: STABLE 2169 buckets: 2170 - 1 2171 - 2 2172 - 4 2173 - 8 2174 - 16 2175 - 32 2176 - 64 2177 - name: queue_incoming_pods_total 2178 subsystem: scheduler 2179 help: Number of pods added to scheduling queues by event and queue type. 2180 type: Counter 2181 stabilityLevel: STABLE 2182 labels: 2183 - event 2184 - queue 2185 - name: schedule_attempts_total 2186 subsystem: scheduler 2187 help: Number of attempts to schedule pods, by the result. 'unschedulable' means 2188 a pod could not be scheduled, while 'error' means an internal scheduler problem. 2189 type: Counter 2190 stabilityLevel: STABLE 2191 labels: 2192 - profile 2193 - result 2194 - name: scheduling_attempt_duration_seconds 2195 subsystem: scheduler 2196 help: Scheduling attempt latency in seconds (scheduling algorithm + binding) 2197 type: Histogram 2198 stabilityLevel: STABLE 2199 labels: 2200 - profile 2201 - result 2202 buckets: 2203 - 0.001 2204 - 0.002 2205 - 0.004 2206 - 0.008 2207 - 0.016 2208 - 0.032 2209 - 0.064 2210 - 0.128 2211 - 0.256 2212 - 0.512 2213 - 1.024 2214 - 2.048 2215 - 4.096 2216 - 8.192 2217 - 16.384 2218 - name: operations_seconds 2219 subsystem: csi 2220 help: Container Storage Interface operation duration with gRPC error code status 2221 total 2222 type: Histogram 2223 stabilityLevel: ALPHA 2224 labels: 2225 - driver_name 2226 - grpc_status_code 2227 - method_name 2228 - migrated 2229 buckets: 2230 - 0.1 2231 - 0.25 2232 - 0.5 2233 - 1 2234 - 2.5 2235 - 5 2236 - 10 2237 - 15 2238 - 25 2239 - 50 2240 - 120 2241 - 300 2242 - 600 2243 - name: storage_operation_duration_seconds 2244 help: Storage operation duration 2245 type: Histogram 2246 stabilityLevel: ALPHA 2247 labels: 2248 - migrated 2249 - operation_name 2250 - status 2251 - volume_plugin 2252 buckets: 2253 - 0.1 2254 - 0.25 2255 - 0.5 2256 - 1 2257 - 2.5 2258 - 5 2259 - 10 2260 - 15 2261 - 25 2262 - 50 2263 - 120 2264 - 300 2265 - 600 2266 - name: volume_operation_total_seconds 2267 help: Storage operation end to end duration in seconds 2268 type: Histogram 2269 stabilityLevel: ALPHA 2270 labels: 2271 - operation_name 2272 - plugin_name 2273 buckets: 2274 - 0.1 2275 - 0.25 2276 - 0.5 2277 - 1 2278 - 2.5 2279 - 5 2280 - 10 2281 - 15 2282 - 25 2283 - 50 2284 - 120 2285 - 300 2286 - 600 2287 - name: graph_actions_duration_seconds 2288 subsystem: node_authorizer 2289 help: Histogram of duration of graph actions in node authorizer. 2290 type: Histogram 2291 stabilityLevel: ALPHA 2292 labels: 2293 - operation 2294 buckets: 2295 - 0.0001 2296 - 0.0002 2297 - 0.0004 2298 - 0.0008 2299 - 0.0016 2300 - 0.0032 2301 - 0.0064 2302 - 0.0128 2303 - 0.0256 2304 - 0.0512 2305 - 0.1024 2306 - 0.2048 2307 - name: conversion_webhook_duration_seconds 2308 namespace: apiserver 2309 help: Conversion webhook request latency 2310 type: Histogram 2311 stabilityLevel: ALPHA 2312 labels: 2313 - failure_type 2314 - result 2315 buckets: 2316 - 0.005 2317 - 0.01 2318 - 0.02 2319 - 0.05 2320 - 0.1 2321 - 0.2 2322 - 0.5 2323 - 1 2324 - 2 2325 - 5 2326 - 10 2327 - 20 2328 - 30 2329 - 45 2330 - 60 2331 - name: conversion_webhook_request_total 2332 namespace: apiserver 2333 help: Counter for conversion webhook requests with success/failure and failure error 2334 type 2335 type: Counter 2336 stabilityLevel: ALPHA 2337 labels: 2338 - failure_type 2339 - result 2340 - name: apiserver_crd_conversion_webhook_duration_seconds 2341 help: CRD webhook conversion duration in seconds 2342 type: Histogram 2343 stabilityLevel: ALPHA 2344 labels: 2345 - crd_name 2346 - from_version 2347 - succeeded 2348 - to_version 2349 buckets: 2350 - 0.001 2351 - 0.002 2352 - 0.004 2353 - 0.008 2354 - 0.016 2355 - 0.032 2356 - 0.064 2357 - 0.128 2358 - 0.256 2359 - 0.512 2360 - 1.024 2361 - 2.048 2362 - 4.096 2363 - 8.192 2364 - 16.384 2365 - name: apiextensions_openapi_v2_regeneration_count 2366 help: Counter of OpenAPI v2 spec regeneration count broken down by causing CRD name 2367 and reason. 2368 type: Counter 2369 stabilityLevel: ALPHA 2370 labels: 2371 - crd 2372 - reason 2373 - name: apiextensions_openapi_v3_regeneration_count 2374 help: Counter of OpenAPI v3 spec regeneration count broken down by group, version, 2375 causing CRD and reason. 2376 type: Counter 2377 stabilityLevel: ALPHA 2378 labels: 2379 - crd 2380 - group 2381 - reason 2382 - version 2383 - name: match_condition_evaluation_errors_total 2384 subsystem: admission 2385 namespace: apiserver 2386 help: Admission match condition evaluation errors count, identified by name of resource 2387 containing the match condition and broken out for each kind containing matchConditions 2388 (webhook or policy), operation and admission type (validate or admit). 2389 type: Counter 2390 stabilityLevel: ALPHA 2391 labels: 2392 - kind 2393 - name 2394 - operation 2395 - type 2396 - name: match_condition_evaluation_seconds 2397 subsystem: admission 2398 namespace: apiserver 2399 help: Admission match condition evaluation time in seconds, identified by name and 2400 broken out for each kind containing matchConditions (webhook or policy), operation 2401 and type (validate or admit). 2402 type: Histogram 2403 stabilityLevel: ALPHA 2404 labels: 2405 - kind 2406 - name 2407 - operation 2408 - type 2409 buckets: 2410 - 0.001 2411 - 0.005 2412 - 0.01 2413 - 0.025 2414 - 0.1 2415 - 0.2 2416 - 0.25 2417 - name: match_condition_exclusions_total 2418 subsystem: admission 2419 namespace: apiserver 2420 help: Admission match condition evaluation exclusions count, identified by name 2421 of resource containing the match condition and broken out for each kind containing 2422 matchConditions (webhook or policy), operation and admission type (validate or 2423 admit). 2424 type: Counter 2425 stabilityLevel: ALPHA 2426 labels: 2427 - kind 2428 - name 2429 - operation 2430 - type 2431 - name: step_admission_duration_seconds_summary 2432 subsystem: admission 2433 namespace: apiserver 2434 help: Admission sub-step latency summary in seconds, broken out for each operation 2435 and API resource and step type (validate or admit). 2436 type: Summary 2437 stabilityLevel: ALPHA 2438 labels: 2439 - operation 2440 - rejected 2441 - type 2442 maxAge: 18000000000000 2443 - name: webhook_fail_open_count 2444 subsystem: admission 2445 namespace: apiserver 2446 help: Admission webhook fail open count, identified by name and broken out for each 2447 admission type (validating or mutating). 2448 type: Counter 2449 stabilityLevel: ALPHA 2450 labels: 2451 - name 2452 - type 2453 - name: webhook_rejection_count 2454 subsystem: admission 2455 namespace: apiserver 2456 help: Admission webhook rejection count, identified by name and broken out for each 2457 admission type (validating or admit) and operation. Additional labels specify 2458 an error type (calling_webhook_error or apiserver_internal_error if an error occurred; 2459 no_error otherwise) and optionally a non-zero rejection code if the webhook rejects 2460 the request with an HTTP status code (honored by the apiserver when the code is 2461 greater or equal to 400). Codes greater than 600 are truncated to 600, to keep 2462 the metrics cardinality bounded. 2463 type: Counter 2464 stabilityLevel: ALPHA 2465 labels: 2466 - error_type 2467 - name 2468 - operation 2469 - rejection_code 2470 - type 2471 - name: webhook_request_total 2472 subsystem: admission 2473 namespace: apiserver 2474 help: Admission webhook request total, identified by name and broken out for each 2475 admission type (validating or mutating) and operation. Additional labels specify 2476 whether the request was rejected or not and an HTTP status code. Codes greater 2477 than 600 are truncated to 600, to keep the metrics cardinality bounded. 2478 type: Counter 2479 stabilityLevel: ALPHA 2480 labels: 2481 - code 2482 - name 2483 - operation 2484 - rejected 2485 - type 2486 - name: check_duration_seconds 2487 subsystem: validating_admission_policy 2488 namespace: apiserver 2489 help: Validation admission latency for individual validation expressions in seconds, 2490 labeled by policy and further including binding, state and enforcement action 2491 taken. 2492 type: Histogram 2493 stabilityLevel: ALPHA 2494 labels: 2495 - enforcement_action 2496 - policy 2497 - policy_binding 2498 - state 2499 buckets: 2500 - 5e-07 2501 - 0.001 2502 - 0.01 2503 - 0.1 2504 - 1 2505 - name: check_total 2506 subsystem: validating_admission_policy 2507 namespace: apiserver 2508 help: Validation admission policy check total, labeled by policy and further identified 2509 by binding, enforcement action taken, and state. 2510 type: Counter 2511 stabilityLevel: ALPHA 2512 labels: 2513 - enforcement_action 2514 - policy 2515 - policy_binding 2516 - state 2517 - name: definition_total 2518 subsystem: validating_admission_policy 2519 namespace: apiserver 2520 help: Validation admission policy count total, labeled by state and enforcement 2521 action. 2522 type: Counter 2523 stabilityLevel: ALPHA 2524 labels: 2525 - enforcement_action 2526 - state 2527 - name: controller_admission_duration_seconds 2528 subsystem: admission 2529 namespace: apiserver 2530 help: Admission controller latency histogram in seconds, identified by name and 2531 broken out for each operation and API resource and type (validate or admit). 2532 type: Histogram 2533 stabilityLevel: STABLE 2534 labels: 2535 - name 2536 - operation 2537 - rejected 2538 - type 2539 buckets: 2540 - 0.005 2541 - 0.025 2542 - 0.1 2543 - 0.5 2544 - 1 2545 - 2.5 2546 - name: step_admission_duration_seconds 2547 subsystem: admission 2548 namespace: apiserver 2549 help: Admission sub-step latency histogram in seconds, broken out for each operation 2550 and API resource and step type (validate or admit). 2551 type: Histogram 2552 stabilityLevel: STABLE 2553 labels: 2554 - operation 2555 - rejected 2556 - type 2557 buckets: 2558 - 0.005 2559 - 0.025 2560 - 0.1 2561 - 0.5 2562 - 1 2563 - 2.5 2564 - name: webhook_admission_duration_seconds 2565 subsystem: admission 2566 namespace: apiserver 2567 help: Admission webhook latency histogram in seconds, identified by name and broken 2568 out for each operation and API resource and type (validate or admit). 2569 type: Histogram 2570 stabilityLevel: STABLE 2571 labels: 2572 - name 2573 - operation 2574 - rejected 2575 - type 2576 buckets: 2577 - 0.005 2578 - 0.025 2579 - 0.1 2580 - 0.5 2581 - 1 2582 - 2.5 2583 - 10 2584 - 25 2585 - name: aggregator_discovery_aggregation_count_total 2586 help: Counter of number of times discovery was aggregated 2587 type: Counter 2588 stabilityLevel: ALPHA 2589 - name: error_total 2590 subsystem: apiserver_audit 2591 help: Counter of audit events that failed to be audited properly. Plugin identifies 2592 the plugin affected by the error. 2593 type: Counter 2594 stabilityLevel: ALPHA 2595 labels: 2596 - plugin 2597 - name: event_total 2598 subsystem: apiserver_audit 2599 help: Counter of audit events generated and sent to the audit backend. 2600 type: Counter 2601 stabilityLevel: ALPHA 2602 - name: level_total 2603 subsystem: apiserver_audit 2604 help: Counter of policy levels for audit events (1 per request). 2605 type: Counter 2606 stabilityLevel: ALPHA 2607 labels: 2608 - level 2609 - name: requests_rejected_total 2610 subsystem: apiserver_audit 2611 help: Counter of apiserver requests rejected due to an error in audit logging backend. 2612 type: Counter 2613 stabilityLevel: ALPHA 2614 - name: compilation_duration_seconds 2615 subsystem: cel 2616 namespace: apiserver 2617 help: CEL compilation time in seconds. 2618 type: Histogram 2619 stabilityLevel: ALPHA 2620 - name: evaluation_duration_seconds 2621 subsystem: cel 2622 namespace: apiserver 2623 help: CEL evaluation time in seconds. 2624 type: Histogram 2625 stabilityLevel: ALPHA 2626 - name: certificate_expiration_seconds 2627 subsystem: client 2628 namespace: apiserver 2629 help: Distribution of the remaining lifetime on the certificate used to authenticate 2630 a request. 2631 type: Histogram 2632 stabilityLevel: ALPHA 2633 buckets: 2634 - 0 2635 - 1800 2636 - 3600 2637 - 7200 2638 - 21600 2639 - 43200 2640 - 86400 2641 - 172800 2642 - 345600 2643 - 604800 2644 - 2.592e+06 2645 - 7.776e+06 2646 - 1.5552e+07 2647 - 3.1104e+07 2648 - name: apiserver_delegated_authn_request_duration_seconds 2649 help: Request latency in seconds. Broken down by status code. 2650 type: Histogram 2651 stabilityLevel: ALPHA 2652 labels: 2653 - code 2654 buckets: 2655 - 0.25 2656 - 0.5 2657 - 0.7 2658 - 1 2659 - 1.5 2660 - 3 2661 - 5 2662 - 10 2663 - name: apiserver_delegated_authn_request_total 2664 help: Number of HTTP requests partitioned by status code. 2665 type: Counter 2666 stabilityLevel: ALPHA 2667 labels: 2668 - code 2669 - name: apiserver_delegated_authz_request_duration_seconds 2670 help: Request latency in seconds. Broken down by status code. 2671 type: Histogram 2672 stabilityLevel: ALPHA 2673 labels: 2674 - code 2675 buckets: 2676 - 0.25 2677 - 0.5 2678 - 0.7 2679 - 1 2680 - 1.5 2681 - 3 2682 - 5 2683 - 10 2684 - name: apiserver_delegated_authz_request_total 2685 help: Number of HTTP requests partitioned by status code. 2686 type: Counter 2687 stabilityLevel: ALPHA 2688 labels: 2689 - code 2690 - name: active_fetch_count 2691 subsystem: token_cache 2692 namespace: authentication 2693 type: Gauge 2694 stabilityLevel: ALPHA 2695 labels: 2696 - status 2697 - name: fetch_total 2698 subsystem: token_cache 2699 namespace: authentication 2700 type: Counter 2701 stabilityLevel: ALPHA 2702 labels: 2703 - status 2704 - name: request_duration_seconds 2705 subsystem: token_cache 2706 namespace: authentication 2707 type: Histogram 2708 stabilityLevel: ALPHA 2709 labels: 2710 - status 2711 - name: request_total 2712 subsystem: token_cache 2713 namespace: authentication 2714 type: Counter 2715 stabilityLevel: ALPHA 2716 labels: 2717 - status 2718 - name: cache_list_fetched_objects_total 2719 namespace: apiserver 2720 help: Number of objects read from watch cache in the course of serving a LIST request 2721 type: Counter 2722 stabilityLevel: ALPHA 2723 labels: 2724 - index 2725 - resource_prefix 2726 - name: cache_list_returned_objects_total 2727 namespace: apiserver 2728 help: Number of objects returned for a LIST request from watch cache 2729 type: Counter 2730 stabilityLevel: ALPHA 2731 labels: 2732 - resource_prefix 2733 - name: cache_list_total 2734 namespace: apiserver 2735 help: Number of LIST requests served from watch cache 2736 type: Counter 2737 stabilityLevel: ALPHA 2738 labels: 2739 - index 2740 - resource_prefix 2741 - name: current_inqueue_requests 2742 subsystem: apiserver 2743 help: Maximal number of queued requests in this apiserver per request kind in last 2744 second. 2745 type: Gauge 2746 stabilityLevel: ALPHA 2747 labels: 2748 - request_kind 2749 - name: dial_duration_seconds 2750 subsystem: egress_dialer 2751 namespace: apiserver 2752 help: Dial latency histogram in seconds, labeled by the protocol (http-connect or 2753 grpc), transport (tcp or uds) 2754 type: Histogram 2755 stabilityLevel: ALPHA 2756 labels: 2757 - protocol 2758 - transport 2759 buckets: 2760 - 0.005 2761 - 0.025 2762 - 0.1 2763 - 0.5 2764 - 2.5 2765 - 12.5 2766 - name: dial_failure_count 2767 subsystem: egress_dialer 2768 namespace: apiserver 2769 help: Dial failure count, labeled by the protocol (http-connect or grpc), transport 2770 (tcp or uds), and stage (connect or proxy). The stage indicates at which stage 2771 the dial failed 2772 type: Counter 2773 stabilityLevel: ALPHA 2774 labels: 2775 - protocol 2776 - stage 2777 - transport 2778 - name: dial_start_total 2779 subsystem: egress_dialer 2780 namespace: apiserver 2781 help: Dial starts, labeled by the protocol (http-connect or grpc) and transport 2782 (tcp or uds). 2783 type: Counter 2784 stabilityLevel: ALPHA 2785 labels: 2786 - protocol 2787 - transport 2788 - name: automatic_reload_failures_total 2789 subsystem: encryption_config_controller 2790 namespace: apiserver 2791 help: Total number of failed automatic reloads of encryption configuration split 2792 by apiserver identity. 2793 type: Counter 2794 stabilityLevel: ALPHA 2795 labels: 2796 - apiserver_id_hash 2797 - name: automatic_reload_last_timestamp_seconds 2798 subsystem: encryption_config_controller 2799 namespace: apiserver 2800 help: Timestamp of the last successful or failed automatic reload of encryption 2801 configuration split by apiserver identity. 2802 type: Gauge 2803 stabilityLevel: ALPHA 2804 labels: 2805 - apiserver_id_hash 2806 - status 2807 - name: automatic_reload_success_total 2808 subsystem: encryption_config_controller 2809 namespace: apiserver 2810 help: Total number of successful automatic reloads of encryption configuration split 2811 by apiserver identity. 2812 type: Counter 2813 stabilityLevel: ALPHA 2814 labels: 2815 - apiserver_id_hash 2816 - name: init_events_total 2817 namespace: apiserver 2818 help: Counter of init events processed in watch cache broken by resource type. 2819 type: Counter 2820 stabilityLevel: ALPHA 2821 labels: 2822 - resource 2823 - name: request_aborts_total 2824 subsystem: apiserver 2825 help: Number of requests which apiserver aborted possibly due to a timeout, for 2826 each group, version, verb, resource, subresource and scope 2827 type: Counter 2828 stabilityLevel: ALPHA 2829 labels: 2830 - group 2831 - resource 2832 - scope 2833 - subresource 2834 - verb 2835 - version 2836 - name: request_body_sizes 2837 subsystem: apiserver 2838 help: Apiserver request body sizes broken out by size. 2839 type: Histogram 2840 stabilityLevel: ALPHA 2841 labels: 2842 - resource 2843 - verb 2844 buckets: 2845 - 50000 2846 - 150000 2847 - 250000 2848 - 350000 2849 - 450000 2850 - 550000 2851 - 650000 2852 - 750000 2853 - 850000 2854 - 950000 2855 - 1.05e+06 2856 - 1.15e+06 2857 - 1.25e+06 2858 - 1.35e+06 2859 - 1.45e+06 2860 - 1.55e+06 2861 - 1.65e+06 2862 - 1.75e+06 2863 - 1.85e+06 2864 - 1.95e+06 2865 - 2.05e+06 2866 - 2.15e+06 2867 - 2.25e+06 2868 - 2.35e+06 2869 - 2.45e+06 2870 - 2.55e+06 2871 - 2.65e+06 2872 - 2.75e+06 2873 - 2.85e+06 2874 - 2.95e+06 2875 - 3.05e+06 2876 - name: request_filter_duration_seconds 2877 subsystem: apiserver 2878 help: Request filter latency distribution in seconds, for each filter type 2879 type: Histogram 2880 stabilityLevel: ALPHA 2881 labels: 2882 - filter 2883 buckets: 2884 - 0.0001 2885 - 0.0003 2886 - 0.001 2887 - 0.003 2888 - 0.01 2889 - 0.03 2890 - 0.1 2891 - 0.3 2892 - 1 2893 - 5 2894 - 10 2895 - 15 2896 - 30 2897 - name: request_post_timeout_total 2898 subsystem: apiserver 2899 help: Tracks the activity of the request handlers after the associated requests 2900 have been timed out by the apiserver 2901 type: Counter 2902 stabilityLevel: ALPHA 2903 labels: 2904 - source 2905 - status 2906 - name: request_sli_duration_seconds 2907 subsystem: apiserver 2908 help: Response latency distribution (not counting webhook duration and priority 2909 & fairness queue wait times) in seconds for each verb, group, version, resource, 2910 subresource, scope and component. 2911 type: Histogram 2912 stabilityLevel: ALPHA 2913 labels: 2914 - component 2915 - group 2916 - resource 2917 - scope 2918 - subresource 2919 - verb 2920 - version 2921 buckets: 2922 - 0.05 2923 - 0.1 2924 - 0.2 2925 - 0.4 2926 - 0.6 2927 - 0.8 2928 - 1 2929 - 1.25 2930 - 1.5 2931 - 2 2932 - 3 2933 - 4 2934 - 5 2935 - 6 2936 - 8 2937 - 10 2938 - 15 2939 - 20 2940 - 30 2941 - 45 2942 - 60 2943 - name: request_slo_duration_seconds 2944 subsystem: apiserver 2945 help: Response latency distribution (not counting webhook duration and priority 2946 & fairness queue wait times) in seconds for each verb, group, version, resource, 2947 subresource, scope and component. 2948 type: Histogram 2949 deprecatedVersion: 1.27.0 2950 stabilityLevel: ALPHA 2951 labels: 2952 - component 2953 - group 2954 - resource 2955 - scope 2956 - subresource 2957 - verb 2958 - version 2959 buckets: 2960 - 0.05 2961 - 0.1 2962 - 0.2 2963 - 0.4 2964 - 0.6 2965 - 0.8 2966 - 1 2967 - 1.25 2968 - 1.5 2969 - 2 2970 - 3 2971 - 4 2972 - 5 2973 - 6 2974 - 8 2975 - 10 2976 - 15 2977 - 20 2978 - 30 2979 - 45 2980 - 60 2981 - name: request_terminations_total 2982 subsystem: apiserver 2983 help: Number of requests which apiserver terminated in self-defense. 2984 type: Counter 2985 stabilityLevel: ALPHA 2986 labels: 2987 - code 2988 - component 2989 - group 2990 - resource 2991 - scope 2992 - subresource 2993 - verb 2994 - version 2995 - name: request_timestamp_comparison_time 2996 subsystem: apiserver 2997 help: Time taken for comparison of old vs new objects in UPDATE or PATCH requests 2998 type: Histogram 2999 stabilityLevel: ALPHA 3000 labels: 3001 - code_path 3002 buckets: 3003 - 0.0001 3004 - 0.0003 3005 - 0.001 3006 - 0.003 3007 - 0.01 3008 - 0.03 3009 - 0.1 3010 - 0.3 3011 - 1 3012 - 5 3013 - name: selfrequest_total 3014 subsystem: apiserver 3015 help: Counter of apiserver self-requests broken out for each verb, API resource 3016 and subresource. 3017 type: Counter 3018 stabilityLevel: ALPHA 3019 labels: 3020 - resource 3021 - subresource 3022 - verb 3023 - name: storage_db_total_size_in_bytes 3024 subsystem: apiserver 3025 help: Total size of the storage database file physically allocated in bytes. 3026 type: Gauge 3027 deprecatedVersion: 1.28.0 3028 stabilityLevel: ALPHA 3029 labels: 3030 - endpoint 3031 - name: storage_decode_errors_total 3032 namespace: apiserver 3033 help: Number of stored object decode errors split by object type 3034 type: Counter 3035 stabilityLevel: ALPHA 3036 labels: 3037 - resource 3038 - name: storage_events_received_total 3039 subsystem: apiserver 3040 help: Number of etcd events received split by kind. 3041 type: Counter 3042 stabilityLevel: ALPHA 3043 labels: 3044 - resource 3045 - name: apiserver_storage_list_evaluated_objects_total 3046 help: Number of objects tested in the course of serving a LIST request from storage 3047 type: Counter 3048 stabilityLevel: ALPHA 3049 labels: 3050 - resource 3051 - name: apiserver_storage_list_fetched_objects_total 3052 help: Number of objects read from storage in the course of serving a LIST request 3053 type: Counter 3054 stabilityLevel: ALPHA 3055 labels: 3056 - resource 3057 - name: apiserver_storage_list_returned_objects_total 3058 help: Number of objects returned for a LIST request from storage 3059 type: Counter 3060 stabilityLevel: ALPHA 3061 labels: 3062 - resource 3063 - name: apiserver_storage_list_total 3064 help: Number of LIST requests served from storage 3065 type: Counter 3066 stabilityLevel: ALPHA 3067 labels: 3068 - resource 3069 - name: apiserver_storage_size_bytes 3070 help: Size of the storage database file physically allocated in bytes. 3071 type: Custom 3072 stabilityLevel: ALPHA 3073 labels: 3074 - cluster 3075 - name: terminated_watchers_total 3076 namespace: apiserver 3077 help: Counter of watchers closed due to unresponsiveness broken by resource type. 3078 type: Counter 3079 stabilityLevel: ALPHA 3080 labels: 3081 - resource 3082 - name: tls_handshake_errors_total 3083 subsystem: apiserver 3084 help: Number of requests dropped with 'TLS handshake error from' error 3085 type: Counter 3086 stabilityLevel: ALPHA 3087 - name: events_dispatched_total 3088 subsystem: watch_cache 3089 namespace: apiserver 3090 help: Counter of events dispatched in watch cache broken by resource type. 3091 type: Counter 3092 stabilityLevel: ALPHA 3093 labels: 3094 - resource 3095 - name: events_received_total 3096 subsystem: watch_cache 3097 namespace: apiserver 3098 help: Counter of events received in watch cache broken by resource type. 3099 type: Counter 3100 stabilityLevel: ALPHA 3101 labels: 3102 - resource 3103 - name: initializations_total 3104 subsystem: watch_cache 3105 namespace: apiserver 3106 help: Counter of watch cache initializations broken by resource type. 3107 type: Counter 3108 stabilityLevel: ALPHA 3109 labels: 3110 - resource 3111 - name: watch_events_sizes 3112 subsystem: apiserver 3113 help: Watch event size distribution in bytes 3114 type: Histogram 3115 stabilityLevel: ALPHA 3116 labels: 3117 - group 3118 - kind 3119 - version 3120 buckets: 3121 - 1024 3122 - 2048 3123 - 4096 3124 - 8192 3125 - 16384 3126 - 32768 3127 - 65536 3128 - 131072 3129 - name: watch_events_total 3130 subsystem: apiserver 3131 help: Number of events sent in watch clients 3132 type: Counter 3133 stabilityLevel: ALPHA 3134 labels: 3135 - group 3136 - kind 3137 - version 3138 - name: authenticated_user_requests 3139 help: Counter of authenticated requests broken out by username. 3140 type: Counter 3141 stabilityLevel: ALPHA 3142 labels: 3143 - username 3144 - name: authentication_attempts 3145 help: Counter of authenticated attempts. 3146 type: Counter 3147 stabilityLevel: ALPHA 3148 labels: 3149 - result 3150 - name: authentication_duration_seconds 3151 help: Authentication duration in seconds broken out by result. 3152 type: Histogram 3153 stabilityLevel: ALPHA 3154 labels: 3155 - result 3156 buckets: 3157 - 0.001 3158 - 0.002 3159 - 0.004 3160 - 0.008 3161 - 0.016 3162 - 0.032 3163 - 0.064 3164 - 0.128 3165 - 0.256 3166 - 0.512 3167 - 1.024 3168 - 2.048 3169 - 4.096 3170 - 8.192 3171 - 16.384 3172 - name: authorization_attempts_total 3173 help: Counter of authorization attempts broken down by result. It can be either 3174 'allowed', 'denied', 'no-opinion' or 'error'. 3175 type: Counter 3176 stabilityLevel: ALPHA 3177 labels: 3178 - result 3179 - name: authorization_duration_seconds 3180 help: Authorization duration in seconds broken out by result. 3181 type: Histogram 3182 stabilityLevel: ALPHA 3183 labels: 3184 - result 3185 buckets: 3186 - 0.001 3187 - 0.002 3188 - 0.004 3189 - 0.008 3190 - 0.016 3191 - 0.032 3192 - 0.064 3193 - 0.128 3194 - 0.256 3195 - 0.512 3196 - 1.024 3197 - 2.048 3198 - 4.096 3199 - 8.192 3200 - 16.384 3201 - name: etcd_bookmark_counts 3202 help: Number of etcd bookmarks (progress notify events) split by kind. 3203 type: Gauge 3204 stabilityLevel: ALPHA 3205 labels: 3206 - resource 3207 - name: etcd_lease_object_counts 3208 help: Number of objects attached to a single etcd lease. 3209 type: Histogram 3210 stabilityLevel: ALPHA 3211 buckets: 3212 - 10 3213 - 50 3214 - 100 3215 - 500 3216 - 1000 3217 - 2500 3218 - 5000 3219 - name: etcd_request_duration_seconds 3220 help: Etcd request latency in seconds for each operation and object type. 3221 type: Histogram 3222 stabilityLevel: ALPHA 3223 labels: 3224 - operation 3225 - type 3226 buckets: 3227 - 0.005 3228 - 0.025 3229 - 0.05 3230 - 0.1 3231 - 0.2 3232 - 0.4 3233 - 0.6 3234 - 0.8 3235 - 1 3236 - 1.25 3237 - 1.5 3238 - 2 3239 - 3 3240 - 4 3241 - 5 3242 - 6 3243 - 8 3244 - 10 3245 - 15 3246 - 20 3247 - 30 3248 - 45 3249 - 60 3250 - name: etcd_request_errors_total 3251 help: Etcd failed request counts for each operation and object type. 3252 type: Counter 3253 stabilityLevel: ALPHA 3254 labels: 3255 - operation 3256 - type 3257 - name: etcd_requests_total 3258 help: Etcd request counts for each operation and object type. 3259 type: Counter 3260 stabilityLevel: ALPHA 3261 labels: 3262 - operation 3263 - type 3264 - name: field_validation_request_duration_seconds 3265 help: Response latency distribution in seconds for each field validation value 3266 type: Histogram 3267 stabilityLevel: ALPHA 3268 labels: 3269 - field_validation 3270 buckets: 3271 - 0.05 3272 - 0.1 3273 - 0.2 3274 - 0.4 3275 - 0.6 3276 - 0.8 3277 - 1 3278 - 1.25 3279 - 1.5 3280 - 2 3281 - 3 3282 - 4 3283 - 5 3284 - 6 3285 - 8 3286 - 10 3287 - 15 3288 - 20 3289 - 30 3290 - 45 3291 - 60 3292 - name: capacity 3293 subsystem: watch_cache 3294 help: Total capacity of watch cache broken by resource type. 3295 type: Gauge 3296 stabilityLevel: ALPHA 3297 labels: 3298 - resource 3299 - name: capacity_decrease_total 3300 subsystem: watch_cache 3301 help: Total number of watch cache capacity decrease events broken by resource type. 3302 type: Counter 3303 stabilityLevel: ALPHA 3304 labels: 3305 - resource 3306 - name: capacity_increase_total 3307 subsystem: watch_cache 3308 help: Total number of watch cache capacity increase events broken by resource type. 3309 type: Counter 3310 stabilityLevel: ALPHA 3311 labels: 3312 - resource 3313 - name: current_inflight_requests 3314 subsystem: apiserver 3315 help: Maximal number of currently used inflight request limit of this apiserver 3316 per request kind in last second. 3317 type: Gauge 3318 stabilityLevel: STABLE 3319 labels: 3320 - request_kind 3321 - name: longrunning_requests 3322 subsystem: apiserver 3323 help: Gauge of all active long-running apiserver requests broken out by verb, group, 3324 version, resource, scope and component. Not all requests are tracked this way. 3325 type: Gauge 3326 stabilityLevel: STABLE 3327 labels: 3328 - component 3329 - group 3330 - resource 3331 - scope 3332 - subresource 3333 - verb 3334 - version 3335 - name: request_duration_seconds 3336 subsystem: apiserver 3337 help: Response latency distribution in seconds for each verb, dry run value, group, 3338 version, resource, subresource, scope and component. 3339 type: Histogram 3340 stabilityLevel: STABLE 3341 labels: 3342 - component 3343 - dry_run 3344 - group 3345 - resource 3346 - scope 3347 - subresource 3348 - verb 3349 - version 3350 buckets: 3351 - 0.005 3352 - 0.025 3353 - 0.05 3354 - 0.1 3355 - 0.2 3356 - 0.4 3357 - 0.6 3358 - 0.8 3359 - 1 3360 - 1.25 3361 - 1.5 3362 - 2 3363 - 3 3364 - 4 3365 - 5 3366 - 6 3367 - 8 3368 - 10 3369 - 15 3370 - 20 3371 - 30 3372 - 45 3373 - 60 3374 - name: request_total 3375 subsystem: apiserver 3376 help: Counter of apiserver requests broken out for each verb, dry run value, group, 3377 version, resource, scope, component, and HTTP response code. 3378 type: Counter 3379 stabilityLevel: STABLE 3380 labels: 3381 - code 3382 - component 3383 - dry_run 3384 - group 3385 - resource 3386 - scope 3387 - subresource 3388 - verb 3389 - version 3390 - name: requested_deprecated_apis 3391 subsystem: apiserver 3392 help: Gauge of deprecated APIs that have been requested, broken out by API group, 3393 version, resource, subresource, and removed_release. 3394 type: Gauge 3395 stabilityLevel: STABLE 3396 labels: 3397 - group 3398 - removed_release 3399 - resource 3400 - subresource 3401 - version 3402 - name: response_sizes 3403 subsystem: apiserver 3404 help: Response size distribution in bytes for each group, version, verb, resource, 3405 subresource, scope and component. 3406 type: Histogram 3407 stabilityLevel: STABLE 3408 labels: 3409 - component 3410 - group 3411 - resource 3412 - scope 3413 - subresource 3414 - verb 3415 - version 3416 buckets: 3417 - 1000 3418 - 10000 3419 - 100000 3420 - 1e+06 3421 - 1e+07 3422 - 1e+08 3423 - 1e+09 3424 - name: apiserver_storage_objects 3425 help: Number of stored objects at the time of last check split by kind. 3426 type: Gauge 3427 stabilityLevel: STABLE 3428 labels: 3429 - resource 3430 - name: dek_cache_fill_percent 3431 subsystem: envelope_encryption 3432 namespace: apiserver 3433 help: Percent of the cache slots currently occupied by cached DEKs. 3434 type: Gauge 3435 stabilityLevel: ALPHA 3436 - name: dek_cache_inter_arrival_time_seconds 3437 subsystem: envelope_encryption 3438 namespace: apiserver 3439 help: Time (in seconds) of inter arrival of transformation requests. 3440 type: Histogram 3441 stabilityLevel: ALPHA 3442 labels: 3443 - transformation_type 3444 buckets: 3445 - 60 3446 - 120 3447 - 240 3448 - 480 3449 - 960 3450 - 1920 3451 - 3840 3452 - 7680 3453 - 15360 3454 - 30720 3455 - name: dek_source_cache_size 3456 subsystem: envelope_encryption 3457 namespace: apiserver 3458 help: Number of records in data encryption key (DEK) source cache. On a restart, 3459 this value is an approximation of the number of decrypt RPC calls the server will 3460 make to the KMS plugin. 3461 type: Gauge 3462 stabilityLevel: ALPHA 3463 labels: 3464 - provider_name 3465 - name: invalid_key_id_from_status_total 3466 subsystem: envelope_encryption 3467 namespace: apiserver 3468 help: Number of times an invalid keyID is returned by the Status RPC call split 3469 by error. 3470 type: Counter 3471 stabilityLevel: ALPHA 3472 labels: 3473 - error 3474 - provider_name 3475 - name: key_id_hash_last_timestamp_seconds 3476 subsystem: envelope_encryption 3477 namespace: apiserver 3478 help: The last time in seconds when a keyID was used. 3479 type: Gauge 3480 stabilityLevel: ALPHA 3481 labels: 3482 - apiserver_id_hash 3483 - key_id_hash 3484 - provider_name 3485 - transformation_type 3486 - name: key_id_hash_status_last_timestamp_seconds 3487 subsystem: envelope_encryption 3488 namespace: apiserver 3489 help: The last time in seconds when a keyID was returned by the Status RPC call. 3490 type: Gauge 3491 stabilityLevel: ALPHA 3492 labels: 3493 - apiserver_id_hash 3494 - key_id_hash 3495 - provider_name 3496 - name: key_id_hash_total 3497 subsystem: envelope_encryption 3498 namespace: apiserver 3499 help: Number of times a keyID is used split by transformation type, provider, and 3500 apiserver identity. 3501 type: Counter 3502 stabilityLevel: ALPHA 3503 labels: 3504 - apiserver_id_hash 3505 - key_id_hash 3506 - provider_name 3507 - transformation_type 3508 - name: kms_operations_latency_seconds 3509 subsystem: envelope_encryption 3510 namespace: apiserver 3511 help: KMS operation duration with gRPC error code status total. 3512 type: Histogram 3513 stabilityLevel: ALPHA 3514 labels: 3515 - grpc_status_code 3516 - method_name 3517 - provider_name 3518 buckets: 3519 - 0.0001 3520 - 0.0002 3521 - 0.0004 3522 - 0.0008 3523 - 0.0016 3524 - 0.0032 3525 - 0.0064 3526 - 0.0128 3527 - 0.0256 3528 - 0.0512 3529 - 0.1024 3530 - 0.2048 3531 - 0.4096 3532 - 0.8192 3533 - 1.6384 3534 - 3.2768 3535 - 6.5536 3536 - 13.1072 3537 - 26.2144 3538 - 52.4288 3539 - name: current_inqueue_seats 3540 subsystem: flowcontrol 3541 namespace: apiserver 3542 help: Number of seats currently pending in queues of the API Priority and Fairness 3543 subsystem 3544 type: Gauge 3545 stabilityLevel: ALPHA 3546 labels: 3547 - flow_schema 3548 - priority_level 3549 - name: current_limit_seats 3550 subsystem: flowcontrol 3551 namespace: apiserver 3552 help: current derived number of execution seats available to each priority level 3553 type: Gauge 3554 stabilityLevel: ALPHA 3555 labels: 3556 - priority_level 3557 - name: current_r 3558 subsystem: flowcontrol 3559 namespace: apiserver 3560 help: R(time of last change) 3561 type: Gauge 3562 stabilityLevel: ALPHA 3563 labels: 3564 - priority_level 3565 - name: demand_seats 3566 subsystem: flowcontrol 3567 namespace: apiserver 3568 help: Observations, at the end of every nanosecond, of (the number of seats each 3569 priority level could use) / (nominal number of seats for that level) 3570 type: TimingRatioHistogram 3571 stabilityLevel: ALPHA 3572 labels: 3573 - priority_level 3574 buckets: 3575 - 0.2 3576 - 0.4 3577 - 0.6 3578 - 0.8 3579 - 1 3580 - 1.2 3581 - 1.4 3582 - 1.7 3583 - 2 3584 - 2.8 3585 - 4 3586 - 6 3587 - name: demand_seats_average 3588 subsystem: flowcontrol 3589 namespace: apiserver 3590 help: Time-weighted average, over last adjustment period, of demand_seats 3591 type: Gauge 3592 stabilityLevel: ALPHA 3593 labels: 3594 - priority_level 3595 - name: demand_seats_high_watermark 3596 subsystem: flowcontrol 3597 namespace: apiserver 3598 help: High watermark, over last adjustment period, of demand_seats 3599 type: Gauge 3600 stabilityLevel: ALPHA 3601 labels: 3602 - priority_level 3603 - name: demand_seats_smoothed 3604 subsystem: flowcontrol 3605 namespace: apiserver 3606 help: Smoothed seat demands 3607 type: Gauge 3608 stabilityLevel: ALPHA 3609 labels: 3610 - priority_level 3611 - name: demand_seats_stdev 3612 subsystem: flowcontrol 3613 namespace: apiserver 3614 help: Time-weighted standard deviation, over last adjustment period, of demand_seats 3615 type: Gauge 3616 stabilityLevel: ALPHA 3617 labels: 3618 - priority_level 3619 - name: dispatch_r 3620 subsystem: flowcontrol 3621 namespace: apiserver 3622 help: R(time of last dispatch) 3623 type: Gauge 3624 stabilityLevel: ALPHA 3625 labels: 3626 - priority_level 3627 - name: epoch_advance_total 3628 subsystem: flowcontrol 3629 namespace: apiserver 3630 help: Number of times the queueset's progress meter jumped backward 3631 type: Counter 3632 stabilityLevel: ALPHA 3633 labels: 3634 - priority_level 3635 - success 3636 - name: latest_s 3637 subsystem: flowcontrol 3638 namespace: apiserver 3639 help: S(most recently dispatched request) 3640 type: Gauge 3641 stabilityLevel: ALPHA 3642 labels: 3643 - priority_level 3644 - name: lower_limit_seats 3645 subsystem: flowcontrol 3646 namespace: apiserver 3647 help: Configured lower bound on number of execution seats available to each priority 3648 level 3649 type: Gauge 3650 stabilityLevel: ALPHA 3651 labels: 3652 - priority_level 3653 - name: next_discounted_s_bounds 3654 subsystem: flowcontrol 3655 namespace: apiserver 3656 help: min and max, over queues, of S(oldest waiting request in queue) - estimated 3657 work in progress 3658 type: Gauge 3659 stabilityLevel: ALPHA 3660 labels: 3661 - bound 3662 - priority_level 3663 - name: next_s_bounds 3664 subsystem: flowcontrol 3665 namespace: apiserver 3666 help: min and max, over queues, of S(oldest waiting request in queue) 3667 type: Gauge 3668 stabilityLevel: ALPHA 3669 labels: 3670 - bound 3671 - priority_level 3672 - name: priority_level_request_utilization 3673 subsystem: flowcontrol 3674 namespace: apiserver 3675 help: Observations, at the end of every nanosecond, of number of requests (as a 3676 fraction of the relevant limit) waiting or in any stage of execution (but only 3677 initial stage for WATCHes) 3678 type: TimingRatioHistogram 3679 stabilityLevel: ALPHA 3680 labels: 3681 - phase 3682 - priority_level 3683 buckets: 3684 - 0 3685 - 0.001 3686 - 0.003 3687 - 0.01 3688 - 0.03 3689 - 0.1 3690 - 0.25 3691 - 0.5 3692 - 0.75 3693 - 1 3694 - name: priority_level_seat_utilization 3695 subsystem: flowcontrol 3696 namespace: apiserver 3697 help: Observations, at the end of every nanosecond, of utilization of seats for 3698 any stage of execution (but only initial stage for WATCHes) 3699 type: TimingRatioHistogram 3700 stabilityLevel: ALPHA 3701 labels: 3702 - priority_level 3703 buckets: 3704 - 0 3705 - 0.1 3706 - 0.2 3707 - 0.3 3708 - 0.4 3709 - 0.5 3710 - 0.6 3711 - 0.7 3712 - 0.8 3713 - 0.9 3714 - 0.95 3715 - 0.99 3716 - 1 3717 constLabels: 3718 phase: executing 3719 - name: read_vs_write_current_requests 3720 subsystem: flowcontrol 3721 namespace: apiserver 3722 help: Observations, at the end of every nanosecond, of the number of requests (as 3723 a fraction of the relevant limit) waiting or in regular stage of execution 3724 type: TimingRatioHistogram 3725 stabilityLevel: ALPHA 3726 labels: 3727 - phase 3728 - request_kind 3729 buckets: 3730 - 0 3731 - 0.001 3732 - 0.01 3733 - 0.1 3734 - 0.2 3735 - 0.3 3736 - 0.4 3737 - 0.5 3738 - 0.6 3739 - 0.7 3740 - 0.8 3741 - 0.9 3742 - 0.95 3743 - 0.99 3744 - 1 3745 - name: request_concurrency_in_use 3746 subsystem: flowcontrol 3747 namespace: apiserver 3748 help: Concurrency (number of seats) occupied by the currently executing (initial 3749 stage for a WATCH, any stage otherwise) requests in the API Priority and Fairness 3750 subsystem 3751 type: Gauge 3752 deprecatedVersion: 1.31.0 3753 stabilityLevel: ALPHA 3754 labels: 3755 - flow_schema 3756 - priority_level 3757 - name: request_concurrency_limit 3758 subsystem: flowcontrol 3759 namespace: apiserver 3760 help: Nominal number of execution seats configured for each priority level 3761 type: Gauge 3762 deprecatedVersion: 1.30.0 3763 stabilityLevel: ALPHA 3764 labels: 3765 - priority_level 3766 - name: request_dispatch_no_accommodation_total 3767 subsystem: flowcontrol 3768 namespace: apiserver 3769 help: Number of times a dispatch attempt resulted in a non accommodation due to 3770 lack of available seats 3771 type: Counter 3772 stabilityLevel: ALPHA 3773 labels: 3774 - flow_schema 3775 - priority_level 3776 - name: request_execution_seconds 3777 subsystem: flowcontrol 3778 namespace: apiserver 3779 help: Duration of initial stage (for a WATCH) or any (for a non-WATCH) stage of 3780 request execution in the API Priority and Fairness subsystem 3781 type: Histogram 3782 stabilityLevel: ALPHA 3783 labels: 3784 - flow_schema 3785 - priority_level 3786 - type 3787 buckets: 3788 - 0 3789 - 0.005 3790 - 0.02 3791 - 0.05 3792 - 0.1 3793 - 0.2 3794 - 0.5 3795 - 1 3796 - 2 3797 - 5 3798 - 10 3799 - 15 3800 - 30 3801 - name: request_queue_length_after_enqueue 3802 subsystem: flowcontrol 3803 namespace: apiserver 3804 help: Length of queue in the API Priority and Fairness subsystem, as seen by each 3805 request after it is enqueued 3806 type: Histogram 3807 stabilityLevel: ALPHA 3808 labels: 3809 - flow_schema 3810 - priority_level 3811 buckets: 3812 - 0 3813 - 10 3814 - 25 3815 - 50 3816 - 100 3817 - 250 3818 - 500 3819 - 1000 3820 - name: seat_fair_frac 3821 subsystem: flowcontrol 3822 namespace: apiserver 3823 help: Fair fraction of server's concurrency to allocate to each priority level that 3824 can use it 3825 type: Gauge 3826 stabilityLevel: ALPHA 3827 - name: target_seats 3828 subsystem: flowcontrol 3829 namespace: apiserver 3830 help: Seat allocation targets 3831 type: Gauge 3832 stabilityLevel: ALPHA 3833 labels: 3834 - priority_level 3835 - name: upper_limit_seats 3836 subsystem: flowcontrol 3837 namespace: apiserver 3838 help: Configured upper bound on number of execution seats available to each priority 3839 level 3840 type: Gauge 3841 stabilityLevel: ALPHA 3842 labels: 3843 - priority_level 3844 - name: watch_count_samples 3845 subsystem: flowcontrol 3846 namespace: apiserver 3847 help: count of watchers for mutating requests in API Priority and Fairness 3848 type: Histogram 3849 stabilityLevel: ALPHA 3850 labels: 3851 - flow_schema 3852 - priority_level 3853 buckets: 3854 - 0 3855 - 1 3856 - 10 3857 - 100 3858 - 1000 3859 - 10000 3860 - name: work_estimated_seats 3861 subsystem: flowcontrol 3862 namespace: apiserver 3863 help: Number of estimated seats (maximum of initial and final seats) associated 3864 with requests in API Priority and Fairness 3865 type: Histogram 3866 stabilityLevel: ALPHA 3867 labels: 3868 - flow_schema 3869 - priority_level 3870 buckets: 3871 - 1 3872 - 2 3873 - 4 3874 - 10 3875 - name: rerouted_request_total 3876 subsystem: apiserver 3877 help: Total number of requests that were proxied to a peer kube apiserver because 3878 the local apiserver was not capable of serving it 3879 type: Counter 3880 stabilityLevel: ALPHA 3881 labels: 3882 - code 3883 - name: data_key_generation_duration_seconds 3884 subsystem: storage 3885 namespace: apiserver 3886 help: Latencies in seconds of data encryption key(DEK) generation operations. 3887 type: Histogram 3888 stabilityLevel: ALPHA 3889 buckets: 3890 - 5e-06 3891 - 1e-05 3892 - 2e-05 3893 - 4e-05 3894 - 8e-05 3895 - 0.00016 3896 - 0.00032 3897 - 0.00064 3898 - 0.00128 3899 - 0.00256 3900 - 0.00512 3901 - 0.01024 3902 - 0.02048 3903 - 0.04096 3904 - name: data_key_generation_failures_total 3905 subsystem: storage 3906 namespace: apiserver 3907 help: Total number of failed data encryption key(DEK) generation operations. 3908 type: Counter 3909 stabilityLevel: ALPHA 3910 - name: envelope_transformation_cache_misses_total 3911 subsystem: storage 3912 namespace: apiserver 3913 help: Total number of cache misses while accessing key decryption key(KEK). 3914 type: Counter 3915 stabilityLevel: ALPHA 3916 - name: transformation_duration_seconds 3917 subsystem: storage 3918 namespace: apiserver 3919 help: Latencies in seconds of value transformation operations. 3920 type: Histogram 3921 stabilityLevel: ALPHA 3922 labels: 3923 - transformation_type 3924 - transformer_prefix 3925 buckets: 3926 - 5e-06 3927 - 1e-05 3928 - 2e-05 3929 - 4e-05 3930 - 8e-05 3931 - 0.00016 3932 - 0.00032 3933 - 0.00064 3934 - 0.00128 3935 - 0.00256 3936 - 0.00512 3937 - 0.01024 3938 - 0.02048 3939 - 0.04096 3940 - 0.08192 3941 - 0.16384 3942 - 0.32768 3943 - 0.65536 3944 - 1.31072 3945 - 2.62144 3946 - 5.24288 3947 - 10.48576 3948 - 20.97152 3949 - 41.94304 3950 - 83.88608 3951 - name: transformation_operations_total 3952 subsystem: storage 3953 namespace: apiserver 3954 help: Total number of transformations. Successful transformation will have a status 3955 'OK' and a varied status string when the transformation fails. This status and 3956 transformation_type fields may be used for alerting on encryption/decryption failure 3957 using transformation_type from_storage for decryption and to_storage for encryption 3958 type: Counter 3959 stabilityLevel: ALPHA 3960 labels: 3961 - status 3962 - transformation_type 3963 - transformer_prefix 3964 - name: x509_insecure_sha1_total 3965 subsystem: webhooks 3966 namespace: apiserver 3967 help: Counts the number of requests to servers with insecure SHA1 signatures in 3968 their serving certificate OR the number of connection failures due to the insecure 3969 SHA1 signatures (either/or, based on the runtime environment) 3970 type: Counter 3971 stabilityLevel: ALPHA 3972 - name: x509_missing_san_total 3973 subsystem: webhooks 3974 namespace: apiserver 3975 help: Counts the number of requests to servers missing SAN extension in their serving 3976 certificate OR the number of connection failures due to the lack of x509 certificate 3977 SAN extension missing (either/or, based on the runtime environment) 3978 type: Counter 3979 stabilityLevel: ALPHA 3980 - name: current_executing_requests 3981 subsystem: flowcontrol 3982 namespace: apiserver 3983 help: Number of requests in initial (for a WATCH) or any (for a non-WATCH) execution 3984 stage in the API Priority and Fairness subsystem 3985 type: Gauge 3986 stabilityLevel: BETA 3987 labels: 3988 - flow_schema 3989 - priority_level 3990 - name: current_executing_seats 3991 subsystem: flowcontrol 3992 namespace: apiserver 3993 help: Concurrency (number of seats) occupied by the currently executing (initial 3994 stage for a WATCH, any stage otherwise) requests in the API Priority and Fairness 3995 subsystem 3996 type: Gauge 3997 stabilityLevel: BETA 3998 labels: 3999 - flow_schema 4000 - priority_level 4001 - name: current_inqueue_requests 4002 subsystem: flowcontrol 4003 namespace: apiserver 4004 help: Number of requests currently pending in queues of the API Priority and Fairness 4005 subsystem 4006 type: Gauge 4007 stabilityLevel: BETA 4008 labels: 4009 - flow_schema 4010 - priority_level 4011 - name: dispatched_requests_total 4012 subsystem: flowcontrol 4013 namespace: apiserver 4014 help: Number of requests executed by API Priority and Fairness subsystem 4015 type: Counter 4016 stabilityLevel: BETA 4017 labels: 4018 - flow_schema 4019 - priority_level 4020 - name: nominal_limit_seats 4021 subsystem: flowcontrol 4022 namespace: apiserver 4023 help: Nominal number of execution seats configured for each priority level 4024 type: Gauge 4025 stabilityLevel: BETA 4026 labels: 4027 - priority_level 4028 - name: rejected_requests_total 4029 subsystem: flowcontrol 4030 namespace: apiserver 4031 help: Number of requests rejected by API Priority and Fairness subsystem 4032 type: Counter 4033 stabilityLevel: BETA 4034 labels: 4035 - flow_schema 4036 - priority_level 4037 - reason 4038 - name: request_wait_duration_seconds 4039 subsystem: flowcontrol 4040 namespace: apiserver 4041 help: Length of time a request spent waiting in its queue 4042 type: Histogram 4043 stabilityLevel: BETA 4044 labels: 4045 - execute 4046 - flow_schema 4047 - priority_level 4048 buckets: 4049 - 0 4050 - 0.005 4051 - 0.02 4052 - 0.05 4053 - 0.1 4054 - 0.2 4055 - 0.5 4056 - 1 4057 - 2 4058 - 5 4059 - 10 4060 - 15 4061 - 30 4062 - name: request_duration_seconds 4063 subsystem: cloud_provider_webhook 4064 help: Request latency in seconds. Broken down by status code. 4065 type: Histogram 4066 stabilityLevel: ALPHA 4067 labels: 4068 - code 4069 - webhook 4070 buckets: 4071 - 0.25 4072 - 0.5 4073 - 0.7 4074 - 1 4075 - 1.5 4076 - 3 4077 - 5 4078 - 10 4079 - name: request_total 4080 subsystem: cloud_provider_webhook 4081 help: Number of HTTP requests partitioned by status code. 4082 type: Counter 4083 stabilityLevel: ALPHA 4084 labels: 4085 - code 4086 - webhook 4087 - name: cloud_provider_taint_removal_delay_seconds 4088 subsystem: node_controller 4089 help: Number of seconds after node creation when NodeController removed the cloud-provider 4090 taint of a single node. 4091 type: Histogram 4092 stabilityLevel: ALPHA 4093 buckets: 4094 - 1 4095 - 4 4096 - 16 4097 - 64 4098 - 256 4099 - 1024 4100 - name: initial_node_sync_delay_seconds 4101 subsystem: node_controller 4102 help: Number of seconds after node creation when NodeController finished the initial 4103 synchronization of a single node. 4104 type: Histogram 4105 stabilityLevel: ALPHA 4106 buckets: 4107 - 1 4108 - 4 4109 - 16 4110 - 64 4111 - 256 4112 - 1024 4113 - name: loadbalancer_sync_total 4114 subsystem: service_controller 4115 help: A metric counting the amount of times any load balancer has been configured, 4116 as an effect of service/node changes on the cluster 4117 type: Counter 4118 stabilityLevel: ALPHA 4119 - name: nodesync_error_total 4120 subsystem: service_controller 4121 help: A metric counting the amount of times any load balancer has been configured 4122 and errored, as an effect of node changes on the cluster 4123 type: Counter 4124 stabilityLevel: ALPHA 4125 - name: nodesync_latency_seconds 4126 subsystem: service_controller 4127 help: A metric measuring the latency for nodesync which updates loadbalancer hosts 4128 on cluster node updates. 4129 type: Histogram 4130 stabilityLevel: ALPHA 4131 buckets: 4132 - 1 4133 - 2 4134 - 4 4135 - 8 4136 - 16 4137 - 32 4138 - 64 4139 - 128 4140 - 256 4141 - 512 4142 - 1024 4143 - 2048 4144 - 4096 4145 - 8192 4146 - 16384 4147 - name: update_loadbalancer_host_latency_seconds 4148 subsystem: service_controller 4149 help: A metric measuring the latency for updating each load balancer hosts. 4150 type: Histogram 4151 stabilityLevel: ALPHA 4152 buckets: 4153 - 1 4154 - 2 4155 - 4 4156 - 8 4157 - 16 4158 - 32 4159 - 64 4160 - 128 4161 - 256 4162 - 512 4163 - 1024 4164 - 2048 4165 - 4096 4166 - 8192 4167 - 16384 4168 - name: kubernetes_build_info 4169 help: A metric with a constant '1' value labeled by major, minor, git version, git 4170 commit, git tree state, build date, Go version, and compiler from which Kubernetes 4171 was built, and platform on which it is running. 4172 type: Gauge 4173 stabilityLevel: ALPHA 4174 labels: 4175 - build_date 4176 - compiler 4177 - git_commit 4178 - git_tree_state 4179 - git_version 4180 - go_version 4181 - major 4182 - minor 4183 - platform 4184 - name: leader_election_master_status 4185 help: Gauge of if the reporting system is master of the relevant lease, 0 indicates 4186 backup, 1 indicates master. 'name' is the string used to identify the lease. Please 4187 make sure to group by name. 4188 type: Gauge 4189 stabilityLevel: ALPHA 4190 labels: 4191 - name 4192 - name: rest_client_dns_resolution_duration_seconds 4193 help: DNS resolver latency in seconds. Broken down by host. 4194 type: Histogram 4195 stabilityLevel: ALPHA 4196 labels: 4197 - host 4198 buckets: 4199 - 0.005 4200 - 0.025 4201 - 0.1 4202 - 0.25 4203 - 0.5 4204 - 1 4205 - 2 4206 - 4 4207 - 8 4208 - 15 4209 - 30 4210 - name: rest_client_exec_plugin_call_total 4211 help: Number of calls to an exec plugin, partitioned by the type of event encountered 4212 (no_error, plugin_execution_error, plugin_not_found_error, client_internal_error) 4213 and an optional exit code. The exit code will be set to 0 if and only if the plugin 4214 call was successful. 4215 type: Counter 4216 stabilityLevel: ALPHA 4217 labels: 4218 - call_status 4219 - code 4220 - name: rest_client_exec_plugin_certificate_rotation_age 4221 help: Histogram of the number of seconds the last auth exec plugin client certificate 4222 lived before being rotated. If auth exec plugin client certificates are unused, 4223 histogram will contain no data. 4224 type: Histogram 4225 stabilityLevel: ALPHA 4226 buckets: 4227 - 600 4228 - 1800 4229 - 3600 4230 - 14400 4231 - 86400 4232 - 604800 4233 - 2.592e+06 4234 - 7.776e+06 4235 - 1.5552e+07 4236 - 3.1104e+07 4237 - 1.24416e+08 4238 - name: rest_client_exec_plugin_ttl_seconds 4239 help: Gauge of the shortest TTL (time-to-live) of the client certificate(s) managed 4240 by the auth exec plugin. The value is in seconds until certificate expiry (negative 4241 if already expired). If auth exec plugins are unused or manage no TLS certificates, 4242 the value will be +INF. 4243 type: Gauge 4244 stabilityLevel: ALPHA 4245 - name: rest_client_rate_limiter_duration_seconds 4246 help: Client side rate limiter latency in seconds. Broken down by verb, and host. 4247 type: Histogram 4248 stabilityLevel: ALPHA 4249 labels: 4250 - host 4251 - verb 4252 buckets: 4253 - 0.005 4254 - 0.025 4255 - 0.1 4256 - 0.25 4257 - 0.5 4258 - 1 4259 - 2 4260 - 4 4261 - 8 4262 - 15 4263 - 30 4264 - 60 4265 - name: rest_client_request_duration_seconds 4266 help: Request latency in seconds. Broken down by verb, and host. 4267 type: Histogram 4268 stabilityLevel: ALPHA 4269 labels: 4270 - host 4271 - verb 4272 buckets: 4273 - 0.005 4274 - 0.025 4275 - 0.1 4276 - 0.25 4277 - 0.5 4278 - 1 4279 - 2 4280 - 4 4281 - 8 4282 - 15 4283 - 30 4284 - 60 4285 - name: rest_client_request_retries_total 4286 help: Number of request retries, partitioned by status code, verb, and host. 4287 type: Counter 4288 stabilityLevel: ALPHA 4289 labels: 4290 - code 4291 - host 4292 - verb 4293 - name: rest_client_request_size_bytes 4294 help: Request size in bytes. Broken down by verb and host. 4295 type: Histogram 4296 stabilityLevel: ALPHA 4297 labels: 4298 - host 4299 - verb 4300 buckets: 4301 - 64 4302 - 256 4303 - 512 4304 - 1024 4305 - 4096 4306 - 16384 4307 - 65536 4308 - 262144 4309 - 1.048576e+06 4310 - 4.194304e+06 4311 - 1.6777216e+07 4312 - name: rest_client_requests_total 4313 help: Number of HTTP requests, partitioned by status code, method, and host. 4314 type: Counter 4315 stabilityLevel: ALPHA 4316 labels: 4317 - code 4318 - host 4319 - method 4320 - name: rest_client_response_size_bytes 4321 help: Response size in bytes. Broken down by verb and host. 4322 type: Histogram 4323 stabilityLevel: ALPHA 4324 labels: 4325 - host 4326 - verb 4327 buckets: 4328 - 64 4329 - 256 4330 - 512 4331 - 1024 4332 - 4096 4333 - 16384 4334 - 65536 4335 - 262144 4336 - 1.048576e+06 4337 - 4.194304e+06 4338 - 1.6777216e+07 4339 - name: rest_client_transport_cache_entries 4340 help: Number of transport entries in the internal cache. 4341 type: Gauge 4342 stabilityLevel: ALPHA 4343 - name: rest_client_transport_create_calls_total 4344 help: 'Number of calls to get a new transport, partitioned by the result of the 4345 operation hit: obtained from the cache, miss: created and added to the cache, 4346 uncacheable: created and not cached' 4347 type: Counter 4348 stabilityLevel: ALPHA 4349 labels: 4350 - result 4351 - name: running_managed_controllers 4352 help: Indicates where instances of a controller are currently running 4353 type: Gauge 4354 stabilityLevel: ALPHA 4355 labels: 4356 - manager 4357 - name 4358 - name: adds_total 4359 subsystem: workqueue 4360 help: Total number of adds handled by workqueue 4361 type: Counter 4362 stabilityLevel: ALPHA 4363 labels: 4364 - name 4365 - name: depth 4366 subsystem: workqueue 4367 help: Current depth of workqueue 4368 type: Gauge 4369 stabilityLevel: ALPHA 4370 labels: 4371 - name 4372 - name: longest_running_processor_seconds 4373 subsystem: workqueue 4374 help: How many seconds has the longest running processor for workqueue been running. 4375 type: Gauge 4376 stabilityLevel: ALPHA 4377 labels: 4378 - name 4379 - name: queue_duration_seconds 4380 subsystem: workqueue 4381 help: How long in seconds an item stays in workqueue before being requested. 4382 type: Histogram 4383 stabilityLevel: ALPHA 4384 labels: 4385 - name 4386 buckets: 4387 - 1e-08 4388 - 1e-07 4389 - 1e-06 4390 - 9.999999999999999e-06 4391 - 9.999999999999999e-05 4392 - 0.001 4393 - 0.01 4394 - 0.1 4395 - 1 4396 - 10 4397 - name: retries_total 4398 subsystem: workqueue 4399 help: Total number of retries handled by workqueue 4400 type: Counter 4401 stabilityLevel: ALPHA 4402 labels: 4403 - name 4404 - name: unfinished_work_seconds 4405 subsystem: workqueue 4406 help: How many seconds of work has done that is in progress and hasn't been observed 4407 by work_duration. Large values indicate stuck threads. One can deduce the number 4408 of stuck threads by observing the rate at which this increases. 4409 type: Gauge 4410 stabilityLevel: ALPHA 4411 labels: 4412 - name 4413 - name: work_duration_seconds 4414 subsystem: workqueue 4415 help: How long in seconds processing an item from workqueue takes. 4416 type: Histogram 4417 stabilityLevel: ALPHA 4418 labels: 4419 - name 4420 buckets: 4421 - 1e-08 4422 - 1e-07 4423 - 1e-06 4424 - 9.999999999999999e-06 4425 - 9.999999999999999e-05 4426 - 0.001 4427 - 0.01 4428 - 0.1 4429 - 1 4430 - 10 4431 - name: disabled_metrics_total 4432 help: The count of disabled metrics. 4433 type: Counter 4434 stabilityLevel: BETA 4435 - name: hidden_metrics_total 4436 help: The count of hidden metrics. 4437 type: Counter 4438 stabilityLevel: BETA 4439 - name: feature_enabled 4440 namespace: kubernetes 4441 help: This metric records the data about the stage and enablement of a k8s feature. 4442 type: Gauge 4443 stabilityLevel: BETA 4444 labels: 4445 - name 4446 - stage 4447 - name: healthcheck 4448 namespace: kubernetes 4449 help: This metric records the result of a single healthcheck. 4450 type: Gauge 4451 stabilityLevel: BETA 4452 labels: 4453 - name 4454 - type 4455 - name: healthchecks_total 4456 namespace: kubernetes 4457 help: This metric records the results of all healthcheck. 4458 type: Counter 4459 stabilityLevel: BETA 4460 labels: 4461 - name 4462 - status 4463 - type 4464 - name: registered_metrics_total 4465 help: The count of registered metrics broken by stability level and deprecation 4466 version. 4467 type: Counter 4468 stabilityLevel: BETA 4469 labels: 4470 - deprecated_version 4471 - stability_level 4472 - name: x509_insecure_sha1_total 4473 subsystem: kube_aggregator 4474 namespace: apiserver 4475 help: Counts the number of requests to servers with insecure SHA1 signatures in 4476 their serving certificate OR the number of connection failures due to the insecure 4477 SHA1 signatures (either/or, based on the runtime environment) 4478 type: Counter 4479 stabilityLevel: ALPHA 4480 - name: x509_missing_san_total 4481 subsystem: kube_aggregator 4482 namespace: apiserver 4483 help: Counts the number of requests to servers missing SAN extension in their serving 4484 certificate OR the number of connection failures due to the lack of x509 certificate 4485 SAN extension missing (either/or, based on the runtime environment) 4486 type: Counter 4487 stabilityLevel: ALPHA 4488 - name: changes 4489 subsystem: endpoint_slice_controller 4490 help: Number of EndpointSlice changes 4491 type: Counter 4492 stabilityLevel: ALPHA 4493 labels: 4494 - operation 4495 - name: desired_endpoint_slices 4496 subsystem: endpoint_slice_controller 4497 help: Number of EndpointSlices that would exist with perfect endpoint allocation 4498 type: Gauge 4499 stabilityLevel: ALPHA 4500 - name: endpoints_added_per_sync 4501 subsystem: endpoint_slice_controller 4502 help: Number of endpoints added on each Service sync 4503 type: Histogram 4504 stabilityLevel: ALPHA 4505 buckets: 4506 - 2 4507 - 4 4508 - 8 4509 - 16 4510 - 32 4511 - 64 4512 - 128 4513 - 256 4514 - 512 4515 - 1024 4516 - 2048 4517 - 4096 4518 - 8192 4519 - 16384 4520 - 32768 4521 - name: endpoints_desired 4522 subsystem: endpoint_slice_controller 4523 help: Number of endpoints desired 4524 type: Gauge 4525 stabilityLevel: ALPHA 4526 - name: endpoints_removed_per_sync 4527 subsystem: endpoint_slice_controller 4528 help: Number of endpoints removed on each Service sync 4529 type: Histogram 4530 stabilityLevel: ALPHA 4531 buckets: 4532 - 2 4533 - 4 4534 - 8 4535 - 16 4536 - 32 4537 - 64 4538 - 128 4539 - 256 4540 - 512 4541 - 1024 4542 - 2048 4543 - 4096 4544 - 8192 4545 - 16384 4546 - 32768 4547 - name: endpointslices_changed_per_sync 4548 subsystem: endpoint_slice_controller 4549 help: Number of EndpointSlices changed on each Service sync 4550 type: Histogram 4551 stabilityLevel: ALPHA 4552 labels: 4553 - topology 4554 - name: num_endpoint_slices 4555 subsystem: endpoint_slice_controller 4556 help: Number of EndpointSlices 4557 type: Gauge 4558 stabilityLevel: ALPHA 4559 - name: syncs 4560 subsystem: endpoint_slice_controller 4561 help: Number of EndpointSlice syncs 4562 type: Counter 4563 stabilityLevel: ALPHA 4564 labels: 4565 - result 4566 - name: aggregator_openapi_v2_regeneration_count 4567 help: Counter of OpenAPI v2 spec regeneration count broken down by causing APIService 4568 name and reason. 4569 type: Counter 4570 stabilityLevel: ALPHA 4571 labels: 4572 - apiservice 4573 - reason 4574 - name: aggregator_openapi_v2_regeneration_duration 4575 help: Gauge of OpenAPI v2 spec regeneration duration in seconds. 4576 type: Gauge 4577 stabilityLevel: ALPHA 4578 labels: 4579 - reason 4580 - name: aggregator_unavailable_apiservice 4581 help: Gauge of APIServices which are marked as unavailable broken down by APIService 4582 name. 4583 type: Custom 4584 stabilityLevel: ALPHA 4585 labels: 4586 - name 4587 - name: aggregator_unavailable_apiservice_total 4588 help: Counter of APIServices which are marked as unavailable broken down by APIService 4589 name and reason. 4590 type: Counter 4591 stabilityLevel: ALPHA 4592 labels: 4593 - name 4594 - reason 4595 - name: api_request_duration_seconds 4596 namespace: cloudprovider_azure 4597 help: Latency of an Azure API call 4598 type: Histogram 4599 stabilityLevel: ALPHA 4600 labels: 4601 - request 4602 - resource_group 4603 - source 4604 - subscription_id 4605 buckets: 4606 - 0.1 4607 - 0.25 4608 - 0.5 4609 - 1 4610 - 2.5 4611 - 5 4612 - 10 4613 - 15 4614 - 25 4615 - 50 4616 - 120 4617 - 300 4618 - 600 4619 - 1200 4620 - name: api_request_errors 4621 namespace: cloudprovider_azure 4622 help: Number of errors for an Azure API call 4623 type: Counter 4624 stabilityLevel: ALPHA 4625 labels: 4626 - request 4627 - resource_group 4628 - source 4629 - subscription_id 4630 - name: api_request_ratelimited_count 4631 namespace: cloudprovider_azure 4632 help: Number of rate limited Azure API calls 4633 type: Counter 4634 stabilityLevel: ALPHA 4635 labels: 4636 - request 4637 - resource_group 4638 - source 4639 - subscription_id 4640 - name: api_request_throttled_count 4641 namespace: cloudprovider_azure 4642 help: Number of throttled Azure API calls 4643 type: Counter 4644 stabilityLevel: ALPHA 4645 labels: 4646 - request 4647 - resource_group 4648 - source 4649 - subscription_id 4650 - name: op_duration_seconds 4651 namespace: cloudprovider_azure 4652 help: Latency of an Azure service operation 4653 type: Histogram 4654 stabilityLevel: ALPHA 4655 labels: 4656 - request 4657 - resource_group 4658 - source 4659 - subscription_id 4660 buckets: 4661 - 0.1 4662 - 0.2 4663 - 0.5 4664 - 1 4665 - 10 4666 - 20 4667 - 30 4668 - 40 4669 - 50 4670 - 60 4671 - 100 4672 - 200 4673 - 300 4674 - name: op_failure_count 4675 namespace: cloudprovider_azure 4676 help: Number of failed Azure service operations 4677 type: Counter 4678 stabilityLevel: ALPHA 4679 labels: 4680 - request 4681 - resource_group 4682 - source 4683 - subscription_id 4684 - name: cloudprovider_gce_api_request_duration_seconds 4685 help: Latency of a GCE API call 4686 type: Histogram 4687 stabilityLevel: ALPHA 4688 labels: 4689 - region 4690 - request 4691 - version 4692 - zone 4693 - name: cloudprovider_gce_api_request_errors 4694 help: Number of errors for an API call 4695 type: Counter 4696 stabilityLevel: ALPHA 4697 labels: 4698 - region 4699 - request 4700 - version 4701 - zone 4702 - name: cloudprovider_vsphere_api_request_duration_seconds 4703 help: Latency of vsphere api call 4704 type: Histogram 4705 stabilityLevel: ALPHA 4706 labels: 4707 - request 4708 - name: cloudprovider_vsphere_api_request_errors 4709 help: vsphere Api errors 4710 type: Counter 4711 stabilityLevel: ALPHA 4712 labels: 4713 - request 4714 - name: cloudprovider_vsphere_operation_duration_seconds 4715 help: Latency of vsphere operation call 4716 type: Histogram 4717 stabilityLevel: ALPHA 4718 labels: 4719 - operation 4720 - name: cloudprovider_vsphere_operation_errors 4721 help: vsphere operation errors 4722 type: Counter 4723 stabilityLevel: ALPHA 4724 labels: 4725 - operation 4726 - name: cloudprovider_vsphere_vcenter_versions 4727 help: Versions for connected vSphere vCenters 4728 type: Custom 4729 stabilityLevel: ALPHA 4730 labels: 4731 - hostname 4732 - version 4733 - build 4734 - name: get_token_count 4735 help: Counter of total Token() requests to the alternate token source 4736 type: Counter 4737 stabilityLevel: ALPHA 4738 - name: get_token_fail_count 4739 help: Counter of failed Token() requests to the alternate token source 4740 type: Counter 4741 stabilityLevel: ALPHA 4742 - name: number_of_l4_ilbs 4743 help: Number of L4 ILBs 4744 type: Gauge 4745 stabilityLevel: ALPHA 4746 labels: 4747 - feature 4748 - name: pod_security_errors_total 4749 help: Number of errors preventing normal evaluation. Non-fatal errors may result 4750 in the latest restricted profile being used for evaluation. 4751 type: Counter 4752 stabilityLevel: ALPHA 4753 labels: 4754 - fatal 4755 - request_operation 4756 - resource 4757 - subresource 4758 - name: pod_security_evaluations_total 4759 help: Number of policy evaluations that occurred, not counting ignored or exempt 4760 requests. 4761 type: Counter 4762 stabilityLevel: ALPHA 4763 labels: 4764 - decision 4765 - mode 4766 - policy_level 4767 - policy_version 4768 - request_operation 4769 - resource 4770 - subresource 4771 - name: pod_security_exemptions_total 4772 help: Number of exempt requests, not counting ignored or out of scope requests. 4773 type: Counter 4774 stabilityLevel: ALPHA 4775 labels: 4776 - request_operation 4777 - resource 4778 - subresource