Configure the Prometheus receiver to collect NetApp metrics

Configure the Prometheus receiver to collect NetApp metrics.

Cisco AI PODs are designed to integrate with high-performance storage solutions. You can monitor the performance of Cisco AI PODs that use NetApp storage resources with NetApp Trident, a storage orchestrator and management tool for containers and Kubernetes distributions.

Splunk Observability Cloud uses the Prometheus receiver to collect metrics from NetApp Trident, which exposes a /metrics endpoint that publishes Prometheus-compatible metrics.

To configure the Prometheus receiver to collect NetApp Trident metrics, you must meet the following requirements.

  1. Activate the Prometheus receiver for NetApp Trident manually in the Splunk Distribution of the OpenTelemetry Collector configuration by making the following changes to your values.yaml configuration file.
    1. Add prometheus/trident to the receiver_creator/cisco_ai_pods section. For example:
      YAML
      receiver_creator/cisco-ai-pods: 
        # Name of the extensions to watch for endpoints to start and stop. 
        watch_observers: [ k8s_observer ] 
        receivers: 
          prometheus/trident: 
            config: 
              config: 
                scrape_configs: 
                  - job_name: trident-metrics 
                    scrape_interval: 10s 
                    metrics_path: /metrics 
                    static_configs: 
                      - targets: 
                        - '`endpoint`:8001' 
               rule: type == "pod" && labels["app"] == "controller.csi.trident.netapp.io"
    2. Add the receiver_creator/cisco-ai-pods receiver in the metrics pipeline of the service section, if it's not already present. For example:
      YAML
      metrics/cisco-ai-pods: 
        exporters: 
          - signalfx 
        processors: 
          - memory_limiter 
          - batch 
          - resourcedetection 
          - resource 
        receivers: 
          - receiver_creator/cisco-ai-pods
  2. Use Helm to apply the configuration changes:
    CODE
    helm upgrade ucs-otel-collector \ 
    --set="clusterName=$CLUSTER_NAME" \ 
    --set="environment=$ENVIRONMENT_NAME" \ 
    --set="splunkObservability.accessToken$SPLUNK_ACCESS_TOKEN\ 
    --set="splunkObservability.realm=$SPLUNKREALM" \ 
    --set="splunkPlatform.endpoint=$SPLUNK_HEC_URL" \ 
    --set="splunkPlatform.token=$SPLUNK_HEC_TOKEN" \ 
    --set="splunkPlatform.index=$SPLUNK_INDEX" \ 
    
    -n otel \ 
    -f ./base-otel-collector-config/values.yaml \ 
    splunk-otel-collector-chart/splunk-otel-collector

Configuration settings

Learn about the configuration options for the Prometheus receiver.

To view the configuration options for the Prometheus receiver, see Settings.

Metrics

The following metrics are available for NetApp. These metrics fall under the default metric category.

For more information on these metrics, see Monitor Trident in the NetApp Trident documentation and the metrics.go file in the NetApp Trident GitHub repository.
Metric name Description
trident_backend_count The total number of backends.
trident_node_count The total number of nodes.
trident_operation_duration_milliseconds_count The total count of observed operations.
trident_operation_duration_milliseconds_quantile The latency quantile for operation events.
trident_operation_duration_milliseconds_sum The total duration of all observed operations.
trident_storageclass_count The total number of storage classes.
trident_volume_allocated_bytes The total allocated number of bytes grouped by backends and volumes.
trident_volume_count The total number of volumes.
trident_volume_total_bytes The total number of bytes in all volumes.