Prometheus Scraping

Good day.

Im new to the confluent cloud and scraping the metrics with the Prometheus config, which lists each cluster, connector, or ksql db by the ID. If any of those objects are removed for some reason. The scrape will break with a 403 unauthorized error as that specified object no longer exists.

Is there a way to scrape at an ORG level and not be dependent on listing each clusterID.
or if that ID is not available, continue without failing

Thanks

Trevor

hey @trevorwebster

welcome :slight_smile:

could you elaborate a bit more?
who did you configure prometheus?
I guess you are using the metrics api?

best,
michael

This is correct, using the metrics api with the generated config

scrape_configs:

  • job_name: Confluent Cloud
    scrape_interval: 1m
    scrape_timeout: 1m
    honor_timestamps: true
    static_configs:
    • targets:
      • api.telemetry.confluent.cloud
        scheme: https
        basic_auth:
        username:
        password:
        metrics_path: /v2/metrics/cloud/export
        params:
        “resource.kafka.id”:
      • lkc-r
      • lkc-y
      • lkc-d
      • lkc-3
      • lkc-6

thanks for the input.

need to check by myself whether it’s possible.
will come back once tested :slight_smile:

same error here

afaik there is no possibility to automate this out of the box

one thing which might work is to get the cluster ids via confluent cli and create the prometheus.yml by script or similar.

will try this tomorrow

best,
michael