A Prometheus exporter to provide "extended" metrics about a Ceph cluster's running components (e.g., RGW).
Due to the closure of Koor Technologies, Inc. this repository has been made to continue the work on the extended-ceph-exporter project.
-
Needs a Ceph cluster up and running.
-
Needs an admin user
radosgw-admin user create --uid extended-ceph-exporter --display-name "extended-ceph-exporter admin user" --caps "buckets=read;users=read;usage=read;metadata=read;zone=read" # Access key / "Username" radosgw-admin user info --uid extended-ceph-exporter | jq '.keys[0].access_key' # Secret key / "Password radosgw-admin user info --uid extended-ceph-exporter | jq '.keys[0].secret_key'
If using Rook to manage RGW, the admin user may also be created using a CephOjectStoreUser
resource:
apiVersion: ceph.rook.io/v1
kind: CephObjectStoreUser
metadata:
name: extended-ceph-exporter
namespace: rook-ceph
spec:
store: <objectstore-name>
clusterNamespace: rook-ceph
displayName: extended-ceph-exporter
capabilities:
buckets: read
users: read
usage: read
metadata: read
zone: read
Applying this will create an user with all permissions needed.
-
Clone the repository, download release binary or pull the container image:
git clone https://github.com/galexrt/extended-ceph-exporter cd extended-ceph-exporter
-
Create a copy of the
.env.example
file and name it.env
. Configure your RGW credentials and endpoint in the.env
file. -
Configure Prometheus to collect metrics from the exporter from
:9138/metrics
endpoint using a static configuration, here's a sample scrape job from theprometheus.yml
:# For more information on Prometheus scrape_configs: # https://prometheus.io/docs/prometheus/latest/configuration/configuration/#scrape_config scrape_configs: - job_name: "extended-ceph-metrics" # Override the global default and scrape targets from this job every 30 seconds. scrape_interval: 30s static_configs: # Please change the ip address `127.0.0.1` to target the exporter is running - targets: ['127.0.0.1:9138']
-
To run the exporter locally, run
go run .
-
Should you have Grafana running for metrics visulization, check out the Grafana dashboards.
To install the exporter to Kubernetes using Helm, check out the extended-ceph-exporter Helm Chart.
There is varying support for collectors. The tables below list all existing collectors and the required Ceph components.
Name | Description | Ceph Component |
---|---|---|
rgw_buckets |
Exposes RGW Bucket Usage and Quota metrics from the Ceph cluster. | RGW |
rgw_user_quota |
Exposes RGW User Quota metrics from the Ceph cluster. | RGW |
Name | Description | Ceph Component |
---|---|---|
rbd_volumes |
Exposes RBD volumes size (volume pool, id, and name are available as labels). | RBD |
You can use the exporter to scrape metrics from multiple RGW realms by providing multiple RGWs in the realm config file.
An example realm config file can be found here realms.example.yaml
.
$ extended-ceph-exporter --help
Usage of exporter:
--collectors-enabled strings List of enabled collectors (please refer to the readme for a list of all available collectors) (default [rgw_user_quota,rgw_buckets])
--config config.yaml Config file path (default name config.yaml , current and `/config` directory).
--realms-config --multi-realm-config Path to your realms.yaml config file (old flag name: --multi-realm-config) (default "realms.yaml")
--version Show version info and exit
pflag: help requested
exit status 2
- Golang 1.23.x (or higher should work)
- Ceph development files (
librados
,librdb
)- If you are using
nix
, theflake.nix
should be satisfy these lib dependencies.
- If you are using
helm
When changing anything in the Helm Chart, the version in the Chart.yaml
needs to be increased according to Semver.
Additionally make helm-doc
must be run afterwards and the changes to the Helm Chart's README.md
must be commited as well.
A VSCode debug config is available to run and debug the project.
To make the exporter talk with a Ceph RGW S3 endpoint, create a copy of the .env.example
file and name it .env
.
Be sure ot add your Ceph RGW S3 endpoint and credentials in it.