Anemometer is a tool for running SQL queries and pushing results as metrics to Datadog
Why "Anemometer"
An anemometer is a device used for measuring wind speed and direction.
This project was originally created to help us monitor some tables in Airflow, but was later updated so it could work generically with any database.
Supported Databases
We currently support the following databases:
- Postgres
- Vertica
- BigQuery
- SQLite (mostly for local development and testing)
Adding support for another database
Support for any of the databases listed here can be added fairly easily!
- Read How to contribute
- Add the driver to
go.mod, and ensurego.sumgets updated - Add import for the new driver to
monitor.go - Update this README to add the new database
- Submit a Pull Request
Setup
The latest version of Anemometer can be found on the Releases tab.
Example Configuration:
statsd: address: 127.0.0.1:8125 tags: - environment:production monitors: # Gauge metric (default) - values that can go up or down - name: airflow-dag-disabled database: type: postgres uri: postgresql://username:password@localhost:5432/database?sslmode=disable sleep_duration: 300 metric: airflow.dag.disabled metric_type: gauge # optional - this is the default sql: > SELECT dag_id AS dag_id, CASE WHEN is_paused AND NOT is_subdag THEN 1 ELSE 0 END AS metric FROM dag # Histogram metric - statistical distribution on each host - name: airflow-task-queued-seconds database: type: postgres uri: postgresql://username:password@localhost:5432/database?sslmode=disable sleep_duration: 300 metric: airflow.task.queued_seconds metric_type: histogram sql: > SELECT dag_id AS dag_id, task_id AS task_id, EXTRACT(EPOCH FROM (current_timestamp - queued_dttm)) AS metric FROM task_instance WHERE state = 'queued' # Count metric - track number of events - name: failed-tasks-count database: type: postgres uri: postgresql://username:password@localhost:5432/database?sslmode=disable sleep_duration: 60 metric: airflow.task.failed metric_type: count sql: > SELECT dag_id AS dag_id, COUNT(*) AS metric FROM task_instance WHERE state = 'failed' AND end_date > NOW() - INTERVAL '1 hour' GROUP BY dag_id
statsd
This is where you tell Anemometer where to send StatsD metrics
address- The address:port on which StatsD is listening (usually127.0.0.1:8125)tags- Default tags to send with every metric, optional
monitors
This is where you tell Anemometer about the monitor(s) configuration
name- The name of this monitor, mainly used in loggingdatabase.type- The type of database connection to be used (postgres,vertica,bigquery, andsqlite3are currently supported)database.uri- The URI connection string used to connect to the database (usually followsprotocol://username:password@hostname:port/database)sleep_duration- How long to wait between pushes to StatsD (in seconds)metric- The name of the metric to be sent to StatsDmetric_type- The type of metric to send to Datadog (optional, defaults togauge)sql- The SQL query to execute when populating the metric's values/tags (see SQL Query Structure)
SQL Query Structure
Anemometer makes the following assumptions about the results of your query:
- Exactly one column will be named
metric, and the value is convertable tofloat64(no strings) - An optional column named
timestampcan be included to explicitly provide a timestamp for the metrics (only supported forgaugeandcounttypes) - All other columns will be aggregated into tags and sent to StatsD
- The tags will take the form of
column_name:value
Metric Types
Anemometer supports all four Datadog metric types. You can specify the metric
type using the metric_type configuration option:
Gauge (default)
- Use case: Values that can go up or down (e.g., CPU usage, memory usage, queue depth)
- Configuration:
metric_type: gauge(or omit for default) - StatsD format:
metric_name:value|g|#tags - Timestamp support: ✅ Supports optional
timestampcolumn
Count
- Use case: Track how many times something happened (e.g., requests, errors, events)
- Configuration:
metric_type: count - StatsD format:
metric_name:value|c|#tags - Note: Values are converted to integers
- Timestamp support: ✅ Supports optional
timestampcolumn
Histogram
- Use case: Track statistical distribution of values on each host (e.g., request latency, file sizes)
- Configuration:
metric_type: histogram - StatsD format:
metric_name:value|h|#tags - Timestamp support: ❌ Uses current time only
Distribution
- Use case: Track statistical distribution of values across your infrastructure (e.g., request latency across all hosts)
- Configuration:
metric_type: distribution - StatsD format:
metric_name:value|d|#tags - Timestamp support: ❌ Uses current time only
Note: The metric_type field is optional and defaults to gauge for
backwards compatibility. Existing configurations will continue to work without
any changes.
Timestamp Support
Anemometer supports custom timestamps for gauge and count metrics by including an optional timestamp column in your SQL query results. This allows you to send metrics with specific timestamps rather than using the current time.
Supported timestamp formats:
- RFC3339 strings:
"2023-12-25T10:30:00Z" - Unix timestamps:
1703505000(asint64,int32,int, orfloat64) - Database time objects: Direct
time.Timeorsql.NullTimevalues
Example with timestamp:
SELECT 'production' AS environment, COUNT(*) AS metric, '2023-12-25T10:30:00Z' AS timestamp FROM users WHERE created_at BETWEEN '2023-12-25 00:00:00' AND '2023-12-25 23:59:59'
Important Datadog considerations:
⚠️ Historical Data Limitations: By default, Datadog will reject metrics with timestamps older than 4 hours from the current time. To send historical metrics, you need to enable "Allow metrics with timestamps in past" in your Datadog organization settings.
📖 For more information, see Datadog's documentation on historical metric in gestion.
Query Example
Single row result
To monitor the number of records in your user's table you might do something like this:
SELECT 'production' AS environment, 'users' AS table_name, COUNT(0) AS metric FROM users
Resulting in the following:
environment | table_name | metric
-------------+------------+--------
production | users | 99
Assuming we named our metric table.records with metric_type: gauge, this
would result in the following data being sent to StatsD:
table.records:99|g|#environment:production,table_name:users
Multiple row result
To monitor the number of queries each user is running in your database you might do something like this:
SELECT 'production' AS environment, usename AS user_name, COUNT(0) AS metric FROM pg_stat_activity WHERE query != '<IDLE>' GROUP BY usename
Resulting in the following:
environment | user_name | metric
-------------+-----------+--------
production | cjonesy | 160
production | postgres | 6
Assuming we named our metric database.queries with metric_type: count, this
would result in the following data being sent to StatsD:
database.queries:160|c|#environment:production,user_name:cjonesy
database.queries:6|c|#environment:production,user_name:postgres
Notice that one metric is sent for each row in the query.
Usage
Basic Usage
Anemometer (A SQL -> StatsD metrics generator)
Usage:
anemometer [command]
Available Commands:
help Help about any command
start Start the Anemometer agent
version Print the version number
Flags:
-h, --help help for anemometer
Use "anemometer [command] --help" for more information about a command.
To start the agent:
anemometer start -c /path/to/your/config.yml
Using Docker
You can run Anemometer using Docker with the image from GitHub Container Registry:
docker run -v /path/to/your/config.yml:/config.yml ghcr.io/simplifi/anemometer start -c /config.yml
This mounts your local configuration file into the container and runs the start command.
Development
Testing locally
If you want to test this out locally you can run the following to start Anemometer:
anemometer start -c /path/to/config.yml
You can see the metrics that would be sent by watching the statsd port on localhost:
Compiling
Running Tests
To run all the standard tests:
Releasing
This project is using goreleaser. GitHub release creation is automated using Travis CI. New releases are automatically created when new tags are pushed to the repo.
How to contribute
This project has some clear Contribution Guidelines and expectations that you can read here (CONTRIBUTING).
The contribution guidelines outline the process that you'll need to follow to get a patch merged.
And you don't just have to write code. You can help out by writing documentation, tests, or even by giving feedback about this work.
Thank you for contributing!
