diff --git a/content/en/database_monitoring/setup_postgres/rds/_index.md b/content/en/database_monitoring/setup_postgres/rds/_index.md index a6cbfe71f4b..2ffe8a57ad7 100644 --- a/content/en/database_monitoring/setup_postgres/rds/_index.md +++ b/content/en/database_monitoring/setup_postgres/rds/_index.md @@ -6,7 +6,7 @@ further_reading: tag: "Documentation" text: "Basic Postgres Integration" - link: "/database_monitoring/guide/rds_autodiscovery" - tag: "Documenation" + tag: "Documentation" text: "Autodiscovery for RDS" - link: "/database_monitoring/guide/parameterized_queries/" tag: "Documentation" @@ -40,7 +40,7 @@ Performance impact Database Monitoring runs as an integration on top of the base Agent ([see benchmarks][1]). Proxies, load balancers, and connection poolers -: The Datadog Agent must connect directly to the host being monitored. For self-hosted databases, `127.0.0.1` or the socket is preferred. The Agent should not connect to the database through a proxy, load balancer, or connection pooler such as `pgbouncer`. If the Agent connects to different hosts while it is running (as in the case of failover, load balancing, and so on), the Agent calculates the difference in statistics between two hosts, producing inaccurate metrics. +: The Datadog Agent must connect directly to the host being monitored. For self-hosted databases, use `127.0.0.1` or the socket. The Agent should not connect to the database through a proxy, load balancer, or connection pooler such as `pgbouncer`. If the Agent connects to different hosts while it is running (as in the case of failover, load balancing, and so on), the Agent calculates the difference in statistics between two hosts, producing inaccurate metrics. Data security considerations : See [Sensitive information][2] for information about what data the Agent collects from your databases and how to ensure it is secure. @@ -53,20 +53,27 @@ Enable **Resource Collection** in the **Resource Collection** section of your [A Configure the following [parameters][4] in the [DB parameter group][5] and then **restart the server** for the settings to take effect. For more information about these parameters, see the [Postgres documentation][6]. +**Required parameters** + | Parameter | Value | Description | | --- | --- | --- | | `shared_preload_libraries` | `pg_stat_statements` | Required for `postgresql.queries.*` metrics. Enables collection of query metrics using the [pg_stat_statements][6] extension. | | `track_activity_query_size` | `4096` | Required for collection of larger queries. Increases the size of SQL text in `pg_stat_activity`. If left at the default value then queries longer than `1024` characters will not be collected. | -| `pg_stat_statements.track` | `ALL` | Optional. Enables tracking of statements within stored procedures and functions. | -| `pg_stat_statements.max` | `10000` | Optional. Increases the number of normalized queries tracked in `pg_stat_statements`. This setting is recommended for high-volume databases that see many different types of queries from many different clients. | -| `pg_stat_statements.track_utility` | `off` | Optional. Disables utility commands like PREPARE and EXPLAIN. Setting this value to `off` means only queries like SELECT, UPDATE, and DELETE are tracked. | -| `track_io_timing` | `on` | Optional. Enables collection of block read and write times for queries. | + +**Optional parameters** + +| Parameter | Value | Description | +| --- | --- | --- | +| `pg_stat_statements.track` | `ALL` | Enables tracking of statements within stored procedures and functions. | +| `pg_stat_statements.max` | `10000` | Increases the number of normalized queries tracked in `pg_stat_statements`. Recommended for high-volume databases that see many different types of queries from many different clients. | +| `pg_stat_statements.track_utility` | `off` | Disables utility commands like PREPARE and EXPLAIN. Setting this value to `off` means only queries like SELECT, UPDATE, and DELETE are tracked. | +| `track_io_timing` | `on` | Enables collection of block read and write times for queries. | ### Enable `auto_explain` (optional) By default, the agent only gathers [`EXPLAIN`][15] plans for a sampling of in-flight queries. These plans are of a more general nature, especially when application code uses prepared statements. -To collect full `EXPLAIN ANALYZE` plans taken from all queries, you need to use [`auto_explain`][16], a first-party extension bundled with PostgreSQL available in all major providers. _Logging collection is a prerequisite to `auto_explain` collection_, so be sure to enable it before continuing. +To collect full `EXPLAIN ANALYZE` plans taken from all queries, you need to use [`auto_explain`][16], a first-party extension bundled with PostgreSQL available in all major providers. _Logging collection is a prerequisite to `auto_explain` collection_, so enable it before continuing.
auto_explain produces log lines that may contain sensitive application data, similar to raw values in non-obfuscated SQL. Use the dbm_parameterized_queries_read permission to control access to the resulting plans. To restrict visibility of the log lines themselves—which are visible to all users in your Datadog organization by default—also configure RBAC for Logs. Datadog recommends using both permissions to protect sensitive information effectively.
@@ -94,11 +101,11 @@ To collect full `EXPLAIN ANALYZE` plans taken from all queries, you need to use
## Grant the Agent access
-The Datadog Agent requires read-only access to the database server in order to collect statistics and queries.
+The Datadog Agent requires read-only access to the database server to collect statistics and queries.
-The following SQL commands should be executed on the **primary** database server (the writer) in the cluster if Postgres is replicated. Choose a PostgreSQL database on the server for the Agent to connect to. The Agent can collect telemetry from all databases on the database server regardless of which one it connects to, so a good option is to use the default `postgres` database. Choose a different database only if you need the Agent to run [custom queries against data unique to that database][7].
+Run the following SQL commands on the **primary** database server (the writer) in the cluster if Postgres is replicated. The Agent can collect telemetry from all databases on the server regardless of which database it connects to. Use the default `postgres` database unless you need the Agent to run [custom queries against data unique to a different database][7].
-Connect to the chosen database as a superuser (or another user with sufficient permissions). For example, if your chosen database is `postgres`, connect as the `postgres` user using [psql][8] by running:
+Connect to your chosen database as a superuser (or another user with sufficient permissions). For example, to connect to the `postgres` database using [psql][8]:
```bash
psql -h mydb.example.com -d postgres -U postgres
@@ -110,7 +117,7 @@ Create the `datadog` user:
CREATE USER datadog WITH password 'SELECT permission on those tables to the datadog user. Example: grant SELECT on <TABLE_NAME> to datadog;. See PostgreSQL custom metric collection for more information.