mirror of
https://github.com/elastic/kibana.git
synced 2025-04-24 01:38:56 -04:00
* [DOC] Updated documentation for newly added monitoring alerts (#91272) * Documentation for recently added alerts * [DOCS] Fixes broken link * Addressed review feedback Co-authored-by: lcawl <lcawley@elastic.co> Co-authored-by: Kibana Machine <42973632+kibanamachine@users.noreply.github.com> # Conflicts: # docs/user/monitoring/cluster-alerts.asciidoc * [DOC] Removes broken link Co-authored-by: Kibana Machine <42973632+kibanamachine@users.noreply.github.com> Co-authored-by: lcawl <lcawley@elastic.co>
This commit is contained in:
parent
297c3b44a8
commit
e95faff24a
10 changed files with 79 additions and 80 deletions
|
@ -192,7 +192,8 @@ The maximum height that a cell occupies in a table. Set to 0 to disable
|
|||
truncation.
|
||||
|
||||
[[xpack-defaultadminemail]]`xPack:defaultAdminEmail`::
|
||||
**Deprecated. Use <<cluster-alert-email-notifications,Email Notifications>> instead.**
|
||||
**Deprecated. Instead, use {stack-manage-app} > {alerts-ui} to review and
|
||||
modify all the available alerts.**
|
||||
Email address for {xpack} admin operations, such as cluster alert notifications
|
||||
from *{stack-monitor-app}*.
|
||||
|
||||
|
|
|
@ -37,11 +37,6 @@ For more information, see
|
|||
monitoring back-end does not run and {kib} stats are not sent to the monitoring
|
||||
cluster.
|
||||
|
||||
a|`monitoring.cluster_alerts.`
|
||||
`email_notifications.email_address` {ess-icon}
|
||||
| Specifies the email address where you want to receive cluster alerts.
|
||||
See <<cluster-alert-email-notifications, email notifications>> for details.
|
||||
|
||||
| `monitoring.ui.elasticsearch.hosts`
|
||||
| Specifies the location of the {es} cluster where your monitoring data is stored.
|
||||
By default, this is the same as <<elasticsearch-hosts, `elasticsearch.hosts`>>. This setting enables
|
||||
|
|
|
@ -1,64 +0,0 @@
|
|||
[role="xpack"]
|
||||
[[cluster-alerts]]
|
||||
= Cluster Alerts
|
||||
|
||||
The *Stack Monitoring > Clusters* page in {kib} summarizes the status of your
|
||||
{stack}. You can drill down into the metrics to view more information about your
|
||||
cluster and specific nodes, instances, and indices.
|
||||
|
||||
The Top Cluster Alerts shown on the Clusters page notify you of
|
||||
conditions that require your attention:
|
||||
|
||||
* {es} Cluster Health Status is Yellow (missing at least one replica)
|
||||
or Red (missing at least one primary).
|
||||
* {es} Version Mismatch. You have {es} nodes with
|
||||
different versions in the same cluster.
|
||||
* {kib} Version Mismatch. You have {kib} instances with different
|
||||
versions running against the same {es} cluster.
|
||||
* Logstash Version Mismatch. You have Logstash nodes with different
|
||||
versions reporting stats to the same monitoring cluster.
|
||||
* {es} Nodes Changed. You have {es} nodes that were recently added or removed.
|
||||
* {es} License Expiration. The cluster's license is about to expire.
|
||||
+
|
||||
--
|
||||
If you do not preserve the data directory when upgrading a {kib} or
|
||||
Logstash node, the instance is assigned a new persistent UUID and shows up
|
||||
as a new instance
|
||||
--
|
||||
* {xpack} License Expiration. When the {xpack} license expiration date
|
||||
approaches, you will get notifications with a severity level relative to how
|
||||
soon the expiration date is:
|
||||
** 60 days: Informational alert
|
||||
** 30 days: Low-level alert
|
||||
** 15 days: Medium-level alert
|
||||
** 7 days: Severe-level alert
|
||||
+
|
||||
The 60-day and 30-day thresholds are skipped for Trial licenses, which are only
|
||||
valid for 30 days.
|
||||
|
||||
The {monitor-features} check the cluster alert conditions every minute. Cluster
|
||||
alerts are automatically dismissed when the condition is resolved.
|
||||
|
||||
NOTE: {watcher} must be enabled to view cluster alerts. If you have a Basic
|
||||
license, Top Cluster Alerts are not displayed.
|
||||
|
||||
[float]
|
||||
[[cluster-alert-email-notifications]]
|
||||
== Email Notifications
|
||||
To receive email notifications for the Cluster Alerts:
|
||||
|
||||
1. Configure an email account as described in
|
||||
{ref}/actions-email.html#configuring-email[Configuring email accounts].
|
||||
2. Configure the
|
||||
`monitoring.cluster_alerts.email_notifications.email_address` setting in
|
||||
`kibana.yml` with your email address.
|
||||
+
|
||||
--
|
||||
TIP: If you have separate production and monitoring clusters and separate {kib}
|
||||
instances for those clusters, you must put the
|
||||
`monitoring.cluster_alerts.email_notifications.email_address` setting in
|
||||
the {kib} instance that is associated with the production cluster.
|
||||
|
||||
--
|
||||
|
||||
Email notifications are sent only when Cluster Alerts are triggered and resolved.
|
|
@ -1,6 +1,5 @@
|
|||
include::xpack-monitoring.asciidoc[]
|
||||
include::beats-details.asciidoc[leveloffset=+1]
|
||||
include::cluster-alerts.asciidoc[leveloffset=+1]
|
||||
include::elasticsearch-details.asciidoc[leveloffset=+1]
|
||||
include::kibana-alerts.asciidoc[leveloffset=+1]
|
||||
include::kibana-details.asciidoc[leveloffset=+1]
|
||||
|
|
|
@ -29,7 +29,7 @@ To review and modify all the available alerts, use
|
|||
This alert is triggered when a node runs a consistently high CPU load. By
|
||||
default, the trigger condition is set at 85% or more averaged over the last 5
|
||||
minutes. The alert is grouped across all the nodes of the cluster by running
|
||||
checks on a schedule time of 1 minute with a re-notify internal of 1 day.
|
||||
checks on a schedule time of 1 minute with a re-notify interval of 1 day.
|
||||
|
||||
[discrete]
|
||||
[[kibana-alerts-disk-usage-threshold]]
|
||||
|
@ -38,7 +38,7 @@ checks on a schedule time of 1 minute with a re-notify internal of 1 day.
|
|||
This alert is triggered when a node is nearly at disk capacity. By
|
||||
default, the trigger condition is set at 80% or more averaged over the last 5
|
||||
minutes. The alert is grouped across all the nodes of the cluster by running
|
||||
checks on a schedule time of 1 minute with a re-notify internal of 1 day.
|
||||
checks on a schedule time of 1 minute with a re-notify interval of 1 day.
|
||||
|
||||
[discrete]
|
||||
[[kibana-alerts-jvm-memory-threshold]]
|
||||
|
@ -47,7 +47,7 @@ checks on a schedule time of 1 minute with a re-notify internal of 1 day.
|
|||
This alert is triggered when a node runs a consistently high JVM memory usage. By
|
||||
default, the trigger condition is set at 85% or more averaged over the last 5
|
||||
minutes. The alert is grouped across all the nodes of the cluster by running
|
||||
checks on a schedule time of 1 minute with a re-notify internal of 1 day.
|
||||
checks on a schedule time of 1 minute with a re-notify interval of 1 day.
|
||||
|
||||
[discrete]
|
||||
[[kibana-alerts-missing-monitoring-data]]
|
||||
|
@ -56,7 +56,72 @@ checks on a schedule time of 1 minute with a re-notify internal of 1 day.
|
|||
This alert is triggered when any stack products nodes or instances stop sending
|
||||
monitoring data. By default, the trigger condition is set to missing for 15 minutes
|
||||
looking back 1 day. The alert is grouped across all the nodes of the cluster by running
|
||||
checks on a schedule time of 1 minute with a re-notify internal of 6 hours.
|
||||
checks on a schedule time of 1 minute with a re-notify interval of 6 hours.
|
||||
|
||||
[discrete]
|
||||
[[kibana-alerts-thread-pool-rejections]]
|
||||
== Thread pool rejections (search/write)
|
||||
|
||||
This alert is triggered when a node experiences thread pool rejections. By
|
||||
default, the trigger condition is set at 300 or more over the last 5
|
||||
minutes. The alert is grouped across all the nodes of the cluster by running
|
||||
checks on a schedule time of 1 minute with a re-notify interval of 1 day.
|
||||
Thresholds can be set independently for `search` and `write` type rejections.
|
||||
|
||||
[discrete]
|
||||
[[kibana-alerts-ccr-read-exceptions]]
|
||||
== CCR read exceptions
|
||||
|
||||
This alert is triggered if a read exception has been detected on any of the
|
||||
replicated clusters. The trigger condition is met if 1 or more read exceptions
|
||||
are detected in the last hour. The alert is grouped across all replicated clusters
|
||||
by running checks on a schedule time of 1 minute with a re-notify interval of 6 hours.
|
||||
|
||||
[discrete]
|
||||
[[kibana-alerts-large-shard-size]]
|
||||
== Large shard size
|
||||
|
||||
This alert is triggered if a large (primary) shard size is found on any of the
|
||||
specified index patterns. The trigger condition is met if an index's shard size is
|
||||
55gb or higher in the last 5 minutes. The alert is grouped across all indices that match
|
||||
the default patter of `*` by running checks on a schedule time of 1 minute with a re-notify
|
||||
interval of 12 hours.
|
||||
|
||||
[discrete]
|
||||
[[kibana-alerts-cluster-alerts]]
|
||||
== Cluster alerts
|
||||
|
||||
These alerts summarize the current status of your {stack}. You can drill down into the metrics
|
||||
to view more information about your cluster and specific nodes, instances, and indices.
|
||||
|
||||
An alert will be triggered if any of the following conditions are met within the last minute:
|
||||
|
||||
* {es} cluster health status is yellow (missing at least one replica)
|
||||
or red (missing at least one primary).
|
||||
* {es} version mismatch. You have {es} nodes with
|
||||
different versions in the same cluster.
|
||||
* {kib} version mismatch. You have {kib} instances with different
|
||||
versions running against the same {es} cluster.
|
||||
* Logstash version mismatch. You have Logstash nodes with different
|
||||
versions reporting stats to the same monitoring cluster.
|
||||
* {es} nodes changed. You have {es} nodes that were recently added or removed.
|
||||
* {es} license expiration. The cluster's license is about to expire.
|
||||
+
|
||||
--
|
||||
If you do not preserve the data directory when upgrading a {kib} or
|
||||
Logstash node, the instance is assigned a new persistent UUID and shows up
|
||||
as a new instance
|
||||
--
|
||||
* Subscription license expiration. When the expiration date
|
||||
approaches, you will get notifications with a severity level relative to how
|
||||
soon the expiration date is:
|
||||
** 60 days: Informational alert
|
||||
** 30 days: Low-level alert
|
||||
** 15 days: Medium-level alert
|
||||
** 7 days: Severe-level alert
|
||||
+
|
||||
The 60-day and 30-day thresholds are skipped for Trial licenses, which are only
|
||||
valid for 30 days.
|
||||
|
||||
NOTE: Some action types are subscription features, while others are free.
|
||||
For a comparison of the Elastic subscription levels, see the alerting section of
|
||||
|
|
|
@ -201,12 +201,15 @@ export class DocLinksService {
|
|||
guide: `${ELASTIC_WEBSITE_URL}guide/en/kibana/${DOC_LINK_VERSION}/maps.html`,
|
||||
},
|
||||
monitoring: {
|
||||
alertsCluster: `${ELASTIC_WEBSITE_URL}guide/en/kibana/${DOC_LINK_VERSION}/cluster-alerts.html`,
|
||||
alertsKibana: `${ELASTIC_WEBSITE_URL}guide/en/kibana/${DOC_LINK_VERSION}/kibana-alerts.html`,
|
||||
alertsKibanaCpuThreshold: `${ELASTIC_WEBSITE_URL}guide/en/kibana/${DOC_LINK_VERSION}/kibana-alerts.html#kibana-alerts-cpu-threshold`,
|
||||
alertsKibanaDiskThreshold: `${ELASTIC_WEBSITE_URL}guide/en/kibana/${DOC_LINK_VERSION}/kibana-alerts.html#kibana-alerts-disk-usage-threshold`,
|
||||
alertsKibanaJvmThreshold: `${ELASTIC_WEBSITE_URL}guide/en/kibana/${DOC_LINK_VERSION}/kibana-alerts.html#kibana-alerts-jvm-memory-threshold`,
|
||||
alertsKibanaMissingData: `${ELASTIC_WEBSITE_URL}guide/en/kibana/${DOC_LINK_VERSION}/kibana-alerts.html#kibana-alerts-missing-monitoring-data`,
|
||||
alertsKibanaThreadpoolRejections: `${ELASTIC_WEBSITE_URL}guide/en/kibana/${DOC_LINK_VERSION}/kibana-alerts.html#kibana-alerts-thread-pool-rejections`,
|
||||
alertsKibanaCCRReadExceptions: `${ELASTIC_WEBSITE_URL}guide/en/kibana/${DOC_LINK_VERSION}/kibana-alerts.html#kibana-alerts-ccr-read-exceptions`,
|
||||
alertsKibanaLargeShardSize: `${ELASTIC_WEBSITE_URL}guide/en/kibana/${DOC_LINK_VERSION}/kibana-alerts.html#kibana-alerts-large-shard-size`,
|
||||
alertsKibanaClusterAlerts: `${ELASTIC_WEBSITE_URL}guide/en/kibana/${DOC_LINK_VERSION}/kibana-alerts.html#kibana-alerts-cluster-alerts`,
|
||||
metricbeatBlog: `${ELASTIC_WEBSITE_URL}blog/external-collection-for-elastic-stack-monitoring-is-now-available-via-metricbeat`,
|
||||
monitorElasticsearch: `${ELASTICSEARCH_DOCS}configuring-metricbeat.html`,
|
||||
monitorKibana: `${ELASTIC_WEBSITE_URL}guide/en/kibana/${DOC_LINK_VERSION}/monitoring-metricbeat.html`,
|
||||
|
|
|
@ -42,7 +42,7 @@ export function createCCRReadExceptionsAlertType(): AlertTypeModel<ValidateOptio
|
|||
description: ALERT_DETAILS[ALERT_CCR_READ_EXCEPTIONS].description,
|
||||
iconClass: 'bell',
|
||||
documentationUrl(docLinks) {
|
||||
return `${docLinks.links.monitoring.alertsKibana}`;
|
||||
return `${docLinks.links.monitoring.alertsKibanaCCRReadExceptions}`;
|
||||
},
|
||||
alertParamsExpression: (props: Props) => (
|
||||
<Expression {...props} paramDetails={ALERT_DETAILS[ALERT_CCR_READ_EXCEPTIONS].paramDetails} />
|
||||
|
|
|
@ -42,7 +42,7 @@ export function createLargeShardSizeAlertType(): AlertTypeModel<ValidateOptions>
|
|||
description: ALERT_DETAILS[ALERT_LARGE_SHARD_SIZE].description,
|
||||
iconClass: 'bell',
|
||||
documentationUrl(docLinks) {
|
||||
return `${docLinks.links.monitoring.alertsKibana}`;
|
||||
return `${docLinks.links.monitoring.alertsKibanaLargeShardSize}`;
|
||||
},
|
||||
alertParamsExpression: (props: Props) => (
|
||||
<Expression {...props} paramDetails={ALERT_DETAILS[ALERT_LARGE_SHARD_SIZE].paramDetails} />
|
||||
|
|
|
@ -23,7 +23,7 @@ export function createLegacyAlertTypes(): AlertTypeModel[] {
|
|||
description: LEGACY_ALERT_DETAILS[legacyAlert].description,
|
||||
iconClass: 'bell',
|
||||
documentationUrl(docLinks) {
|
||||
return `${docLinks.links.monitoring.alertsCluster}`;
|
||||
return `${docLinks.links.monitoring.alertsKibanaClusterAlerts}`;
|
||||
},
|
||||
alertParamsExpression: () => (
|
||||
<Fragment>
|
||||
|
|
|
@ -33,7 +33,7 @@ export function createThreadPoolRejectionsAlertType(
|
|||
description: threadPoolAlertDetails.description,
|
||||
iconClass: 'bell',
|
||||
documentationUrl(docLinks) {
|
||||
return `${docLinks.links.monitoring.alertsKibana}`;
|
||||
return `${docLinks.links.monitoring.alertsKibanaThreadpoolRejections}`;
|
||||
},
|
||||
alertParamsExpression: (props: Props) => (
|
||||
<>
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue