You are viewing docs on Elastic's new documentation system, currently in technical preview. For all other Elastic docs, visit elastic.co/guide.

Detect metric anomalies

Detect and inspect memory usage and network traffic anomalies for hosts and Kubernetes pods.

Required role

The Editor role or higher is required to create machine learning jobs. To learn more, refer to Assign user roles and privileges.

You can create machine learning jobs to detect and inspect memory usage and network traffic anomalies for hosts and Kubernetes pods.

You can model system memory usage, along with inbound and outbound network traffic across hosts or pods. You can detect unusual increases in memory usage and unusually high inbound or outbound traffic across hosts or pods.

Enable machine learning jobs for hosts or Kubernetes pods

Create a machine learning job to detect anomalous memory usage and network traffic automatically.

After creating machine learning jobs, you cannot change the settings. You can recreate these jobs later. However, you will remove any previously detected anomalies.

  1. In your Observability project, go to InfrastructureInventory and click the Anomaly detection link at the top of the page.
  2. Under Hosts or Kubernetes Pods, click Enable to create a machine learning job.
  3. Choose a start date for the machine learning analysis. Machine learning jobs analyze the last four weeks of data and continue to run indefinitely.
  4. Select a partition field. Partitions allow you to create independent models for different groups of data that share similar behavior. For example, you may want to build separate models for machine type or cloud availability zone so that anomalies are not weighted equally across groups.
  5. By default, machine learning jobs analyze all of your metric data. You can filter this list to view only the jobs or metrics that you are interested in. For example, you can filter by job name and node name to view specific anomaly detection jobs for that host.
  6. Click Enable jobs.
  7. You're now ready to explore your metric anomalies. Click Anomalies.

The Anomalies table displays a list of each single metric anomaly detection job for the specific host or Kubernetes pod. By default, anomaly jobs are sorted by time to show the most recent job.

Along with each anomaly job and the node name, detected anomalies with a severity score equal to 50 or higher are listed. These scores represent a severity of "warning" or higher in the selected time period. The summary value represents the increase between the actual value and the expected ("typical") value of the metric in the anomaly record result.

To drill down and analyze the metric anomaly, select Actions → Open in Anomaly Explorer to view the Anomaly Explorer. You can also select ActionsShow in Inventory to view the host or Kubernetes pods Inventory page, filtered by the specific metric.

Note

These predefined anomaly detection jobs use custom rules. To update the rules in the Anomaly Explorer, select ActionsConfigure rules. The changes only take effect for new results. If you want to apply the changes to existing results, clone and rerun the job.

History chart

On the Inventory page, click Show history to view the metric values within the selected time frame. Detected anomalies with an anomaly score equal to 50 or higher are highlighted in red. To examine the detected anomalies, use the Anomaly Explorer.

On this page