Collect logs and metrics for Amazon Elastic Compute Cloud service with Elastic Agent
What is an Elastic integration?
This integration is powered by Elastic Agent. Elastic Agent is a single, unified way to add monitoring for logs, metrics, and other types of data to a host. It can also protect hosts from security threats, query data from operating systems, forward data from remote services or hardware, and more. Refer to our documentation for a detailed comparison between Beats and Elastic Agent.
Prefer to use Beats for this use case? See Filebeat modules for logs or Metricbeat modules for metrics.
See the integrations quick start guides to get started:
The Amazon EC2 integration allows you to monitor Amazon Elastic Compute Cloud (Amazon EC2)—a cloud compute platform.
Use the Amazon EC2 integration to collect logs and metrics related to your EC2 instances. Then visualize that data in Kibana, create alerts to notify you if something goes wrong, and reference the logs and metrics when troubleshooting an issue.
For example, you could use this data to track Amazon EC2 CPU utilization. Then you can alert when utilization for an instance crosses a predefined threshold.
IMPORTANT: Extra AWS charges on AWS API requests will be generated by this integration. Please refer to the AWS integration for more details.
The Amazon EC2 integration collects two types of data: logs and metrics.
Logs help you keep a record of events happening in Amazon EC2. Logs collected by the Amazon EC2 integration include the region in which an instance is running, the operating system architecture, container information, and more. See more details in the Logs reference.
Metrics give you insight into the state of your Amazon EC2 instances. Metrics collected by the Amazon EC2 integration include the Amazon EC2 instance ID, the number of earned CPU credits that an instance has accrued since it was launched or started, and more. See more details in the Metrics reference.
You need Elasticsearch for storing and searching your data and Kibana for visualizing and managing it. You can use our hosted Elasticsearch Service on Elastic Cloud, which is recommended, or self-manage the Elastic Stack on your own hardware.
Before using any AWS integration you will need:
For more details about these requirements, see the AWS integration documentation.
Use this integration if you only need to collect data from the Amazon EC2 service.
If you want to collect data from two or more AWS services, consider using the AWS integration. When you configure the AWS integration, you can collect data from as many AWS services as you'd like.
For step-by-step instructions on how to set up an integration, see the Getting started guide.
The ec2
data stream supports both EC2 logs stored in AWS CloudWatch and EC2 logs stored in Amazon S3.
For logs stored in S3, you must export logs from log groups to an Amazon S3 bucket which has SQS notification setup already.
With this data stream, EC2 logs will be parsed into fields like ip_address
and process.name
. For logs from other services, please use the AWS CloudWatch integration.
Exported fields
Field | Description | Type |
---|---|---|
@timestamp | Event timestamp. | date |
aws.ec2.ip_address | The internet address of the requester. | keyword |
cloud.account.id | The cloud account or organization id used to identify different entities in a multi-tenant environment. Examples: AWS account id, Google Cloud ORG Id, or other unique identifier. | keyword |
cloud.availability_zone | Availability zone in which this host, resource, or service is located. | keyword |
cloud.image.id | Image ID for the cloud instance. | keyword |
cloud.instance.id | Instance ID of the host machine. | keyword |
cloud.instance.name | Instance name of the host machine. | keyword |
cloud.machine.type | Machine type of the host machine. | keyword |
cloud.project.id | The cloud project identifier. Examples: Google Cloud Project id, Azure Project id. | keyword |
cloud.provider | Name of the cloud provider. Example values are aws, azure, gcp, or digitalocean. | keyword |
cloud.region | Region in which this host, resource, or service is located. | keyword |
container.id | Unique container id. | keyword |
container.image.name | Name of the image the container was built on. | keyword |
container.labels | Image labels. | object |
container.name | Container name. | keyword |
data_stream.dataset | Data stream dataset. | constant_keyword |
data_stream.namespace | Data stream namespace. | constant_keyword |
data_stream.type | Data stream type. | constant_keyword |
ecs.version | ECS version this event conforms to. ecs.version is a required field and must exist in all events. When querying across multiple indices -- which may conform to slightly different ECS versions -- this field lets integrations adjust to the schema version of the events. | keyword |
error.message | Error message. | match_only_text |
event.dataset | Event dataset | constant_keyword |
event.module | Event module | constant_keyword |
host.architecture | Operating system architecture. | keyword |
host.containerized | If the host is a container. | boolean |
host.domain | Name of the domain of which the host is a member. For example, on Windows this could be the host's Active Directory domain or NetBIOS domain name. For Linux this could be the domain of the host's LDAP provider. | keyword |
host.hostname | Hostname of the host. It normally contains what the hostname command returns on the host machine. | keyword |
host.id | Unique host id. As hostname is not always unique, use values that are meaningful in your environment. Example: The current usage of beat.name . | keyword |
host.ip | Host ip addresses. | ip |
host.mac | Host MAC addresses. The notation format from RFC 7042 is suggested: Each octet (that is, 8-bit byte) is represented by two [uppercase] hexadecimal digits giving the value of the octet as an unsigned integer. Successive octets are separated by a hyphen. | keyword |
host.name | Name of the host. It can contain what hostname returns on Unix systems, the fully qualified domain name, or a name specified by the user. The sender decides which value to use. | keyword |
host.os.build | OS build information. | keyword |
host.os.codename | OS codename, if any. | keyword |
host.os.family | OS family (such as redhat, debian, freebsd, windows). | keyword |
host.os.kernel | Operating system kernel version as a raw string. | keyword |
host.os.name | Operating system name, without the version. | keyword |
host.os.name.text | Multi-field of host.os.name . | match_only_text |
host.os.platform | Operating system platform (such centos, ubuntu, windows). | keyword |
host.os.version | Operating system version as a raw string. | keyword |
host.type | Type of host. For Cloud providers this can be the machine type like t2.medium . If vm, this could be the container, for example, or other information meaningful in your environment. | keyword |
message | For log events the message field contains the log message, optimized for viewing in a log viewer. For structured logs without an original message field, other fields can be concatenated to form a human-readable summary of the event. If multiple messages exist, they can be combined into one message. | match_only_text |
process.name | Process name. | keyword |
tags | List of keywords used to tag each event. | keyword |
An example event for ec2
looks as following:
{
"data_stream": {
"namespace": "default",
"type": "logs",
"dataset": "aws.ec2_logs"
},
"process": {
"name": "systemd"
},
"@timestamp": "2020-02-20T07:01:01.000Z",
"ecs": {
"version": "8.0.0"
},
"event": {
"ingested": "2021-07-19T21:47:04.871450600Z",
"original": "2020-02-20T07:01:01.000Z Feb 20 07:01:01 ip-172-31-81-156 systemd: Stopping User Slice of root."
},
"aws": {
"ec2": {
"ip_address": "ip-172-31-81-156"
}
},
"message": "Stopping User Slice of root.",
"tags": [
"preserve_original_event"
]
}
An example event for ec2
looks as following:
{
"@timestamp": "2022-05-26T12:40:00.000Z",
"agent": {
"name": "docker-fleet-agent",
"id": "24d50340-a9d0-4d5d-9f42-fe9cb4b8c95d",
"type": "metricbeat",
"ephemeral_id": "f8282deb-ebc7-4d1f-9386-207f56657244",
"version": "8.2.0"
},
"elastic_agent": {
"id": "24d50340-a9d0-4d5d-9f42-fe9cb4b8c95d",
"version": "8.2.0",
"snapshot": false
},
"cloud": {
"availability_zone": "us-east-1c",
"instance": {
"name": "elastic-package-test-33138",
"id": "i-0de58890d94dda2e3"
},
"provider": "aws",
"machine": {
"type": "t1.micro"
},
"region": "us-east-1",
"account": {
"name": "elastic-beats",
"id": "428152502467"
}
},
"ecs": {
"version": "8.0.0"
},
"service": {
"type": "aws"
},
"data_stream": {
"namespace": "default",
"type": "metrics",
"dataset": "aws.ec2_metrics"
},
"host": {
"hostname": "docker-fleet-agent",
"disk": {
"read": {
"bytes": 0
},
"write": {
"bytes": 0
}
},
"os": {
"kernel": "5.10.47-linuxkit",
"codename": "focal",
"name": "Ubuntu",
"type": "linux",
"family": "debian",
"version": "20.04.4 LTS (Focal Fossa)",
"platform": "ubuntu"
},
"containerized": true,
"ip": [
"192.168.80.7"
],
"name": "elastic-package-test-33138",
"cpu": {
"usage": 0.08265027322397175
},
"id": "i-0de58890d94dda2e3",
"mac": [
"02:42:c0:a8:50:07"
],
"architecture": "x86_64",
"network": {
"ingress": {
"bytes": 1992
},
"egress": {
"bytes": 1800
}
}
},
"metricset": {
"period": 300000,
"name": "cloudwatch"
},
"event": {
"duration": 15936070000,
"agent_id_status": "verified",
"ingested": "2022-05-26T12:44:52Z",
"module": "aws",
"dataset": "aws.ec2_metrics"
},
"aws": {
"ec2": {
"diskio": {
"read": {
"count_per_sec": 0,
"bytes_per_sec": 0
},
"write": {
"count_per_sec": 0,
"bytes_per_sec": 0
}
},
"instance": {
"image": {
"id": "ami-0ff900168d0231cd3"
},
"core": {
"count": 1
},
"private": {
"ip": "172.31.16.45",
"dns_name": "ip-172-31-16-45.ec2.internal"
},
"threads_per_core": 1,
"public": {
"ip": "54.90.153.147",
"dns_name": "ec2-54-90-153-147.compute-1.amazonaws.com"
},
"state": {
"code": 16,
"name": "running"
},
"monitoring": {
"state": "enabled"
}
},
"cpu": {
"credit_balance": 144,
"credit_usage": 0.004121
},
"status": {
"check_failed": 0,
"check_failed_instance": 0,
"check_failed_system": 0
},
"network": {
"in": {
"bytes_per_sec": 33.2
},
"out": {
"bytes_per_sec": 30
}
}
},
"cloudwatch": {
"namespace": "AWS/EC2"
},
"dimensions": {
"InstanceId": "i-0de58890d94dda2e3"
},
"tags": {
"Name": "elastic-package-test-33138"
}
}
}
Exported fields
Field | Description | Type |
---|---|---|
@timestamp | Event timestamp. | date |
aws.*.metrics.*.* | Metrics that returned from Cloudwatch API query. | object |
aws.cloudwatch.namespace | The namespace specified when query cloudwatch api. | keyword |
aws.dimensions.* | Metric dimensions. | object |
aws.dimensions.AutoScalingGroupName | An Auto Scaling group is a collection of instances you define if you're using Auto Scaling. | keyword |
aws.dimensions.ImageId | This dimension filters the data you request for all instances running this Amazon EC2 Amazon Machine Image (AMI) | keyword |
aws.dimensions.InstanceId | Amazon EC2 instance ID | keyword |
aws.dimensions.InstanceType | This dimension filters the data you request for all instances running with this specified instance type. | keyword |
aws.ec2.cpu.credit_balance | The number of earned CPU credits that an instance has accrued since it was launched or started. | long |
aws.ec2.cpu.credit_usage | The number of CPU credits spent by the instance for CPU utilization. | long |
aws.ec2.cpu.surplus_credit_balance | The number of surplus credits that have been spent by an unlimited instance when its CPUCreditBalance value is zero. | long |
aws.ec2.cpu.surplus_credits_charged | The number of spent surplus credits that are not paid down by earned CPU credits, and which thus incur an additional charge. | long |
aws.ec2.diskio.read.bytes_per_sec | Bytes read per second from all instance store volumes available to the instance. | long |
aws.ec2.diskio.read.count_per_sec | Completed read operations per second from all instance store volumes available to the instance in a specified period of time. | long |
aws.ec2.diskio.write.bytes_per_sec | Bytes written per second to all instance store volumes available to the instance. | long |
aws.ec2.diskio.write.count_per_sec | Completed write operations per second to all instance store volumes available to the instance in a specified period of time. | long |
aws.ec2.instance.core.count | The number of CPU cores for the instance. | integer |
aws.ec2.instance.image.id | The ID of the image used to launch the instance. | keyword |
aws.ec2.instance.monitoring.state | Indicates whether detailed monitoring is enabled. | keyword |
aws.ec2.instance.private.dns_name | The private DNS name of the network interface. | keyword |
aws.ec2.instance.private.ip | The private IPv4 address associated with the network interface. | ip |
aws.ec2.instance.public.dns_name | The public DNS name of the instance. | keyword |
aws.ec2.instance.public.ip | The address of the Elastic IP address (IPv4) bound to the network interface. | ip |
aws.ec2.instance.state.code | The state of the instance, as a 16-bit unsigned integer. | integer |
aws.ec2.instance.state.name | The state of the instance (pending | running |
aws.ec2.instance.threads_per_core | The number of threads per CPU core. | integer |
aws.ec2.network.in.bytes_per_sec | The number of bytes per second received on all network interfaces by the instance. | long |
aws.ec2.network.in.packets_per_sec | The number of packets per second sent out on all network interfaces by the instance. | long |
aws.ec2.network.out.bytes_per_sec | The number of bytes per second sent out on all network interfaces by the instance. | long |
aws.ec2.network.out.packets_per_sec | The number of packets per second sent out on all network interfaces by the instance. | long |
aws.ec2.status.check_failed | Reports whether the instance has passed both the instance status check and the system status check in the last minute. | long |
aws.ec2.status.check_failed_instance | Reports whether the instance has passed the instance status check in the last minute. | long |
aws.ec2.status.check_failed_system | Reports whether the instance has passed the system status check in the last minute. | long |
aws.s3.bucket.name | Name of a S3 bucket. | keyword |
aws.tags.* | Tag key value pairs from aws resources. | object |
cloud | Fields related to the cloud or infrastructure the events are coming from. | group |
cloud.account.id | The cloud account or organization id used to identify different entities in a multi-tenant environment. Examples: AWS account id, Google Cloud ORG Id, or other unique identifier. | keyword |
cloud.account.name | The cloud account name or alias used to identify different entities in a multi-tenant environment. Examples: AWS account name, Google Cloud ORG display name. | keyword |
cloud.availability_zone | Availability zone in which this host, resource, or service is located. | keyword |
cloud.image.id | Image ID for the cloud instance. | keyword |
cloud.instance.id | Instance ID of the host machine. | keyword |
cloud.machine.type | Machine type of the host machine. | keyword |
cloud.provider | Name of the cloud provider. Example values are aws, azure, gcp, or digitalocean. | keyword |
cloud.region | Region in which this host, resource, or service is located. | keyword |
data_stream.dataset | Data stream dataset. | constant_keyword |
data_stream.namespace | Data stream namespace. | constant_keyword |
data_stream.type | Data stream type. | constant_keyword |
ecs.version | ECS version this event conforms to. ecs.version is a required field and must exist in all events. When querying across multiple indices -- which may conform to slightly different ECS versions -- this field lets integrations adjust to the schema version of the events. | keyword |
error | These fields can represent errors of any kind. Use them for errors that happen while fetching events or in cases where the event itself contains an error. | group |
error.message | Error message. | match_only_text |
event.dataset | Event dataset | constant_keyword |
event.module | Event module | constant_keyword |
host.containerized | If the host is a container. | boolean |
host.cpu.usage | Percent CPU used which is normalized by the number of CPU cores and it ranges from 0 to 1. Scaling factor: 1000. For example: For a two core host, this value should be the average of the two cores, between 0 and 1. | scaled_float |
host.disk.read.bytes | The total number of bytes (gauge) read successfully (aggregated from all disks) since the last metric collection. | long |
host.disk.write.bytes | The total number of bytes (gauge) written successfully (aggregated from all disks) since the last metric collection. | long |
host.network.egress.bytes | The number of bytes (gauge) sent out on all network interfaces by the host since the last metric collection. | long |
host.network.egress.packets | The number of packets (gauge) sent out on all network interfaces by the host since the last metric collection. | long |
host.network.ingress.bytes | The number of bytes received (gauge) on all network interfaces by the host since the last metric collection. | long |
host.network.ingress.packets | The number of packets (gauge) received on all network interfaces by the host since the last metric collection. | long |
host.os.build | OS build information. | keyword |
host.os.codename | OS codename, if any. | keyword |
service.type | The type of the service data is collected from. The type can be used to group and correlate logs and metrics from one service type. Example: If logs or metrics are collected from Elasticsearch, service.type would be elasticsearch . | keyword |