Elasticbeats serves as a lightweight platform for data shippers that transfer information from the source to axiom and other tools based on the configuration. Before shipping data, beats collects metrics and logs from different sources, which later are deployed to your Axiom deployments.
There are different Elastic Beats you could use to ship logs and Axioms documentation provides a detailed step by step procedure on how to use each Beats.
You'd need to specify the org-id header if you are using Personal Token, it's best to use an API Token to avoid the need to specify the
Learn more about API and Personal Token
Note: You will get a
400 if you use the field name
_time when using the Elastic Search endpoint
Filebeat is a lightweight shipper for logs, it helps you centralize logs, files and can read files from your system.
Filebeats is useful for workloads, system, application log files, and data logs you would like to ingest to Axiom in some way.
In the logging case, it helps centralize logs and files in a structured pattern by reading from your various applications, services, workloads and VMs, then shipping to your Axiom deployments.
When downloading Filebeats, install the OSS version being that the non-oss version doesn't work with Axiom
Axiom lets you ingest data with the ElasticSearch bulk ingest API.
In order for Filebeat to work, index lifecycle management (ILM) must be disabled. To do so, add setup.ilm.enabled: false to the filebeat.yml configuration file.
setup.ilm.enabled: false filebeat.inputs: - type: log # Specify the path of the system log files to be sent to Axiom deployment. paths: - $PATH_TO_LOG_FILE output.elasticsearch: hosts: ['https://cloud.axiom.co:443/api/v1/datasets/$DATASET_NAME/elastic'] # token should be an API token api_key: 'axiom:$TOKEN'
Metricbeat is a lightweight shipper for metrics.
Metricbeat is installed on your systems and services and used for monitoring their performance, as well as different remote packages/utilities running on them.
metricbeat.modules: setup.ilm.enabled: false metricbeat.config.modules: path: -$PATH_TO_LOG_FILE metricbeat.modules: - module: system metricsets: - filesystem - cpu - load - fsstat - memory - network output.elasticsearch: hosts: ["https://cloud.axiom.co:443/api/v1/datasets/$DATASET_NAME/elastic"] # token should be an API token api_key: 'axiom:$TOKEN'
Winlogbeat is an open-source Windows specific event-log shipper that is installed as a Windows service. It can be used to collect and send event logs to Axiom.
Winlogbeat reads from one or more event logs using Windows APIs, filters the events based on user-configured criteria, then sends the event data to the configured outputs.
You can Capture:
- application events
- hardware events
- security events
- system events
- Extract the contents of the zip file into C:\Program Files.
- Rename the winlogbeat-$version directory to Winlogbeat
- Open a PowerShell prompt as an Administrator and run
PS C:\Users\Administrator> cd C:\Program Files\Winlogbeat PS C:\Program Files\Winlogbeat> .\install-service-winlogbeat.ps1
Configuration for Winlogbeat Service is found in the
winlogbeat.yml file in C:\Program Files\Winlogbeat.
Edit the winlogbeat.yml configuration file found in
C:\Program Files\Winlogbeat this will let you send data to Axiom.
What is winlogbeat.yml File? The winlogbeat.yml file contains the configuration on which windows events and service it should monitor and the time required.
winlogbeat.event_logs: - name: Application - name: System - name: Security logging.to_files: true logging.files: path: C:\ProgramData\Winlogbeat\Logs logging.level: info output.elasticsearch: hosts: ['https://cloud.axiom.co:443/api/v1/datasets/$DATASET_NAME/elastic'] # token should be an API token api_key: 'axiom:$TOKEN'
# Check if your configuration is correct PS C:\Program Files\Winlogbeat> .\winlogbeat.exe test config -c .\winlogbeat.yml -e
PS C:\Program Files\Winlogbeat> Start-Service winlogbeat
You can view the status of your service and control it from the Services management console in Windows.
To launch the management console, run this command:
PS C:\Program Files\Winlogbeat> services.msc
PS C:\Program Files\Winlogbeat> Stop-Service winlogbeat
For more information on Winlogbeat event logs visit the Winlogbeat documentation
Heartbeat is a lightweight shipper for uptime monitoring.
It Monitor your services, and ship response time to Axiom. It lets you check periodically check the status of your services and determine whether they are available.
Heartbeat is useful when you need to verify that you’re meeting your service level agreements for service uptime.
Heartbeat currently supports monitors for checking hosts via:
- ICMP (v4 and v6) Echo Requests. Use the
icmp monitorwhen you simply want to check whether a service is available. This monitor requires root access.
- TCP Use the tcp monitor to connect
via TCP.You can optionally configure this monitor to verify the endpoint by sending and/or receiving a custom payload.
- HTTP. Use the http monitor to connect
via HTTP.You can optionally configure this monitor to verify that the service returns the expected response, such as a specific status code, response header, or content.
Visit the Heartbeat download page to install Heartbeat on your system.
Heartbeat provides monitors to check the status of hosts at set intervals. Heartbeat currently provides monitors for ICMP, TCP, and HTTP.
You configure each monitor individually. In
heartbeat.yml, specify the list of monitors that you want to enable. Each item in the list begins with a dash (-).
The example below configures Heartbeat to use three monitors: an icmp monitor, a tcp monitor, and an http monitor. deployed instantly to Axiom.
# Disable index lifecycle management (ILM) setup.ilm.enabled: false heartbeat.monitors: - type: icmp schedule: '*/5 * * * * * *' hosts: ['myhost'] id: my-icmp-service name: My ICMP Service - type: tcp schedule: '@every 5s' hosts: ['myhost:12345'] mode: any id: my-tcp-service - type: http schedule: '@every 5s' urls: ['http://example.net'] service.name: apm-service-name id: my-http-service name: My HTTP Service output.elasticsearch: hosts: ['https://cloud.axiom.co:443/api/v1/datasets/$DATASET_NAME/elastic'] # token should be an API token api_key: 'axiom:$TOKEN'
Auditbeat is a lightweight shipper that ships events in real time to Axiom for further analysis. It Collects your Linux audit framework data and monitor the integrity of your files. It is also used to evaluate the activities of users and processes on your system.
You can also use Auditbeat to detect changes to critical files, like binaries and configuration files, and identify potential security policy violations.
Visit the Auditbeat download page to install Auditbeat on your system.
Auditbeat uses modules to collect audit information:
- File integrity
By default, Auditbeat uses a configuration that’s tailored to the operating system where Auditbeat is running.
To use a different configuration, change the module settings in
The example below configures Auditbeat to use the
file_integrity module configured to generate events whenever a file in one of the specified paths changes on disk. The events contains the file metadata and hashes, and it's deployed instantly to Axiom.
# Disable index lifecycle management (ILM) setup.ilm.enabled: false auditbeat.modules: - module: file_integrity paths: - /usr/bin - /sbin - /usr/sbin - /etc - /bin - /usr/local/sbin output.elasticsearch: hosts: ['https://cloud.axiom.co:443/api/v1/datasets/$DATASET_NAME/elastic'] # token should be an API token api_key: 'axiom:$TOKEN'
Packetbeat is a real-time network packet analyzer that you can integrate with Axiom to provide an application monitoring and performance analytics system between the servers of your network.
With Axiom you can use Packetbeat to capture the network traffic between your application servers, decode the application layer protocols (HTTP, MySQL, Redis, pgsql, thrift, mongodb and so on), and correlate the requests with the responses.
Packetbeat sniffs the traffic between your servers, and parses the application-level protocols on the fly directly into Axiom.
Currently, Packetbeat supports the following protocols:
- ICMP (v4 and v6)
- DHCP (v4)
- AMQP 0.9.1
- SIP/SDP (beta)
Visit the Packetbeat download page to install Packetbeat on your system.
packetbeat.yml, configure the network devices and protocols to capture traffic from.
To see a list of available devices for
packetbeat.yml configuration , run:
Packetbeat supports these sniffer types:
In the protocols section, configure the ports where Packetbeat can find each
protocol. If you use any non-standard ports, add them here. Otherwise, use the default values:
# Disable index lifecycle management (ILM) setup.ilm.enabled: false packetbeat.interfaces.auto_promisc_mode: true packetbeat.flows: timeout: 30s period: 10s protocols: dns: ports:  include_authorities: true include_additionals: true http: ports: [80, 8080, 8081, 5000, 8002] memcache: ports:  mysql: ports:  pgsql: ports:  redis: ports:  thrift: ports:  mongodb: ports:  output.elasticsearch: hosts: ['https://cloud.axiom.co:443/api/v1/datasets/$DATASET_NAME/elastic'] # api_key should be your API token api_key: 'axiom:$TOKEN'
For more information on configuring Packetbeats, visit the documentation
Journalbeat is a lightweight shipper for forwarding and centralizing log data from systemd journals to a log management tool like Axiom.
Journalbeat monitors the journal locations that you specify, collects log events, and eventually forwards the logs to Axiom.
Visit the Journalbeat download page to install Journalbeat on your system.
Before running Journalbeat, specify the location of the systemd journal files and configure how you want the files to be read.
The example below configures Journalbeat to use the
path of your systemd journal files. Each path can be a directory path (to collect events from all journals in a directory), or a file path configured to deploy logs instantly to Axiom.
# Disable index lifecycle management (ILM) setup.ilm.enabled: false journalbeat.inputs: - paths: - "/dev/log" - "/var/log/messages/my-journal-file.journal" seek: head journalbeat.inputs: - paths:  include_matches: - "CONTAINER_TAG=redis" - "_COMM=redis" - "container.image.tag=redis" - "process.name=redis" output.elasticsearch: hosts: ['https://cloud.axiom.co:443/api/v1/datasets/$DATASET_NAME/elastic'] # token should be an API token api_key: 'axiom:$TOKEN'
For more information on configuring Journalbeat, visit the documentation