Quick start: modules for common log formats

edit

Quick start: modules for common log formats

edit

Filebeat provides a set of pre-built modules that you can use to rapidly implement and deploy a log monitoring solution, complete with sample dashboards and data visualizations (when available), in about 5 minutes. These modules support common log formats, such as Nginx, Apache2, and MySQL, and can be run by issuing a simple command.

This topic shows you how to run the basic modules with minimal extra configuration. For detailed documentation and the full list of available modules, see Modules.

Can’t find a module for your log file type? Follow the numbered steps under Getting Started With Filebeat to set up and configure Filebeat manually.

Prerequisites

edit

Before running Filebeat modules:

  • Install and configure the Elastic stack. See Getting started with the Elastic Stack.
  • Complete the Filebeat installation instructions described in Step 1: Install Filebeat. After installing Filebeat, return to this quick start page.
  • Verify that Elasticsearch and Kibana are running and that Elasticsearch is ready to receive data from Filebeat.

Running Filebeat modules

edit

To set up and run Filebeat modules:

  1. In the filebeat.yml config file, set the location of the Elasticsearch installation. By default, Filebeat assumes Elasticsearch is running locally on port 9200.

    • If you’re running our hosted Elasticsearch Service on Elastic Cloud, specify your Cloud ID. For example:

      cloud.id: "staging:dXMtZWFzdC0xLmF3cy5mb3VuZC5pbyRjZWM2ZjI2MWE3NGJmMjRjZTMzYmI4ODExYjg0Mjk0ZiRjNmMyY2E2ZDA0MjI0OWFmMGNjN2Q3YTllOTYyNTc0Mw=="
    • If you’re running Elasticsearch on your own hardware, set the host and port where Filebeat can find the Elasticsearch installation. For example:

      output.elasticsearch:
        hosts: ["myEShost:9200"]
  2. If Elasticsearch and Kibana are secured, set credentials in the filebeat.yml config file before you run the commands that set up and start Filebeat.

    • If you’re running our hosted Elasticsearch Service on Elastic Cloud, specify your cloud auth credentials. For example:

      cloud.auth: "elastic:YOUR_PASSWORD"
    • If you’re running Elasticsearch on your own hardware, specify your Elasticsearch and Kibana credentials:

      output.elasticsearch:
        hosts: ["myEShost:9200"]
        username: "filebeat_internal"
        password: "YOUR_PASSWORD" 
      setup.kibana:
        host: "mykibanahost:5601"
        username: "my_kibana_user"  
        password: "YOUR_PASSWORD"

      This examples shows a hard-coded password, but you should store sensitive values in the secrets keystore.

      The username and password settings for Kibana are optional. If you don’t specify credentials for Kibana, Filebeat uses the username and password specified for the Elasticsearch output.

      To use the pre-built Kibana dashboards, this user must have the kibana_user built-in role or equivalent privileges.

      For more information, see Securing Filebeat.

  3. Enable the modules you want to run. For example, the following command enables the system, nginx, and mysql modules:

    deb and rpm:

    filebeat modules enable system nginx mysql

    mac:

    ./filebeat modules enable system nginx mysql

    brew:

    filebeat modules enable system nginx mysql

    linux:

    ./filebeat modules enable system nginx mysql

    win:

    PS > .\filebeat.exe modules enable system nginx mysql

    This command enables the module configs defined in the modules.d directory. See Specify which modules to run for other ways to enable modules.

    To see a list of enabled and disabled modules, run:

    deb and rpm:

    filebeat modules list

    mac:

    ./filebeat modules list

    brew:

    filebeat modules list

    linux:

    ./filebeat modules list

    win:

    PS > .\filebeat.exe modules list
  4. Set up the initial environment:

    deb and rpm:

    filebeat setup -e

    mac:

    ./filebeat setup -e

    linux:

    ./filebeat setup -e

    brew:

    filebeat setup -e

    win:

    PS > .\filebeat.exe setup -e

    The setup command loads the recommended index template for writing to Elasticsearch and deploys the sample dashboards (if available) for visualizing the data in Kibana. This is a one-time setup step.

    The -e flag is optional and sends output to standard error instead of syslog.

    The ingest pipelines used to parse log lines are set up automatically the first time you run the module, assuming the Elasticsearch output is enabled. If you’re sending events to Logstash, or plan to use Beats central management, also see Load ingest pipelines manually.

  5. Run Filebeat.

    If your logs aren’t in the default location, set the paths variable before running Filebeat.

    deb and rpm:

    service filebeat start

    mac:

    ./filebeat -e

    brew:

    filebeat -e

    linux:

    ./filebeat -e

    win:

    PS > Start-Service filebeat

    If the module is configured correctly, you’ll see INFO Harvester started messages for each file specified in the config.

    Depending on how you’ve installed Filebeat, you might see errors related to file ownership or permissions when you try to run Filebeat modules. See Config File Ownership and Permissions in the Beats Platform Reference for more information.

Example dashboard

edit

Here’s an example of the syslog dashboard:

Syslog dashboard

Set the paths variable

edit

The examples here assume that the logs you’re harvesting are in the location expected for your OS and that the default behavior of Filebeat is appropriate for your environment.

Each module and fileset has variables that you can set to change the default behavior of the module, including the paths where the module looks for log files. You can set the path in configuration or from the command line. For example:

- module: nginx
  access:
    var.paths: ["/var/log/nginx/access.log*"] 

Sets the path for nginx access log files.

To set the path at the command line, use the -M flag. The variable name must include the module and fileset name. For example:

deb and rpm:

filebeat -e -M "nginx.access.var.paths=[/var/log/nginx/access.log*]"

mac:

./filebeat -e -M "nginx.access.var.paths=[/usr/local/var/log/nginx/access.log*]"

brew:

filebeat -e -M "nginx.access.var.paths=[/usr/local/var/log/nginx/access.log*]"

linux:

./filebeat -e -M "nginx.access.var.paths=[/usr/local/var/log/nginx/access.log*]"

win:

PS > .\filebeat.exe -e -M "nginx.access.var.paths=[c:/programdata/nginx/logs/*access.log*]"

You can specify multiple overrides. Each override must start with -M.

If you are running Filebeat as a service, you cannot set paths from the command line. You must set the var.paths option in the module configuration file.

For information about specific variables that you can set for each fileset, see the documentation for the modules.

See Specify which modules to run for more information about setting variables and advanced options.

Load ingest pipelines manually

edit

The ingest pipelines used to parse log lines are set up automatically the first time you run the module, assuming the Elasticsearch output is enabled. If you’re sending events to Logstash, or plan to use Beats central management, you need to load the ingest pipelines manually. To do this, run the setup command with the --pipelines option specified. If you used the modules command to enable modules in the modules.d directory, also specify the --modules flag. For example, the following command loads the ingest pipelines used by all filesets enabled in the system, nginx, and mysql modules:

deb and rpm:

filebeat setup --pipelines --modules system,nginx,mysql

mac:

./filebeat setup --pipelines --modules system,nginx,mysql

linux:

./filebeat setup --pipelines --modules system,nginx,mysql

win:

PS > .\filebeat.exe setup --pipelines --modules system,nginx,mysql

If you’re loading ingest pipelines manually because you want to send events to Logstash, also see Working with Filebeat modules.