Logstash ArcSight Module

edit

Deprecated in 8.16.0.

Replace by Common Event Format (CEF) from Elastic Integrations

The Logstash ArcSight module is an X-Pack feature under the Basic License and is therefore free to use. Please contact [email protected] for questions or more information.

The Logstash ArcSight module enables you to easily integrate your ArcSight data with the Elastic Stack. With a single command, the module taps directly into the ArcSight Smart Connector or the Event Broker, parses and indexes the security events into Elasticsearch, and installs a suite of Kibana dashboards to get you exploring your data immediately.

The Logstash ArsSight Module has been deprecated and replaced by Common Event Format (CEF) from Elastic Integrations.

Prerequisites

edit

These instructions assume that Logstash, Elasticsearch, and Kibana are already installed. The products you need are available to download and easy to install. The Elastic Stack 5.6 (or later) and X-Pack are required for this module. If you are using the Elastic Stack 6.2 and earlier, please see the instructions for those versions.

Deployment Architecture

edit

The Logstash ArcSight module understands CEF (Common Event Format), and can accept, enrich, and index these events for analysis on the Elastic Stack. ADP contains two core data collection components for data streaming:

  • The Smart Connectors (SC) are edge log collectors that parse and normalize data to CEF prior to publishing to the Logstash receiver.
  • The Event Broker is the central hub for incoming data and is based on open source Apache Kafka. The Logstash ArcSight module can consume directly from Event Broker topics.

Getting Started with the Smart Connector

edit

To get started, you can use a basic Elastic Stack setup that reads events from the Smart Connector directly.

ArcSight Smart Connector architecture

Smart Connector has been configured to publish ArcSight data (to TCP port 5000) using the CEF syslog destination.

Logstash, Elasticsearch, and Kibana must run locally. Note that you can also run Elasticsearch, Kibana and Logstash on separate hosts to consume data from ArcSight.

Instructions for Smart Connector

edit
  1. Start the Logstash ArcSight module by running the following command in the Logstash install directory with your respective Smart Connector host and port:

    bin/logstash --modules arcsight --setup \
      -M "arcsight.var.input.smartconnector.port=smart_connect_port"  \
      -M "arcsight.var.elasticsearch.hosts=localhost:9200"  \
      -M "arcsight.var.kibana.host=localhost:5601"

    The --modules arcsight option spins up an ArcSight CEF-aware Logstash pipeline for ingestion. The --setup option creates an arcsight-* index pattern in Elasticsearch and imports Kibana dashboards and visualizations. On subsequent module runs or when scaling out the Logstash deployment, the --setup option should be omitted to avoid overwriting the existing Kibana dashboards.

    See Logstash ArcSight Module Configuration Options for more info.

  2. Explore your data with Kibana:

    1. Open browser @ http://localhost:5601 (username: "elastic"; password: "YOUR_PASSWORD")
    2. Open the [ArcSight] Network Overview Dashboard
    3. See Exploring Your Security Data for additional details on data exploration.

See Configuring the Module if you want to specify additional options that control the behavior of the ArcSight module.

Getting Started with the Event Broker

edit

To get started, you can use a basic Elastic Stack setup that reads events from the Event Broker event stream.

ArcSight Event Broker architecture

By default, the Logstash ArcSight module consumes from the Event Broker "eb-cef" topic. For additional settings, see Logstash ArcSight Module Configuration Options. Consuming from a secured Event Broker port is possible, see Logstash ArcSight Module Configuration Options.

Logstash, Elasticsearch, and Kibana must run locally. Note that you can also run Elasticsearch, Kibana and Logstash on separate hosts to consume data from ArcSight.

Instructions for Event Broker

edit
  1. Start the Logstash ArcSight module by running the following command in the Logstash install directory with your respective Event Broker host and port:

    bin/logstash --modules arcsight --setup \
     -M "arcsight.var.input.eventbroker.bootstrap_servers=event_broker_host:event_broker_port"  \
     -M "arcsight.var.elasticsearch.hosts=localhost:9200"  \
     -M "arcsight.var.kibana.host=localhost:5601"

    The --modules arcsight option spins up an ArcSight CEF-aware Logstash pipeline for ingestion. The --setup option creates an arcsight-* index pattern in Elasticsearch and imports Kibana dashboards and visualizations. On subsequent module runs or when scaling out the Logstash deployment, the --setup option should be omitted to avoid overwriting the existing Kibana dashboards.

    See Logstash ArcSight Module Configuration Options for more info.

  2. Explore your data with Kibana:

    1. Open browser @ http://localhost:5601 (username: "elastic"; password: "YOUR_PASSWORD")
    2. Open the [ArcSight] Network Overview Dashboard
    3. See Exploring Your Security Data for additional details on data exploration.

See Configuring the Module if you want to specify additional options that control the behavior of the ArcSight module.

Exploring Your Security Data

edit

Once the Logstash ArcSight module starts receiving events, you can immediately begin using the packaged Kibana dashboards to explore and visualize your security data. The dashboards rapidly accelerate the time and effort required for security analysts and operators to gain situational and behavioral insights on network, endpoint, and DNS events flowing through the environment. You can use the dashboards as-is, or tailor them to work better with existing use cases and business requirements.

The dashboards have a navigation pane for context switching and drill downs across three core use cases:

  • Network Data

    • Dashboards: Network Overview, Network Suspicious Activity
    • Data Types: Network firewalls, intrusion systems, VPN devices
  • Endpoint Data

    • Dashboards: Endpoint Overview, Endpoint OS Activity
    • Data Types: Operating systems, applications, host intrusion systems
  • DNS Data

    • Dashboards: Microsoft DNS Overview
    • Data Types: Microsoft DNS devices

Example Network Dashboards

edit
Network overview dashboard
Network suspicious activity dashboard

These Kibana visualizations enable you to quickly understand the top devices, endpoints, attackers, and targets. This insight, along with the ability to instantly drill down on a particular host, port, device, or time range, offers a holistic view across the entire environment to identify specific segments that may require immediate attention or action. You can easily discover answers to questions like:

  • Who are my attackers and what are they targeting?
  • Which of my devices or endpoints are the busiest and what services were rendered?
  • How many unique attackers, techniques, signatures, or targets were triggered at any given point in time?
  • What are the top sources, destinations, protocols, and behaviors that are causing the elevated count of failures?

Configuring the Module

edit

You can specify additional options for the Logstash ArcSight module in the logstash.yml configuration file or with overrides through the command line like in the getting started. For more information about configuring modules, see Working with Logstash Modules.

As an example, the following settings can be appended to logstash.yml to configure your module:

modules:
  - name: arcsight
    var.input.eventbroker.bootstrap_servers: "eb_host:39092"
    var.input.eventbroker.topics: "eb_topic"
    var.elasticsearch.hosts: "localhost:9200"
    var.elasticsearch.username: "elastic"
    var.elasticsearch.password: "YOUR_PASSWORD"
    var.kibana.host: “localhost:5601”
    var.kibana.username: "elastic"
    var.kibana.password: "YOUR_PASSWORD"

Logstash ArcSight Module Configuration Options

edit

The ArcSight module provides the following settings for configuring the behavior of the module. These settings include ArcSight-specific options plus common options that are supported by all Logstash modules.

When you override a setting at the command line, remember to prefix the setting with the module name, for example, arcsight.var.inputs instead of var.inputs.

If you don’t specify configuration settings, Logstash uses the defaults.

ArcSight Module Options

var.inputs
  • Value type is string
  • Default value is "eventbroker"

Set the input(s) to expose for the Logstash ArcSight module. Valid settings are "eventbroker", "smartconnector", or "eventbroker,smartconnector" (exposes both inputs concurrently).

ArcSight Module Event Broker specific Options

var.input.eventbroker.bootstrap_servers
  • Value type is string
  • Default value is "localhost:39092"

A list of Event Broker URLs to use for establishing the initial connection to the cluster. This list should be in the form of host1:port1,host2:port2. These URLs are just used for the initial connection to discover the full cluster membership (which may change dynamically). This list need not contain the full set of servers. (You may want more than one in case a server is down.)

var.input.eventbroker.topics
  • Value type is array
  • Default value is ["eb-cef"]

A list of Event Broker topics to subscribe to.

var.input.eventbroker.security_protocol
  • Value can be any of: PLAINTEXT, SSL, SASL_PLAINTEXT, SASL_SSL
  • Default value is "PLAINTEXT"

Security protocol to use, which can be either of PLAINTEXT, SSL, SASL_PLAINTEXT, SASL_SSL. If you specify anything other than PLAINTEXT then you need to also specify some of the options listed below. When specifying SSL or SASL_SSL you should supply values for the options prefixed with ssl_, when specifying SASL_PLAINTEXT or SASL_SSL you should supply values for jaas_path, kerberos_config, sasl_mechanism and sasl_kerberos_service_name.

var.input.eventbroker.ssl_key_password
  • Value type is password
  • There is no default value for this setting.

The password of the private key in the key store file.

var.input.eventbroker.ssl_keystore_location
  • Value type is path
  • There is no default value for this setting.

If client authentication is required, this setting stores the keystore path.

var.input.eventbroker.ssl_keystore_password
  • Value type is password
  • There is no default value for this setting.

If client authentication is required, this setting stores the keystore password.

var.input.eventbroker.ssl_keystore_type
  • Value type is string
  • There is no default value for this setting.

The keystore type.

var.input.eventbroker.ssl_truststore_location
  • Value type is path
  • There is no default value for this setting.

The JKS truststore path to validate the Kafka broker’s certificate.

var.input.eventbroker.ssl_truststore_password
  • Value type is password
  • There is no default value for this setting.

The truststore password.

var.input.eventbroker.ssl_truststore_type
  • Value type is string
  • There is no default value for this setting.

The truststore type.

var.input.eventbroker.sasl_kerberos_service_name
  • Value type is string
  • There is no default value for this setting.

The Kerberos principal name that Kafka broker runs as. This can be defined either in Kafka’s JAAS config or in Kafka’s config.

var.input.eventbroker.sasl_mechanism
  • Value type is string
  • Default value is "GSSAPI"

SASL mechanism used for client connections. This may be any mechanism for which a security provider is available. GSSAPI is the default mechanism.

var.input.eventbroker.jaas_path
  • Value type is path
  • There is no default value for this setting.

The Java Authentication and Authorization Service (JAAS) API supplies user authentication and authorization services for Kafka. This setting provides the path to the JAAS file. Sample JAAS file for Kafka client:

KafkaClient {
  com.sun.security.auth.module.Krb5LoginModule required
  useTicketCache=true
  renewTicket=true
  serviceName="kafka";
  };

Please note that specifying jaas_path and kerberos_config here will add these to the global JVM system properties. This means if you have multiple Kafka inputs, all of them would be sharing the same jaas_path and kerberos_config. If this is not desirable, you would have to run separate instances of Logstash on different JVM instances.

var.input.eventbroker.kerberos_config
  • Value type is path
  • There is no default value for this setting.

Optional path to kerberos config file. This is krb5.conf style as detailed in https://web.mit.edu/kerberos/krb5-1.12/doc/admin/conf_files/krb5_conf.html

ArcSight Module Smart Connector specific Options

var.input.smartconnector.port
  • Value type is number
  • Default value is 5000

The TCP port to listen on when receiving data from SCs.

var.input.smartconnector.ssl_enable
  • Value type is boolean
  • Default value is false

Enable SSL (must be set for other ssl_ options to take effect).

var.input.smartconnector.ssl_cert
  • Value type is path
  • There is no default value for this setting.

SSL certificate path.

var.input.smartconnector.ssl_extra_chain_certs
  • Value type is array
  • Default value is []

An Array of paths to extra X509 certificates to be added to the certificate chain. Useful when the CA chain is not necessary in the system store.

var.input.smartconnector.ssl_key
  • Value type is path
  • There is no default value for this setting.

SSL key path

var.input.smartconnector.ssl_key_passphrase
  • Value type is password
  • Default value is nil

SSL key passphrase

var.input.smartconnector.ssl_verify
  • Value type is boolean
  • Default value is true

Verify the identity of the other end of the SSL connection against the CA. For input, sets the field sslsubject to that of the client certificate.

Common options

The following configuration options are supported by all modules:

var.elasticsearch.hosts
  • Value type is uri
  • Default value is "localhost:9200"

Sets the host(s) of the Elasticsearch cluster. For each host, you must specify the hostname and port. For example, "myhost:9200". If given an array, Logstash will load balance requests across the hosts specified in the hosts parameter. It is important to exclude dedicated master nodes from the hosts list to prevent Logstash from sending bulk requests to the master nodes. So this parameter should only reference either data or client nodes in Elasticsearch.

Any special characters present in the URLs here MUST be URL escaped! This means # should be put in as %23 for instance.

var.elasticsearch.username
  • Value type is string
  • Default value is "elastic"

The username to authenticate to a secure Elasticsearch cluster.

var.elasticsearch.password
  • Value type is string
  • Default value is "changeme"

The password to authenticate to a secure Elasticsearch cluster.

var.elasticsearch.ssl.enabled
  • Value type is boolean
  • There is no default value for this setting.

Enable SSL/TLS secured communication to the Elasticsearch cluster. Leaving this unspecified will use whatever scheme is specified in the URLs listed in hosts. If no explicit protocol is specified, plain HTTP will be used. If SSL is explicitly disabled here, the plugin will refuse to start if an HTTPS URL is given in hosts.

var.elasticsearch.ssl.verification_mode
  • Value type is string
  • Default value is "strict"

The hostname verification setting when communicating with Elasticsearch. Set to disable to turn off hostname verification. Disabling this has serious security concerns.

var.elasticsearch.ssl.certificate_authority
  • Value type is string
  • There is no default value for this setting

The path to an X.509 certificate to use to validate SSL certificates when communicating with Elasticsearch.

var.elasticsearch.ssl.certificate
  • Value type is string
  • There is no default value for this setting

The path to an X.509 certificate to use for client authentication when communicating with Elasticsearch.

var.elasticsearch.ssl.key
  • Value type is string
  • There is no default value for this setting

The path to the certificate key for client authentication when communicating with Elasticsearch.

var.kibana.host
  • Value type is string
  • Default value is "localhost:5601"

Sets the hostname and port of the Kibana instance to use for importing dashboards and visualizations. For example: "myhost:5601".

var.kibana.scheme
  • Value type is string
  • Default value is "http"

Sets the protocol to use for reaching the Kibana instance. The options are: "http" or "https". The default is "http".

var.kibana.username
  • Value type is string
  • Default value is "elastic"

The username to authenticate to a secured Kibana instance.

var.kibana.password
  • Value type is string
  • Default value is "changeme"

The password to authenticate to a secure Kibana instance.

var.kibana.ssl.enabled
  • Value type is boolean
  • Default value is false

Enable SSL/TLS secured communication to the Kibana instance.

var.kibana.ssl.verification_mode
  • Value type is string
  • Default value is "strict"

The hostname verification setting when communicating with Kibana. Set to disable to turn off hostname verification. Disabling this has serious security concerns.

var.kibana.ssl.certificate_authority
  • Value type is string
  • There is no default value for this setting

The path to an X.509 certificate to use to validate SSL certificates when communicating with Kibana.

var.kibana.ssl.certificate
  • Value type is string
  • There is no default value for this setting

The path to an X.509 certificate to use for client authentication when communicating with Kibana.

var.kibana.ssl.key
  • Value type is string
  • There is no default value for this setting

The path to the certificate key for client authentication when communicating with Kibana.