Google Cloud Pub/Sub Output Plugin
editGoogle Cloud Pub/Sub Output Plugin
edit- Plugin version: v1.0.2
- Released on: 2019-09-18
- Changelog
For other versions, see the overview list.
To learn more about Logstash, see the Logstash Reference.
Getting help
editFor questions about the plugin, open a topic in the Discuss forums. For bugs or feature requests, open an issue in Github. For the list of Elastic supported plugins, please consult the Elastic Support Matrix.
Description
editA Logstash plugin to upload log events to Google Cloud Pubsub. Events are batched and uploaded in the background for the sake of efficiency. Message payloads are serialized JSON representations of the events.
Example use-cases:
- Stream events to Dataproc via Pub/Sub for real-time analysis.
- Forward events from an on-prem datacenter to the Logstash in the cloud.
- Use Pub/Sub as an scalable buffer to even out event flow between processing steps.
Note: While this project is partially maintained by Google, this is not an official Google product.
Environment Configuration
editTo use this plugin, you must create a service account and grant it the publish permission on a topic. You MAY also use the Application Default Credentials assigned to a compute instance.
The Pub/Sub topic must exist before you run the plugin.
Example Configurations
editBasic
editA basic configuration which only includes a project, topic, and JSON key file:
output { google_pubsub { # Required attributes project_id => "my_project" topic => "my_topic" # Optional if you're using app default credentials json_key_file => "service_account_key.json" } }
High Volume
editIf you find that uploads are going too slowly, you can increase the message batching:
output { google_pubsub { project_id => "my_project" topic => "my_topic" json_key_file => "service_account_key.json" # Options for configuring the upload message_count_threshold => 1000 delay_threshold_secs => 10 request_byte_threshold => 5000000 } }
Attributes
editYou can attach additional attributes to each request. For example, you could attach a datacenter label to a log message to help with debugging:
output { google_pubsub { project_id => "my_project" topic => "my_topic" json_key_file => "service_account_key.json" attributes => {"origin" => "pacific-datacenter"} } }
Different Codecs
editYou can use codecs with this plugin to change the body of the events:
output { google_pubsub { project_id => "my_project" topic => "my_topic" json_key_file => "service_account_key.json" codec => plain {format => "%{[time]}: %{[message]}"} } }
Additional Resources
editGoogle Cloud Pub/Sub Output Configuration Options
editThis plugin supports the following configuration options plus the Common options described later.
Setting | Input type | Required |
---|---|---|
Yes |
||
Yes |
||
No |
||
No |
||
No |
||
No |
||
No |
Also see Common options for a list of options supported by all input plugins.
project_id
edit- Value type is string
- There is no default value for this setting.
Google Cloud Project ID (name, not number).
topic
edit- Value type is string
- There is no default value for this setting.
Google Cloud Pub/Sub Topic. You must create the topic manually before running this plugin.
json_key_file
edit- Value type is path
- There is no default value for this setting.
The path to the key to authenticate your user to the bucket.
This service user must have the pubsub.topics.publish
permission so it can publish to the topic.
If Logstash is running within Google Compute Engine and no json_key_file
is defined,
the plugin will use GCE’s Application Default Credentials.
Outside of GCE, you must to specify a Service Account JSON key file.
delay_threshold_secs
edit- Value type is number
-
Default is:
5
Send the batch once this delay has passed, from the time the first message is queued. Must be greater than 0.
message_count_threshold
edit- Value type is number
-
Default is:
100
Once this many messages are queued, send all the messages in a single call, even if the delay threshold hasn’t elapsed yet. Must be < 1000. A value of 0 will cause messages to instantly be sent but will reduce total throughput due to overhead.
request_byte_threshold
edit- Value type is bytes
-
Default is:
1000000
Once the number of bytes in the batched request reaches this threshold, send all of the messages in a single call, even if neither the delay or message count thresholds have been exceeded yet. This includes full message payload size, including any attributes set.
Common options
editThese configuration options are supported by all output plugins:
codec
edit- Value type is codec
-
Default value is
"json"
The codec used for output data. Output codecs are a convenient method for encoding your data before it leaves the output without needing a separate filter in your Logstash pipeline.
enable_metric
edit- Value type is boolean
-
Default value is
true
Disable or enable metric logging for this specific plugin instance. By default we record all the metrics we can, but you can disable metrics collection for a specific plugin.
id
edit- Value type is string
- There is no default value for this setting.
Add a unique ID
to the plugin configuration. If no ID is specified, Logstash will generate one.
It is strongly recommended to set this ID in your configuration. This is particularly useful
when you have two or more plugins of the same type. For example, if you have 2 google_pubsub outputs.
Adding a named ID in this case will help in monitoring Logstash when using the monitoring APIs.
output { google_pubsub { id => "my_plugin_id" } }