- Packetbeat Reference: other versions:
- Packetbeat overview
- Quick start: installation and configuration
- Set up and run
- Upgrade Packetbeat
- Configure
- Traffic sniffing
- Network flows
- Protocols
- Processes
- General settings
- Project paths
- Output
- Kerberos
- SSL
- Index lifecycle management (ILM)
- Elasticsearch index template
- Kibana endpoint
- Kibana dashboards
- Processors
- Define processors
- add_cloud_metadata
- add_cloudfoundry_metadata
- add_docker_metadata
- add_fields
- add_host_metadata
- add_id
- add_kubernetes_metadata
- add_labels
- add_locale
- add_network_direction
- add_nomad_metadata
- add_observer_metadata
- add_process_metadata
- add_tags
- community_id
- convert
- copy_fields
- decode_base64_field
- decode_json_fields
- decode_xml
- decode_xml_wineventlog
- decompress_gzip_field
- detect_mime_type
- dissect
- dns
- drop_event
- drop_fields
- extract_array
- fingerprint
- include_fields
- rate_limit
- registered_domain
- rename
- translate_sid
- truncate_fields
- urldecode
- Internal queue
- Logging
- HTTP endpoint
- Instrumentation
- packetbeat.reference.yml
- How to guides
- Exported fields
- AMQP fields
- Beat fields
- Cassandra fields
- Cloud provider metadata fields
- Common fields
- DHCPv4 fields
- DNS fields
- Docker fields
- ECS fields
- Flow Event fields
- Host fields
- HTTP fields
- ICMP fields
- Jolokia Discovery autodiscover provider fields
- Kubernetes fields
- Memcache fields
- MongoDb fields
- MySQL fields
- NFS fields
- PostgreSQL fields
- Process fields
- Raw fields
- Redis fields
- SIP fields
- Thrift-RPC fields
- Detailed TLS fields
- Transaction Event fields
- Measurements (Transactions) fields
- Monitor
- Secure
- Visualize Packetbeat data in Kibana
- Troubleshoot
- Get help
- Debug
- Record a trace
- Common problems
- Dashboard in Kibana is breaking up data fields incorrectly
- Packetbeat doesn’t see any packets when using mirror ports
- Packetbeat can’t capture traffic from Windows loopback interface
- Packetbeat is missing long running transactions
- Packetbeat isn’t capturing MySQL performance data
- Packetbeat uses too much bandwidth
- Error loading config file
- Found unexpected or unknown characters
- Logstash connection doesn’t work
- Publishing to Logstash fails with "connection reset by peer" message
- @metadata is missing in Logstash
- Not sure whether to use Logstash or Beats
- SSL client fails to connect to Logstash
- Monitoring UI shows fewer Beats than expected
- Dashboard could not locate the index-pattern
- Fields show up as nested JSON in Kibana
- Contribute to Beats
A newer version is available. For the latest information, see the
current release documentation.
Decode JSON fields
editDecode JSON fields
editThe decode_json_fields
processor decodes fields containing JSON strings and
replaces the strings with valid JSON objects.
processors: - decode_json_fields: fields: ["field1", "field2", ...] process_array: false max_depth: 1 target: "" overwrite_keys: false add_error_key: true
The decode_json_fields
processor has the following configuration settings:
-
fields
- The fields containing JSON strings to decode.
-
process_array
-
(Optional) A Boolean value that specifies whether to process
arrays. The default is
false
. -
max_depth
-
(Optional) The maximum parsing depth. A value of
1
will decode the JSON objects in fields indicated infields
, a value of2
will also decode the objects embedded in the fields of these parsed documents. The default is1
. -
target
-
(Optional) The field under which the decoded JSON will be written. By
default, the decoded JSON object replaces the string field from which it was
read. To merge the decoded JSON fields into the root of the event, specify
target
with an empty string (target: ""
). Note that thenull
value (target:
) is treated as if the field was not set. -
overwrite_keys
-
(Optional) A Boolean value that specifies whether existing keys in the event are overwritten by keys from the decoded JSON object. The
default value is
false
. -
expand_keys
-
(Optional) A Boolean value that specifies whether keys in the decoded JSON
should be recursively de-dotted and expanded into a hierarchical object structure.
For example,
{"a.b.c": 123}
would be expanded into{"a":{"b":{"c":123}}}
. -
add_error_key
-
(Optional) If set to
true
and an error occurs while decoding JSON keys, theerror
field will become a part of the event with the error message. If set tofalse
, there will not be any error in the event’s field. The default value isfalse
. -
document_id
-
(Optional) JSON key that’s used as the document ID. If configured,
the field will be removed from the original JSON document and stored in
@metadata._id
Was this helpful?
Thank you for your feedback.