- 示例
filebeat.yml
配置,收集Tomcat的catalina.out日志(注释内容留下以便学习)
[vagrant@localhost filebeat-7.7.1]$ vi filebeat.yml
###################### Filebeat Configuration Example #########################
#=========================== Filebeat inputs =============================
filebeat.inputs:
# Each - is an input. Most options can be set at the input level, so
#每个-是一个输入。大多数选项可以在输入级别设置,因此
# you can use different inputs for various configurations.
#您可以为各种配置使用不同的输入。
# Below are the input specific configurations.
#下面是特定于输入的配置。
- type: log
# Change to true to enable this input configuration.
#更改为true以启用此输入配置。
enabled: true
# Paths that should be crawled and fetched. Glob based paths.
#应该被爬取的路径。基础路径。
paths:
#可配置多个路径
- /home/vagrant/apache-tomcat-9.0.20/logs/catalina.*.out
#- c:\programdata\elasticsearch\logs\*
# Exclude lines. A list of regular expressions to match. It drops the lines that are
#排除线路。要匹配的正则表达式列表。它去掉了
# matching any regular expression from the list.
#匹配列表中的任何正则表达式。
#exclude_lines: ['^DBG']
# Include lines. A list of regular expressions to match. It exports the lines that are
#要匹配的正则表达式列表。它导出
# matching any regular expression from the list.
#匹配列表中的任何正则表达式。
#include_lines: ['^INFO','^ERR', '^WARN']
# Exclude files. A list of regular expressions to match. Filebeat drops the files that
#排除的文件。要匹配的正则表达式列表。Filebeat删除的文件
# are matching any regular expression from the list. By default, no files are dropped.
#匹配列表中的任何正则表达式。默认情况下,没有文件被删除。
#exclude_files: ['.gz$']
# Optional additional fields. These fields can be freely picked
#可选的附加字段。这些字段可以自由选择
# to add additional information to the crawled log files for filtering
#添加附加信息到抓取的日志文件进行过滤
#fields:
# level: debug
# review: 1
### Multiline options
# Multiline can be used for log messages spanning multiple lines. This is common
# Multiline可用于记录跨多行的消息。这是常见的
# for Java Stack Traces or C-Line Continuation
#用于Java堆栈跟踪或c行延续
# The regexp Pattern that has to be matched. The example pattern matches all lines starting with [
#必须匹配的regexp模式。示例模式匹配以[开头的所有行
multiline.pattern: ^\[
# Defines if the pattern set under pattern should be negated or not. Default is false.
#定义模式下的模式集是否应该被否定。默认是false
multiline.negate: true
# Match can be set to "after" or "before". It is used to define if lines should be append to a pattern
#Match可以设置为“after”或“before”。它用于定义是否应该将行追加到模式中
# that was (not) matched before or after or as long as a pattern is not matched based on negate.
#在之前或之后匹配的,或者只要模式没有基于negate匹配。
# Note: After is the equivalent to previous and before is the equivalent to to next in Logstash
#注意:在Logstash中,After等同于previous, before等同于next
multiline.match: after
#============================= Filebeat modules ===============================
filebeat.config.modules:
# Glob pattern for configuration loading
#配置加载的Glob模式
path: ${path.config}/modules.d/*.yml
# Set to true to enable config reloading
#设置为true可重新加载配置
reload.enabled: false
# Period on which files under path should be checked for changes
#应该检查path下的文件是否有更改的时间段
#reload.period: 10s
#==================== Elasticsearch template setting ==========================
setup.template.settings:
index.number_of_shards: 1
#index.codec: best_compression
#_source.enabled: false
#================================ General =====================================
# The name of the shipper that publishes the network data. It can be used to group
#应该检查path下文件更改的时间段#发布网络数据的托运人的名称。它可以用来分组
# all the transactions sent by a single shipper in the web interface.
#由一个托运人在web interfac中发送的所有事务
#name:
# The tags of the shipper are included in their own field with each
#每个托运人的标签都包含在它们自己的字段中
# transaction published.
#事务发表。
#tags: ["service-X", "web-tier"]
# Optional fields that you can specify to add additional information to the
#属性中添加附加信息的可选字段
# output.
#fields:
# env: staging
#============================== Dashboards =====================================
# These settings control loading the sample dashboards to the Kibana index. Loading
#这些设置控制将样例指示板加载到Kibana索引。加载
# the dashboards is disabled by default and can be enabled either by setting the
#仪表板在默认情况下是禁用的,可以通过设置
# options here or by using the `setup` command.
#选项或使用' setup '命令。
#setup.dashboards.enabled: false
# The URL from where to download the dashboards archive. By default this URL
#下载仪表板归档文件的URL。默认情况下,这个URL
# has a value which is computed based on the Beat name and version. For released
#有一个基于节拍名称和版本计算的值。对发布的
# versions, this URL points to the dashboard archive on the artifacts.elastic.co
#版本号,此URL指向工件.elastic.co上的仪表板存档
# website.
#setup.dashboards.url:
#============================== Kibana =====================================
# Starting with Beats version 6.0.0, the dashboards are loaded via the Kibana API.
#从Beats 6.0.0版本开始,仪表板是通过Kibana API加载的。
# This requires a Kibana endpoint configuration.
#这需要Kibana端点配置。
setup.kibana:
# Kibana Host
# Scheme and port can be left out and will be set to the default (http and 5601)
# In case you specify and additional path, the scheme is required: http://localhost:5601/path
# IPv6 addresses should always be defined as: https://[2001:db8::1]:5601
host: "192.168.0.140:5601"
# Kibana Space ID
# ID of the Kibana Space into which the dashboards should be loaded. By default,
# the Default Space will be used.
#space.id:
#============================= Elastic Cloud ==================================
# These settings simplify using Filebeat with the Elastic Cloud (https://cloud.elastic.co/).
# The cloud.id setting overwrites the `output.elasticsearch.hosts` and
# `setup.kibana.host` options.
# You can find the `cloud.id` in the Elastic Cloud web UI.
#cloud.id:
# The cloud.auth setting overwrites the `output.elasticsearch.username` and
# `output.elasticsearch.password` settings. The format is `<user>:<pass>`.
#cloud.auth:
#================================ Outputs =====================================
# Configure what output to use when sending the data collected by the beat.
#配置在发送由节拍收集的数据时使用的输出。
#-------------------------- Elasticsearch output ------------------------------
#output.elasticsearch:
# Array of hosts to connect to.
#hosts: ["192.168.0.140:9200"]
# Protocol - either `http` (default) or `https`.
#protocol: "https"
# Authentication credentials - either API key or username/password.
#api_key: "id:api_key"
#username: "elastic"
#password: "changeme"
#----------------------------- Logstash output --------------------------------
output.logstash:
# The Logstash hosts
hosts: ["192.168.0.140:5044"]
# Optional SSL. By default is off.
# List of root certificates for HTTPS server verifications
#ssl.certificate_authorities: ["/etc/pki/root/ca.pem"]
# Certificate for SSL client authentication
#ssl.certificate: "/etc/pki/client/cert.pem"
# Client Certificate Key
#ssl.key: "/etc/pki/client/cert.key"
#================================ Processors =====================================
# Configure processors to enhance or manipulate events generated by the beat.
#配置处理器以增强或操纵节拍生成的事件。
processors:
- add_host_metadata: ~
- add_cloud_metadata: ~
- add_docker_metadata: ~
- add_kubernetes_metadata: ~
#================================ Logging =====================================
# Sets log level. The default log level is info.
# Available log levels are: error, warning, info, debug
#logging.level: debug
# At debug level, you can selectively enable logging only for some components.
# To enable all selectors use ["*"]. Examples of other selectors are "beat",
# "publish", "service".
#logging.selectors: ["*"]
#============================== X-Pack Monitoring ===============================
# filebeat can export internal metrics to a central Elasticsearch monitoring
# cluster. This requires xpack monitoring to be enabled in Elasticsearch. The
# reporting is disabled by default.
# Set to true to enable the monitoring reporter.
#monitoring.enabled: false
# Sets the UUID of the Elasticsearch cluster under which monitoring data for this
# Filebeat instance will appear in the Stack Monitoring UI. If output.elasticsearch
# is enabled, the UUID is derived from the Elasticsearch cluster referenced by output.elasticsearch.
#monitoring.cluster_uuid:
# Uncomment to send the metrics to Elasticsearch. Most settings from the
# Elasticsearch output are accepted here as well.
# Note that the settings should point to your Elasticsearch *monitoring* cluster.
# Any setting that is not set is automatically inherited from the Elasticsearch
# output configuration, so if you have the Elasticsearch output configured such
# that it is pointing to your Elasticsearch monitoring cluster, you can simply
# uncomment the following line.
#monitoring.elasticsearch:
#================================= Migration ==================================
# This allows to enable 6.7 migration aliases
#migration.6_to_7.enabled: true