centos7 yum安装ELK8.X+filebeat

centos7 yum安装ELK8.X+filebeat


环境版本

系统:centos7.9

elasticsearch-8.5.3
kibana-8.5.3
logstash-8.5.3
filebeat-8.5.3

一、ELK下载路径

  1. 下载地址:Elastic官网
    官网下载

  2. 我下载的是rpm格式
    在这里插入图片描述

  3. 在ssh工具上,可以通过wget 命令将4个rpm包进行下载,

wget https://artifacts.elastic.co/downloads/elasticsearch/elasticsearch-8.5.3-x86_64.rpm
wget https://artifacts.elastic.co/downloads/kibana/kibana-8.5.3-x86_64.rpm
wget https://artifacts.elastic.co/downloads/logstash/logstash-8.5.3-x86_64.rpm
wget https://artifacts.elastic.co/downloads/beats/filebeat/filebeat-8.5.3-x86_64.rpm
  1. 下载的包就放在root下,如
    在这里插入图片描述

二、elasticsearch 安装配置

1.elasticsearch 安装

使用yum localinstall 安装本地rpm包,执行代码如下(示例):

yum localinstall elasticsearch-8.5.3-x86_64.rpm

安装完成会生成默认密码,可以记录下,登录kibana时需要
在这里插入图片描述

2.配置elasticsearch.yml文件

配置elasticsearch.yml文件:

vi /etc/elasticsearch/elasticsearch.yml

把配置文件中如下几个更改或添加,并打开注释,如下

cluster.name: wxxya-es
http.port: 9200
network.host: 0.0.0.0
http.host: 0.0.0.0

elasticsearch8中多了sslt和安全中心,xpack版块是安装自动生成的不用管

在这里插入图片描述

3.配置jvm.options文件,设置es占用系统内存大小

vi /etc/elasticsearch/jvm.options

我这当前设置最大小为3g

-Xms3g
-Xmx3g

4.elasticsearch运行

启动停止

sudo systemctl start elasticsearch.service
sudo systemctl stop elasticsearch.service

开机自启

sudo /bin/systemctl daemon-reload
sudo /bin/systemctl enable elasticsearch.service

查看elasticsearch启动日志,选其一就行

journalctl --unit elasticsearch
systemctl status elasticsearch.service

例:
在这里插入图片描述
查看端口是否启动,8.x使用curl请求9200,返回:curl: (52) Empty reply from server,必须是https请求。

netstat -ntlp
curl http://localhost:9200

在这里插入图片描述


三、Kibana 安装配置

1.Kibana 安装

使用yum localinstall 安装本地rpm包,执行代码如下(示例):

yum localinstall kibana-8.5.3-x86_64.rpm

2.配置kibana.yml文件

vi /etc/kibana/kibana.yml

把配置文件中kibana改成中文,设置自己本机IP,其它不变。如下

i18n.locale: "zh-CN"
elasticsearch.hosts: ['https://172.24.67.40:9200']

在这里插入图片描述

3.kibana运行,与es一致名称改一下就行

启动停止

sudo systemctl start kibana.service
sudo systemctl stop kibana.service

查看5601商品是否启动

在这里插入图片描述

4.打开浏览器http://ip:5601,使用elastic用户的密码进行认证

第一次访问改链接的时候需要填入令牌,令牌就是第一次启动elasticsearch时保存的信息中的token,注意这个token只有30分钟的有效期,如果过期了只能进入容器重置token
在这里插入图片描述

重置token:进入容器执行

 /usr/share/elasticsearch/bin/elasticsearch-create-enrollment-token -s kibana --url "https://127.0.0.1:9200"

输入token以后会看到一个验证码框,验证码从kibana的日志中获取
在这里插入图片描述

生成验证码

/usr/share/kibana/bin/kibana-verification-code

输入用户名:elastic,密码:es安装时记录的密码
在这里插入图片描述

四、logstash 安装配置

1.logstash 安装

使用yum localinstall 安装本地rpm包,执行代码如下(示例):

yum localinstall logstash-8.5.3-x86_64.rpm

2.配置logstash.yml文件

vi /etc/logstash/logstash.yml

在kibana.yml文件中增加了

http.host: "0.0.0.0"
http.port: 9600-9700

更改startup.options文件,设置为root用户权限

LS_USER=root
LS_GROUP=root

创建一个logstash.conf配置文件,我这是配置nginx日志和pm2日志,nginx日志格式也得配置一下

vi /etc/logstash/conf.d/logstash.conf

logstash.conf内容如下

# Sample Logstash configuration for creating a simple
# Beats -> Logstash -> Elasticsearch pipeline.

input {
  
   beats {
    port => 4567
  }
   file {
    path => "/var/log/nginx/access.log"
    type => "nginx-accesslog"
    stat_interval => "1"
    start_position => "beginning"
  }

  file {
    path => "/var/log/nginx/error.log"
    type => "nginx-errorlog"
    stat_interval => "1"
    start_position => "beginning"
  }

}

filter {
  if [type] == "nginx-accesslog" {
  grok {
    match => { "message" => ["%{IPORHOST:clientip} - %{DATA:username} \[%{HTTPDATE:request-time}\] \"%{WORD:request-method} %{DATA:request-uri} HTTP/%{NUMBER:http_version}\" %{NUMBER:response_code} %{NUMBER:body_sent_bytes} \"%{DATA:referrer}\" \"%{DATA:useragent}\""] }
    remove_field => "message"
    add_field => { "project" => "magedu"}
  }
  mutate {
    convert => [ "[response_code]", "integer"]
    }
  }
  if [type] == "nginx-errorlog" {
    grok {
      match => { "message" => ["(?<timestamp>%{YEAR}[./]%{MONTHNUM}[./]%{MONTHDAY} %{TIME}) \[%{LOGLEVEL:loglevel}\] %{POSINT:pid}#%{NUMBER:threadid}\: \*%{NUMBER:connectionid} %{GREEDYDATA:message}, client: %{IPV4:clientip}, server: %{GREEDYDATA:server}, request: \"(?:%{WORD:request-method} %{NOTSPACE:request-uri}(?: HTTP/%{NUMBER:httpversion}))\", host: %{GREEDYDATA:domainname}"]}
      remove_field => "message"
    }
  }
}

output {
	if [type] == "nginx-accesslog" {
    elasticsearch {
		hosts => ["https://172.24.67.40:9200"]
		index => "nginx-accesslog-%{+yyyy.MM.dd}"
		ssl => true
		ssl_certificate_verification => false
		user => "elastic"
		password => "tsy123456"
	}}
	if [type] == "nginx-errorlog" {
    elasticsearch {
		hosts => ["https://172.24.67.40:9200"]
		index => "nginx-errorlog-%{+yyyy.MM.dd}"
		ssl => true
		ssl_certificate_verification => false
		user => "elastic"
		password => "tsy123456"
	}}
	if [fields][service] == "pm2" {
    elasticsearch {
		hosts => ["https://172.24.67.40:9200"]
		index => "pm2-log-%{+yyyy.MM.dd}"
		ssl => true
		ssl_certificate_verification => false
		user => "elastic"
		password => "tsy123456"
	}}
}

我之前出过nginx配置和安装:nginx安装教程

vi /etc/nginx/nginx.conf

nginx文件更改内容,设置json日志格式

user root;
worker_processes auto;

pid /run/nginx.pid;


events {
    worker_connections 1024;
}

http {
    log_format  main  '$remote_addr - $remote_user [$time_local] "$request" '
                      '$status $body_bytes_sent "$http_referer" '
                      '"$http_user_agent" "$http_x_forwarded_for"';
					  
	log_format json '{ "@timestamp": "$time_iso8601", '
                     '"remote_addr": "$remote_addr", ' # 客户端的ip地址
                     '"remote_user": "$remote_user", '  # 客户端用户名称
                     '"body_bytes_sent": "$body_bytes_sent", ' # 发送给客户端文件主体内容大小
                     '"request_time": "$request_time", '
                     '"status": "$status", '
                     '"host": "$host", '
                     '"request": "$request", ' # 请求的url与http协议
                     '"request_method": "$request_method", '
                     '"uri": "$uri", '
                     '"http_referrer": "$http_referer", ' # 从那个页面链接访问过来的
                     '"http_x_forwarded_for": "$http_x_forwarded_for", ' # 客户端真实ip地址

                     '"http_user_agent": "$http_user_agent" '  # 客户端浏览器的相关信息

                '}';

    access_log  /var/log/nginx/access.log  json ;
	error_log /var/log/nginx/error.log;


	#autoindex on;   #开启nginx目录浏览功能
    #autoindex_exact_size off;   #文件大小从KB开始显示
    #autoindex_localtime on;   #显示文件修改时间为服务器本地时间
	#charset utf-8,gbk;	#中文目录的话会出现乱码问题,加上

    sendfile            on;
    tcp_nopush          on;
    tcp_nodelay         on;
    keepalive_timeout   65;
    types_hash_max_size 2048;
    server_names_hash_bucket_size 128;
    client_header_buffer_size 32k;
    large_client_header_buffers 4 32k;
    client_max_body_size 500m;
    client_body_buffer_size 512k;

   # 代理的相关参数设置
    proxy_connect_timeout 5;
    proxy_read_timeout 60;
    proxy_send_timeout 5;
    proxy_buffer_size 16k;
    proxy_buffers 4 64k;
    proxy_busy_buffers_size 128k;
    proxy_temp_file_write_size 128k;
	 

   # 启用gzip压缩,提高用户访问速度
    gzip on;
    gzip_min_length 1k;
    gzip_buffers 4 16k;
    gzip_http_version 1.1;
    gzip_comp_level 2;
    gzip_types text/plain application/css application/javascript application/x-javascript text/css application/xml text/javascript application/x-httpd-php image/jpeg image/gif image/png;
    gzip_vary on;

    include             /etc/nginx/mime.types;
    default_type        application/octet-stream;

	
	
	
    include /etc/nginx/conf.d/*.conf;
}

3.logstash运行,与es一致名称改一下就行

启动停止

sudo systemctl start logstash.service
sudo systemctl stop logstash.service

4.在kibana中可看,创建成功的索引

在这里插入图片描述
创建数据视图与索引关联
在这里插入图片描述
数据展示
在这里插入图片描述

五、filebeat 安装配置

1.filebeat 安装

使用yum localinstall 安装本地rpm包,执行代码如下(示例):

yum localinstall filebeat-8.5.3-x86_64.rpm

2.配置filebeat.yml文件

vi /etc/filebeat/filebeat.yml

在filebeat.yml文件中主要配置读取paths设置,和对接到logstash中,filebeat我配置的是读pm2日志,整体配置文件内容如下

###################### Filebeat Configuration Example #########################

# This file is an example configuration file highlighting only the most common
# options. The filebeat.reference.yml file from the same directory contains all the
# supported options with more comments. You can use it as a reference.
#
# You can find the full configuration reference here:
# https://www.elastic.co/guide/en/beats/filebeat/index.html

# For more available modules and options, please see the filebeat.reference.yml sample
# configuration file.

# ============================== Filebeat inputs ===============================

filebeat.inputs:

# Each - is an input. Most options can be set at the input level, so
# you can use different inputs for various configurations.
# Below are the input specific configurations.

# filestream is an input for collecting log messages from files.
- type: filestream

  # Unique ID among all inputs, an ID is required.
  id: my-filestream-id

  # Change to true to enable this input configuration.
  enabled: true

  # Paths that should be crawled and fetched. Glob based paths.
  paths:
    - /root/.pm2/logs/*.log
  input_type: log
  fields.document_type: pm2
  fields.service: pm2
  tags: ["pm2"]

    #- c:\programdata\elasticsearch\logs\*

  # Exclude lines. A list of regular expressions to match. It drops the lines that are
  # matching any regular expression from the list.
  # Line filtering happens after the parsers pipeline. If you would like to filter lines
  # before parsers, use include_message parser.
  #exclude_lines: ['^DBG']

  # Include lines. A list of regular expressions to match. It exports the lines that are
  # matching any regular expression from the list.
  # Line filtering happens after the parsers pipeline. If you would like to filter lines
  # before parsers, use include_message parser.
  #include_lines: ['^ERR', '^WARN']

  # Exclude files. A list of regular expressions to match. Filebeat drops the files that
  # are matching any regular expression from the list. By default, no files are dropped.
  #prospector.scanner.exclude_files: ['.gz$']

  # Optional additional fields. These fields can be freely picked
  # to add additional information to the crawled log files for filtering
  #fields:
  #  level: debug
  #  review: 1

# ============================== Filebeat modules ==============================

filebeat.config.modules:
  # Glob pattern for configuration loading
  path: ${path.config}/modules.d/*.yml

  # Set to true to enable config reloading
  reload.enabled: false

  # Period on which files under path should be checked for changes
  #reload.period: 10s

# ======================= Elasticsearch template setting =======================

setup.template.settings:
  index.number_of_shards: 1
  #index.codec: best_compression
  #_source.enabled: false


# ================================== General ===================================

# The name of the shipper that publishes the network data. It can be used to group
# all the transactions sent by a single shipper in the web interface.
#name:

# The tags of the shipper are included in their own field with each
# transaction published.
#tags: ["service-X", "web-tier"]

# Optional fields that you can specify to add additional information to the
# output.
#fields:
#  env: staging

# ================================= Dashboards =================================
# These settings control loading the sample dashboards to the Kibana index. Loading
# the dashboards is disabled by default and can be enabled either by setting the
# options here or by using the `setup` command.
#setup.dashboards.enabled: false

# The URL from where to download the dashboards archive. By default this URL
# has a value which is computed based on the Beat name and version. For released
# versions, this URL points to the dashboard archive on the artifacts.elastic.co
# website.
#setup.dashboards.url:

# =================================== Kibana ===================================

# Starting with Beats version 6.0.0, the dashboards are loaded via the Kibana API.
# This requires a Kibana endpoint configuration.
setup.kibana:
  host: "localhost:5601"
  # Kibana Host
  # Scheme and port can be left out and will be set to the default (http and 5601)
  # In case you specify and additional path, the scheme is required: http://localhost:5601/path
  # IPv6 addresses should always be defined as: https://[2001:db8::1]:5601
  #host: "localhost:5601"

  # Kibana Space ID
  # ID of the Kibana Space into which the dashboards should be loaded. By default,
  # the Default Space will be used.
  #space.id:

# =============================== Elastic Cloud ================================

# These settings simplify using Filebeat with the Elastic Cloud (https://cloud.elastic.co/).

# The cloud.id setting overwrites the `output.elasticsearch.hosts` and
# `setup.kibana.host` options.
# You can find the `cloud.id` in the Elastic Cloud web UI.
#cloud.id:

# The cloud.auth setting overwrites the `output.elasticsearch.username` and
# `output.elasticsearch.password` settings. The format is `<user>:<pass>`.
#cloud.auth:

# ================================== Outputs ===================================

# Configure what output to use when sending the data collected by the beat.

# ---------------------------- Elasticsearch Output ----------------------------
#output.elasticsearch:
  # Array of hosts to connect to.
#  hosts: ["localhost:9200"]

  # Protocol - either `http` (default) or `https`.
  #protocol: "https"

  # Authentication credentials - either API key or username/password.
  #api_key: "id:api_key"
  #username: "elastic"
  #password: "changeme"

# ------------------------------ Logstash Output -------------------------------
output.logstash:
  # The Logstash hosts
  hosts: ["localhost:4567"]

  # Optional SSL. By default is off.
  # List of root certificates for HTTPS server verifications
  #ssl.certificate_authorities: ["/etc/pki/root/ca.pem"]

  # Certificate for SSL client authentication
  #ssl.certificate: "/etc/pki/client/cert.pem"

  # Client Certificate Key
  #ssl.key: "/etc/pki/client/cert.key"

# ================================= Processors =================================
processors:
  - add_host_metadata:
      when.not.contains.tags: forwarded
  - add_cloud_metadata: ~
  - add_docker_metadata: ~
  - add_kubernetes_metadata: ~


3.filebeat运行,与es一致名称改一下就行

启动停止

sudo systemctl start filebeat.service
sudo systemctl stop filebeat.service

4.查看filebeat加载的日志文件,通过logstash写入的索引

在这里插入图片描述
在这里插入图片描述

总结

以上就是今天要讲的内容,详细的介绍了ELK8全家桶安装细节,有问题的请评论区留言。

评论 1
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值