其他分享
首页 > 其他分享> > 安装ELK-单机版

安装ELK-单机版

作者:互联网

前言:

首先要有一个全面的认识,什么是ELK?

Elastic Stack也就是ELK,ELK是三款软件的集合,分别是Elasticsearch,logstas,Kibana,在发展过程中,有了新的成员Beats加入,所以就形成了Elastic Starck.也是就是说ELK是旧的称呼,Elastic Stack是新的名字。

先通过Beats采集一切的数据如日志文件,网络流量,Win事件日志,服务指标,健康检查等,然后把数据发送给elasticsearch保存起来,也可以发送给logstas处理然后再发送给elasticsearch,最后通过kibana的组件将数据可视化的展示出来。

Elasticsearch
Elasticsearch基于java,是个开源分布式手术引擎,它的特点有:分布式,零配置,自动发现,索引自动分片,索引副本机制,restful风格接口,多数据源,自动搜索负载等。

Logstash
也是基于java,是一个开源的用于收集,分析和存储日志的工具。

Kibana
Kibana基于nodejs,也是开源和免费的工具,Kibana开源为logsash和Elasticsearch提供日志分析友好的web界面,可以汇总,分析和搜索重要的数据日志。

Beats
Bests是elastic公司开源的一款采集系统监控数据的代理agent,是在被监控服务器上以客户端形式运行的数据收集器的统称,可以直接把数据发送给Elasticsearch或者通过Logstash发送给Elasticsearch,然后进行后续的数据分析活动

ELK官网:

直接从官网下载ELK的所有安装包

Free and Open Search: The Creators of Elasticsearch, ELK & Kibana | Elastic

Note: 怎样部署ELK?

ELK需要使用单独的机器部署,为什么? 因为ELK不涉及业务,而且日志量大  会占用业务的资源,所以需要单独部署。

本文中需要安装ELK,以及kafka,zookeeper, kafka作为logstash的数据源。

1.安装单机版Elasticsearch

1、基础环境配置

添加用户

出于安全的考虑,elasticsearch不允许使用root用户来启动,所以需要创建一个新的用户,并为这个账户赋予相应的权限来启动elasticsearch集群(赋予权限的操作在后面启动ES的环节)。 

  #useradd es

修改服务打开文件数

(1).vim /etc/security/limits.conf

root soft nofile 65535  (这个不需要修改成262144,无需改动)
root hard nofile 65535  (这个不需要修改成262144,无需改动)
  *  soft nofile 262144
  *  hard nofile 262144

#下面两句没加,但是看很多文章里加了

* soft memlock unlimited  
* hard memlock unlimited

防止进程虚拟内存不足: 

(2).vim /etc/sysctl.conf

   #文件尾部设置
    vm.max_map_count=262144

使上面的设置生效:

sysctl -p

2、修改ES配置文件elasticsearch.yml

cluster.name: wit
node.name: node-es
node.master: true
node.data: true
network.host: 0.0.0.0
path.data: /usr/local/elasticsearch-7.8.1/data
path.logs: /usr/local/elasticsearch-7.8.1/logs
http.port: 9200
transport.tcp.port: 9300

#写⼊候选主节点的设备地址,在开启服务后可以被选为主节点

#因为要配置三个主节点,所以这里写了三个,如果只需要一个主节点,则这里只写一个

#这里不要写端口,只写IP,写端口的话,集群搭建不成功。

#由于不是集群,这里只写自己就好了

discovery.seed_hosts: ["10.17.64.172"]

#这里填写的是节点的名字,不是节点的IP地址,由于不是集群,这里只写自己就好了
cluster.initial_master_nodes: ["node-es"]

#下面的几个配置还没使用,看情况吧 

http.cors.enabled: true
http.cors.allow-origin: "*"
http.cors.allow-headers: Authorization
xpack.security.enabled: true
xpack.security.transport.ssl.enabled: true

3. 其它配置文件

(1).ES自带java,所以不需要在系统中安装java,但是如果系统中已经有了java,那么ES会用哪一个?

(2).由于Elasticsearch是Java开发的,所以可以通过../elasticsearch/config/jvm.options配置文件来设定JVM的相关设定。如果没有特殊需求按默认即可。 不过其中还是有两项最重要的-Xmx1g 与

-Xms1gJVM的最大最小Heap。如果太小会导致Elasticsearch刚刚启动就立刻停止。太大会拖慢

系统本身。

4.启动es

1、给elasticsearch文件夹分配es权限

chown -R es:es elasticsearch文件路径

2、切换至es用户su - es

3、启动ES

bin/elasticsearch --前台启动 
bin/elasticsearch -d --后台启动 
 

5.验证ES集群:

shell命令行访问 :

虽然不是集群,但是也可以用这个API进行查看。

 curl localhost:9200/_cluster/health?pretty

命令输出信息:

[root@ecs-zhihuiyingxiao-dev config]# curl localhost:9200/_cluster/health?pretty
{
  "cluster_name" : "wit",
  "status" : "yellow",
  "timed_out" : false,
  "number_of_nodes" : 1,
  "number_of_data_nodes" : 1,
  "active_primary_shards" : 7,
  "active_shards" : 7,
  "relocating_shards" : 0,
  "initializing_shards" : 0,
  "unassigned_shards" : 1,
  "delayed_unassigned_shards" : 0,
  "number_of_pending_tasks" : 0,
  "number_of_in_flight_fetch" : 0,
  "task_max_waiting_in_queue_millis" : 0,
  "active_shards_percent_as_number" : 87.5
}

本地http访问:

curl  http://localhost:9200

外部http访问:

开放防火墙端口:

firewall-cmd --zone=public --add-port=9200/tcp --permanent

firewall-cmd --reload

systemctl daemon-reload

导入数据到ES:

json数据文件data.json如下:

它的第一行定义了_index,_type,_id等信息;第二行定义了字段的信息(关于ES数据格式问题,比如要有哪些字段,请自行研究)。 

{ "index" : { "_index" : "test", "_type" : "type1", "_id" : "1" } }
{ "field1" : "value1" }

curl -H "Content-Type: application/json" -X POST localhost:9200/_bulk --data-binary @data.json

执行结果如下:

二、安装Kibana

Kibana是一个针对Elasticsearch的开源分析及可视化平台,使用Kibana可以查询、查看并与存储在ES索引的数据进行交互操作,使用Kibana能执行高级的数据分析,并能以图表、表格和地图的形式查看数据。

1.获取到安装包

获取kibana-7.8.1-linux-x86_64.tar(自己去官网下载,这里不提供),然后解压

tar xvf kibana-7.8.1-linux-x86_64.tar

2.配置yum源:

在/etc/yum.repos.d/ 下新建 kibana.repo 配置 yum 源地址:
新建kibana.repo:

cat>>kibana.repo

新建完成之后,在里面加上如下内容:

[kibana-7.x]
name=Kibana repository for 7.x packages
baseurl=https://artifacts.elastic.co/packages/7.x/yum
gpgcheck=1
gpgkey=https://artifacts.elastic.co/GPG-KEY-elasticsearch
enabled=1
autorefresh=1
type=rpm-md

3.修改配置文件kibana.yml:

修改config文件夹下的kibana.yml,添加信息:

server.port: 5601
server.host: “0.0.0.0”
server.name: tanklog
elasticsearch.hosts: [“http://127.0.0.0:9200/”]

注意,里面的地址是你自己服务器的IP地址。

4.开启端口5601:

如果防火墙是关闭的,首先打开防火墙,否则没办法开启5601端口:

1.输入命令、开启防火墙:

systemctl start firewalld.service

2.开启防火墙后,添加5601端口:

firewall-cmd --permanent --zone=public --add-port=5601/tcp

出现success,说明端口开启成功。

3.重启防火墙:

firewall-cmd --reload

5.启动kibana:

1.启动kibana之前,先启动ElasticSearch:
进入ES安装包下的bin目录,输入命令:

./elasticsearch -d(后台运行)

2.启动Kibana:
进入kibana安装目录下的bin目录,输入命令:

(1).新建kibana账户(推荐这种方式):

  #useradd kibana

授权:

 chown -R kibana:kibana kibana目录

切换账户:

su - kibana 

启动kibana:

nohup ./kibana & 

(2).如果在root账户下启动kibana

需要加--allow-root参数。 

nohup ./kibana --allow-root &  (--allow-root表示允许root用户去启动)

3.关闭防火墙:

kibana成功启动后,关闭防火墙:

systemctl stop firewalld.service

4.访问kibana:

浏览器中输入: http://IP:5601/app/kibana 
 

三、安装Zookeeper

参考:centos7上安装zookeeper - web_bird - 博客园 (cnblogs.com) 

1.需要先安装JDK:

jdk压缩包解压(自己去下载):

tar zxvf  jdk1.8.0_181.tar

配置环境变量:

修改/etc/profile文件,在文件最后加上:

export JAVA_HOME=/usr/local/jdk1.8.0_181
export PATH=$JAVA_HOME/bin:$PATH
export CLASSPATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar

使配置生效:

source /etc/profile

查看java版本:

[root@ecs-dev-0003 ~]# java -version
java version "1.8.0_181"
Java(TM) SE Runtime Environment (build 1.8.0_181-b13)
Java HotSpot(TM) 64-Bit Server VM (build 25.181-b13, mixed mode)
[root@ecs-dev-0003 ~]# javac -version
javac 1.8.0_181
[root@ecs-dev-0003 ~]# 

2.安装zookeeper:

zookeeper下载地址:https://downloads.apache.org/zookeeper/

解压安装包:

tar zxvf  zookeeper-3.4.10.tar.gz

进入到conf文件夹:

cp  zoo_sample.cfg zoo.cfg

修改配置文件zoo.cfg:

tickTime=2000
initLimit=10
syncLimit=5
dataDir=/usr/local/zookeeper-3.4.10/data
datalogDir=/usr/local/zookeeper-3.4.10/logs
clientPort=2181
autopurge.purgeInterval=1
server.1=10.17.****.****:2888:3888

我们需要注意dataDir和datalogDir对应的目录,如果对应的目录不存在,我们需要去创建。

启动zookeeper:

  ./bin/zkServer.sh start    启动zookeeper
  ./bin/zkServer.sh stop   关闭zookeeper
  ./bin/zkServer.sh status   查看zookeeper状态

也可以通过查看zookeeper的进程来判断其是否启动成功

ps -ef | grep zookeeper

3.设置zookeeper开机启动(尚未验证):

我们可以将zookeeper作为一个服务,设置其开机自启,这样每次我们打开虚拟机就可以开启zookeeper,彻底解放双手!设置zookeeper开机自启需要以下几个步骤:

1、进入 /etc/init.d 目录:

[root@ecs-dev conf]# cd /etc/init.d/
[root@ecs-dev init.d]# ls
functions  hostguard  multi-queue-hw  netconsole  network  README

2.创建文件zookeeper,并添加脚本:

vi zookeeper

脚本内容为:

#!/bin/bash
#chkconfig:2345 20 90
#description:zookeeper
#processname:zookeeper
ZK_PATH= 你的zk路径(到zk根目录就行了)
export JAVA_HOME= jdk目录(也是到根目录就好了)
case $1 in
         start) sh  $ZK_PATH/bin/zkServer.sh start;;
         stop)  sh  $ZK_PATH/bin/zkServer.sh stop;;
         status) sh  $ZK_PATH/bin/zkServer.sh status;;
         restart) sh $ZK_PATH/bin/zkServer.sh restart;;
         *)  echo "require start|stop|status|restart"  ;;
esac

3. 将zookeeper注册为服务:

chkconfig --add zookeeper

4.测试其是否生效

这里采用先停服务,再使用命令启动,注意需要修改创建的zookeeper服务权限:

[root@zhiyou init.d]# service zookeeper start
env: /etc/init.d/zookeeper: 权限不够
[root@zhiyou init.d]# 
[root@zhiyou init.d]# chmod +x zookeeper 
[root@zhiyou init.d]# 
[root@zhiyou init.d]# service zookeeper start
ZooKeeper JMX enabled by default
Using config: /opt/soft/zookeeper-3.4.11/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED
[root@zhiyou init.d]# 
[root@zhiyou init.d]# service zookeeper status
ZooKeeper JMX enabled by default
Using config: /opt/soft/zookeeper-3.4.11/bin/../conf/zoo.cfg
Mode: standalone
[root@zhiyou init.d]# 

至此,我们已经完成了在centos7下安装zookeeper的全部步骤,以及设置了zookeeper开机自启动!

四、安装Kafka

解压kafka压缩包:

tar zxvf kafka_2.13-2.6.0.tgz

修改文件夹config中的配置文件server.properties.

配置log.dirs参数的时候,需要自己创建文件夹 /usr/local/kafka_2.13-2.6.0/logs

[root@ecs-zhihuiyingxiao-dev config]# cat server.properties 

# The id of the broker. This must be set to a unique integer for each broker.
broker.id=0

listeners=PLAINTEXT://10.17.64.172:9092

advertised.listeners=PLAINTEXT://10.17.64.172:9092

auto.create.topics.enable=true
num.network.threads=3

# The number of threads that the server uses for processing requests, which may include disk I/O
num.io.threads=8

# The send buffer (SO_SNDBUF) used by the socket server
socket.send.buffer.bytes=102400

# The receive buffer (SO_RCVBUF) used by the socket server
socket.receive.buffer.bytes=102400

# The maximum size of a request that the socket server will accept (protection against OOM)
socket.request.max.bytes=104857600


############################# Log Basics #############################

# A comma separated list of directories under which to store log files
log.dirs=/usr/local/kafka_2.13-2.6.0/logs

# The default number of log partitions per topic. More partitions allow greater
# parallelism for consumption, but this will also result in more files across
# the brokers.
num.partitions=1

# The number of threads per data directory to be used for log recovery at startup and flushing at shutdown.
# This value is recommended to be increased for installations with data dirs located in RAID array.
num.recovery.threads.per.data.dir=1

############################# Internal Topic Settings  #############################
# The replication factor for the group metadata internal topics "__consumer_offsets" and "__transaction_state"
# For anything other than development testing, a value greater than 1 is recommended to ensure availability such as 3.
offsets.topic.replication.factor=1
transaction.state.log.replication.factor=1
transaction.state.log.min.isr=1

############################# Log Flush Policy #############################

# Messages are immediately written to the filesystem but by default we only fsync() to sync
# the OS cache lazily. The following configurations control the flush of data to disk.
# There are a few important trade-offs here:
#    1. Durability: Unflushed data may be lost if you are not using replication.
#    2. Latency: Very large flush intervals may lead to latency spikes when the flush does occur as there will be a lot of data to flush.
#    3. Throughput: The flush is generally the most expensive operation, and a small flush interval may lead to excessive seeks.
# The settings below allow one to configure the flush policy to flush data after a period of time or
# every N messages (or both). This can be done globally and overridden on a per-topic basis.

# The number of messages to accept before forcing a flush of data to disk
log.flush.interval.messages=10000

# The maximum amount of time a message can sit in a log before we force a flush
log.flush.interval.ms=1000

############################# Log Retention Policy #############################

# The following configurations control the disposal of log segments. The policy can
# be set to delete segments after a period of time, or after a given size has accumulated.
# A segment will be deleted whenever *either* of these criteria are met. Deletion always happens
# from the end of the log.

# The minimum age of a log file to be eligible for deletion due to age
log.retention.hours=24

# A size-based retention policy for logs. Segments are pruned from the log unless the remaining
# segments drop below log.retention.bytes. Functions independently of log.retention.hours.
#log.retention.bytes=1073741824

# The maximum size of a log segment file. When this size is reached a new log segment will be created.
log.segment.bytes=1073741824

# The interval at which log segments are checked to see if they can be deleted according
# to the retention policies
log.retention.check.interval.ms=300000

############################# Zookeeper #############################

# Zookeeper connection string (see zookeeper docs for details).
# This is a comma separated host:port pairs, each corresponding to a zk
# server. e.g. "127.0.0.1:3000,127.0.0.1:3001,127.0.0.1:3002".
# You can also append an optional chroot string to the urls to specify the
# root directory for all kafka znodes.
zookeeper.connect=10.17.64.172:2181

# Timeout in ms for connecting to zookeeper
zookeeper.connection.timeout.ms=18000

group.initial.rebalance.delay.ms=0

 

标签:ELK,单机版,log,--,root,zookeeper,kibana,data,安装
来源: https://blog.csdn.net/wdquan19851029/article/details/122702511