1、下载zookeeper
docker pull wurstmeister/zookeeper
2、下载kafka
docker pull wurstmeister/kafka:2.11-0.11.0.3
3、启动zookeeper
docker run -d --name zookeeper --publish 2181:2181 --volume /etc/localtime:/etc/localtime wurstmeister/zookeeper
报错:
docker: Error response from daemon: OCI runtime create failed: container_linux.go:349: starting container process caused "process_linux.go:449: container init caused \"rootfs_linux.go:58: mounting \\\"/etc/localtime\\\" to rootfs \\\"/var/lib/docker/overlay2/7b63cb086d7b04c3b5a979624d6ea3c39672b268e634d0d80224eefde47e51e2/merged\\\" at \\\"/var/lib/docker/overlay2/7b63cb086d7b04c3b5a979624d6ea3c39672b268e634d0d80224eefde47e51e2/merged/etc/localtime\\\" caused \\\"not a directory\\\"\"": unknown: Are you trying to mount a directory onto a file (or vice-versa)? Check if the specified host path exists and is the expected type.
解决:创建一个自定义容器卷 edc-zookeeper-vol 查看所有容器:docker volume ls
docker volume edc-zookeeper-vol
将挂载的目录也修改为 /var/lib/docker/volumes/etc/localtime 再次执行:
docker run -d --name zookeeper --publish 2181:2181 --volume edc-zookeeper-vol:/var/lib/docker/volumes/etc/localtime wurstmeister/zookeeper
4、启动kafa
docker run -d --name kafka --publish 9092:9092 --link zookeeper --env KAFKA_ZOOKEEPER_CONNECT=192.168.0.104:2181 --env KAFKA_ADVERTISED_HOST_NAME=192.168.0.104 --env KAFKA_ADVERTISED_PORT=9092 --volume edc-zookeeper-vol:/var/lib/docker/volumes/etc/localtime wurstmeister/kafka:2.11-0.11.0.3
5、创建topic
以下是因为kafka-topics.sh并不在bin目录下所走的弯路:
查看kafka容器ID:docker ps
进入容器:docker exec -it [容器ID] bin/bash
创建topic
bin/kafka-topics.sh --create --zookeeper 192.168.0.104:2181 --replication-factor 1 --partitions 1 --topic mykafka
问题:创建topics时执行命令报错:bash: bin/kafka-topics.sh: No such file or directory
进入bin文件:cd /bin
创建文件kafka-topic.sh文件:touch kafka-topics.sh
问题:再次执行创建topic时报错:bash: bin/kafka-topics.sh: Permission denied 因为没有权限
解决:进入文件所在的目录执行:chmod a+x *.sh 然后再次执行创建topics的命令
其实kafka-topics.sh文件在 /opt/kafka/bin 目录下
再次创建:
opt/kafka/bin/kafka-topics.sh --create --zookeeper 192.168.0.104:2181 --replication-factor 1 --partitions 1 --topic mykafka
返回:Created topic "mykafka". 说明已创建成功
6、查看topic
opt/kafka/bin/kafka-topics.sh --list --zookeeper 192.168.0.104:2181
返回:mykafka
7、创建生产者
opt/kafka/bin/kafka-console-producer.sh --broker-list 192.168.0.104:9092 --topic mykafka
返回:> 表示进入输入状态
8、重新开一个命令行窗口,并再次进入kafka容器中,创建消费者
opt/kafka/bin/kafka-console-consumer.sh --zookeeper 192.168.0.104:2181 --topic mykafka --from-beginning
9、在生产者的命令行中输入信息,在消费者的命令行窗口就可以实时的接收到信息。
10、安装ELK
docker pull sebp/elk
11、启动ELK
docker run -p 5601:5601 -p 9200:9200 -p 9300:9300 -p 5044:5044 -e ES_MIN_MEM=128m -e ES_MAX_MEM=2048m -d --name elk sebp/elk
过10秒就可以看到ELK相关的面板。:http://localhost:5601/
12、进入ELK容器: docker exec -it 容器ID bin/bash
13、测试 Logstash和ES之间是可以正常联通
/opt/logstash/bin/logstash -e 'input { stdin { } } output { elasticsearch { hosts => ["localhost"] } }'
报错: Logstash could not be started because there is already another instance using the configured data directory. If you wish to run multiple instances, you must change the "path.data" setting.
进入指定的文件夹root@c953e1538736:/opt/logstash/data# 中将.lock文件删除
再次执行命令 /opt/logstash/bin/logstash -e 'input { stdin { } } output { elasticsearch { hosts => ["localhost"] } }'
14、这句话:Successfully started Logstash API endpoint {:port=>9600}表示成功
然后输入一句话:olidigital.com hello world 回车
打开浏览器,输入:http://192.168.0.104:9200/_search?pretty 如图,就会看到我们刚刚输入的日志内容。(并没有找到输入的日志,待解。。。)
15、配置Logstash从Kafka消费消息
1) 找到config文件
cd /opt/logstash/config
2) 编辑配置文件(没有该文件会新建一个)
vi logstash.config
输入内容:
input {
kafka{
bootstrap_servers =>["192.168.0.104:9092"]
client_id => "test" group_id => "test"
consumer_threads => 5
decorate_events => true
topics => "mi"
}
}
filter{
json{
source => "message"
}
}
output {
elasticsearch {
hosts => ["localhost"]
index => "mi-%{app_id}"
codec => "json"
}
}
16、进入elk加载配置文件从kafka上消费
/opt/logstash/bin/logstash -f /opt/logstash/config/logstash.config