Logstash way to explore https:https://cdn.v2ex.com/navatar/b8c3/7e33/1001_normal.png?m=1502861312 https:https://cdn.v2ex.com/navatar/b8c3/7e33/1001_large.png?m=1502861312 2024-05-06T09:40:00Z Copyright © 2010-2018, V2EX 使用 logstash 消费 avro 数据出现部分字节被替换为 ef bf bd 导致解析出错。 tag:www.v2ex.com,2024-05-06:/t/1038102 2024-05-06T07:41:00Z 2024-05-06T09:40:00Z InDom member/InDom 消费渠道 kafka 与 redis 均出现相同问题,直接以 plain => { charset => "BINARY" } 依然出现此问题。

avro 数据原文( 16 进制)

00000000: 00 80 40 00 cc e1 85 98 0e 00 3a 32 30 32 34 2d ..@.......:2024- 00000010: 30 34 2d 31 38 54 30 36 3a 33 32 3a 30 34 2e 30 04-18T06:32:04.0 00000020: 30 30 2b 30 38 3a 30 30 00+08:00 

logstash 获取到的内容

00000000: 00 ef bf bd 40 00 ef bf bd ef bf bd ef bf bd ef ....@........... 00000010: bf bd 0e 00 3a 32 30 32 34 2d 30 34 2d 31 38 54 ....:2024-04-18T 00000020: 30 36 3a 33 32 3a 30 34 2e 30 30 30 2b 30 38 3a 06:32:04.000+08: 00000030: 30 30 0a 00. 

logstash 是通过 docker 起的 8.13.0 ,以下是容器相关配置

docker-compose.yml

 logstash: image: logstash:8.13.0 volumes: - /root/kafka-2-es/logstash.conf:/usr/share/logstash/pipeline/logstash.conf environment: - "XPACK_MONITORING_ENABLED=false" - "KAFKA_GROUP_ID" 

logstash.conf

input { kafka { bootstrap_servers => "kafka:9092" topics => ["urllog-test-01"] codec => plain { charset => "BINARY" } group_id => "${KAFKA_GROUP_ID}" auto_offset_reset => "earliest" } redis { host => "192.168.4.101" port => 6379 password => "Fle7YuG22qIh7ZNPkceopo3oZb1UFZrX" data_type => "list" key => "urllog-test-01" codec => plain { charset => "BINARY" } } } output { stdout { codec => rubydebug } elasticsearch { hosts => ["https://192.168.1.4:9200"] index => "urllog-test-01" user => "elastic" password => "123456" ssl => true ssl_certificate_verification => false manage_template => true } } 

已经确认的是:

  1. 使用 go 从 kafka 中消费获取到的二进制内容是正确的。
  2. 使用 logstash 从 kafka 与 redis 消费相同的内容,均出现特殊字符被替换为 ef bf bd 的情况
  3. 使用 kafka-console-consumer.sh --bootstrap-server localhost:9092 --topic urllog-test-01 --from-beginning >> /tmp/data-1.txt 直接从 kafka 中得到的内容是正确的

有没有大佬知道是这么回事,给指条路呗?万分感谢。

]]> 请教下这种情况下如何 logstash 处理时间问题? tag:www.v2ex.com,2021-08-09:/t/794702 2021-08-09T10:35:59Z 2021-08-10T15:50:36Z watara member/watara 背景:

因为项目更新,打算弃用原来的日志流架构,换用 logstash 处理日志。为了保持后续流程不动,所以需要 logstash 的输出要保持于之前架构的格式一致。

问题:

日志流需要处理的是 nginx 的日志,日志体中已经有一个时间字段(在这里记作 ST1 吧,原始格式是"23/Jul/2021:14:25:29 +0800"),然后输出的内容中会涉及到时间的字段,格式以及来源分别如下:

字段 1:"2021/7/23 14:25",记录 logstash 处理的时间

字段 2:"20210723",根据 nginx 原始日志时间 ST1 转换出的日期

字段 3:"2021/7/23 14:25",根据 nginx 原始日志时间 ST1 转换出的时间

字段 4:"23/Jul/2021:14:25:29 +0800",原始的 nginx 原始日志时间 ST1

我现在能正确填充字段 4,其他 3 个字段(特别是字段 2 )没有思路,请教有大佬能指点一二么?

]]>
各位大大,日志监控告警系统,大家有什么好的方案推荐么? tag:www.v2ex.com,2020-10-26:/t/718792 2020-10-26T10:52:56Z 2020-10-26T10:53:37Z duanlian member/duanlian 1 、大家有用到哪些方案呢? 2 、服务器分布跨区域,国内服务器比较集中,海外是多国分布,需要采集日志,这种情况,有什么好的办法么? 3 、海外和国内各一套系统?

]]>
logstash 如何将时间戳 UNIX_MS 转化为北京时间? tag:www.v2ex.com,2016-12-23:/t/329635 2016-12-23T03:55:00Z 2017-01-04T12:07:47Z wildcat007 member/wildcat007 {"id":"549","msg-type":"activity","date":"1482372059180","thread-id":"628913","query-id":"16363060","user":"uuuser","priv_user":"uuuser","ip":"10.7.32.64","cmd":"show_create_table","query":"SHOW CREATE TABLE `uc_logindata`"}

搭建 elk ,但是想把 date 改成北京时间,不知道 filter 如何去写?

目前的 filter 格式如下
input {
file {
type => "mysql-audit"
path => ["/opt/audit_logs/mysql-audit/*"]
}
}

filter{
json {
source => "message"
target => "json"
}
}


output {
stdout {
codec => rubydebug
}
elasticsearch {
hosts => ["172.19.11.32:9200"]
}
}


输出后的结果如下
{
"path" => "/opt/audit_logs/mysql-audit/test.json",
"@timestamp" => 2016-12-23T03:05:49.752Z,
"@version" => "1",
"host" => "0.0.0.0",
"json" => {
"date" => "1482372059180",
"query-id" => "16363060",
"msg-type" => "activity",
"ip" => "10.7.32.64",
"query" => "SHOW CREATE TABLE `uc_logindata`",
"priv_user" => "uuuser",
"id" => "7",
"cmd" => "show_create_table",
"user" => "uuuser",
"thread-id" => "628913"
},
"message" => "{\"id\":\"7\",\"msg-type\":\"activity\",\"date\":\"1482372059180\",\"thread-id\":\"628913\",\"query-id\":\"16363060\",\"user\":\"uuuser\",\"priv_user\":\"uuuser\",\"ip\":\"10.7.32.64\",\"cmd\":\"show_create_table\",\"query\":\"SHOW CREATE TABLE `uc_logindata`\"}",
"type" => "mysql-audit",
"tags" => []
} ]]>
使用 ELK(Elasticsearch + Logstash + Kibana) 搭建日志集中分析平台实践 tag:www.v2ex.com,2015-08-31:/t/217318 2015-08-31T06:44:21Z 2016-09-02T04:47:22Z wsgzao member/wsgzao

前言

Elasticsearch + Logstash + Kibana ( ELK )是一套开源的日志管理方案,分析网站的访问情况时我们一般会借助 Google/百度 /CNZZ 等方式嵌入 JS 做数据统计,但是当网站访问异常或者被攻击时我们需要在后台分析如 Nginx 的具体日志,而 Nginx 日志分割 /GoAccess/Awstats 都是相对简单的单节点解决方案,针对分布式集群或者数据量级较大时会显得心有余而力不足,而 ELK 的出现可以使我们从容面对新的挑战。

ELK (Elasticsearch + Logstash + Kibana )


更新记录

2015 年 08 月 31 日 - 初稿

阅读原文 - http://wsgzao.github.io/post/elk/

扩展阅读

CentOS 7.x 安装 ELK (Elasticsearch+Logstash+Kibana ) - http://www.chenshake.com/centos-install-7-x-elk-elasticsearchlogstashkibana/
Centos 6.5 安装 nginx 日志分析系统 elasticsearch + logstash + redis + kibana - http://blog.chinaunix.net/xmlrpc.php?r=blog/article&uid=17291169&id=4898582
logstash-forwarder and grok examples - https://www.ulyaoth.net/threads/logstash-forwarder-and-grok-examples.32413/
三斗室 - http://chenlinux.com/
elastic - https://www.elastic.co/guide
LTMP 索引 - http://wsgzao.github.io/index/#LTMP

]]>
logstash+elasticsearch+kibana 问题 tag:www.v2ex.com,2015-08-21:/t/214926 2015-08-21T09:21:00Z 2015-08-21T09:18:00Z zjf061315 member/zjf061315 在用着三个工具搭一个日志处理的平台,目前有两个不同的系统日志需要处理。
开始考虑在一台机器上启动两个 logstash 实例,收集日志后发到 elasticsearch 上,不知道为什么第二个总是连接不上,传不出数据,进而考虑用一个 logstash 实例,然后在配置文件里给文件加 type ,在 output 中根据 type 判断处理,但是还是有一个条件里的不成功啊。。。求助大家,目前输出的配置文件如下
output {

if [type] == "targeted_coupon" { elasticsearch { host => ["127.0.0.1:9300"] cluster => "cluster_spark" protocol => "node" index => "coupon_%{+YYYYMMdd}" max_retries => 10 } #stdout { codec => rubydebug } } else { elasticsearch { host => ["127.0.0.1:9300"] cluster => "cluster_spark" protocol => "node" index => "gift_%{+YYYYMMdd}" document_type => "%{type}" max_retries => 10 } #stdout { codec => rubydebug } } 
]]>
ubao snddm index pchome yahoo rakuten mypaper meadowduck bidyahoo youbao zxmzxm asda bnvcg cvbfg dfscv mmhjk xxddc yybgb zznbn ccubao uaitu acv GXCV ET GDG YH FG BCVB FJFH CBRE CBC GDG ET54 WRWR RWER WREW WRWER RWER SDG EW SF DSFSF fbbs ubao fhd dfg ewr dg df ewwr ewwr et ruyut utut dfg fgd gdfgt etg dfgt dfgd ert4 gd fgg wr 235 wer3 we vsdf sdf gdf ert xcv sdf rwer hfd dfg cvb rwf afb dfh jgh bmn lgh rty gfds cxv xcv xcs vdas fdf fgd cv sdf tert sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf shasha9178 shasha9178 shasha9178 shasha9178 shasha9178 liflif2 liflif2 liflif2 liflif2 liflif2 liblib3 liblib3 liblib3 liblib3 liblib3 zhazha444 zhazha444 zhazha444 zhazha444 zhazha444 dende5 dende denden denden2 denden21 fenfen9 fenf619 fen619 fenfe9 fe619 sdf sdf sdf sdf sdf zhazh90 zhazh0 zhaa50 zha90 zh590 zho zhoz zhozh zhozho zhozho2 lislis lls95 lili95 lils5 liss9 sdf0ty987 sdft876 sdft9876 sdf09876 sd0t9876 sdf0ty98 sdf0976 sdf0ty986 sdf0ty96 sdf0t76 sdf0876 df0ty98 sf0t876 sd0ty76 sdy76 sdf76 sdf0t76 sdf0ty9 sdf0ty98 sdf0ty987 sdf0ty98 sdf6676 sdf876 sd876 sd876 sdf6 sdf6 sdf9876 sdf0t sdf06 sdf0ty9776 sdf0ty9776 sdf0ty76 sdf8876 sdf0t sd6 sdf06 s688876 sd688 sdf86