orgoz
(Olivier Boudet)
May 23, 2018, 10:23am
1
Hi,
I have deployed Filebeat 6.2.3 in Kubernetes. All works great during few hours but something went wrong after a while.
Logs of my ingress-nginx-controller are not indexed anymore, filebeat reports this error :
2018-05-23T06:15:30.508Z ERROR log/harvester.go:243 Read line error: decoding docker JSON: json: cannot unmarshal number into Go value of type reader.dockerLog; File: %!(EXTRA string=/var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log)
The last indexed line was this one :
{"log":"95.213.177.125 - [95.213.177.125] - - [22/May/2018:16:10:23 +0000] \"CONNECT check.proxyradar.com:80 HTTP/1.1\" 400 174 \"-\" \"-\" 0 0.256 [] - - - -\n","stream":"stdout","time":"2018-05-22T16:10:23.210343159Z"}
The next line in the log is this one :
{"log":"00.000.000.00 - [00.000.000.00] - - [22/May/2018:16:47:08 +0000] \"POST /elasticsearch/_msearch HTTP/1.1\" 403 197 \"http://kibana.mydomain.com/app/kibana\" \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/64.0.3282.167 Chrome/64.0.3282.167 Safari/537.36\" 543 0.000 [monitoring-kibana-5601] - - - -\n","stream":"stdout","time":"2018-05-22T16:47:08.279680383Z"}
The filebeat configuration for this container is as follow :
autodiscover:
providers:
- condition:
contains:
docker.container.image: "nginx-ingress-controller"
config:
- module: nginx
access:
prospector:
type: docker
containers.stream: stdout
containers.ids:
- "${data.docker.container.id}"
processors:
- add_kubernetes_metadata:
in_cluster: true
error:
prospector:
type: docker
containers.stream: stderr
containers.ids:
- "${data.docker.container.id}"
processors:
- add_kubernetes_metadata:
in_cluster: true
Do you have any idea why my logs are not parsed correctly anymore ?
Thank you very much.
orgoz
(Olivier Boudet)
May 23, 2018, 10:32am
2
I just deleted the pods and since a new one was started, the logs are correctly ingested...
kvch
(Noémi Ványi)
May 23, 2018, 10:40am
3
Great! Let us know if the problem show up again.
orgoz
(Olivier Boudet)
May 23, 2018, 11:49am
4
The error shows up again !
2018-05-23T10:39:55.491Z INFO log/harvester.go:216 Harvester started for file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b42612
57d97f3321-json.log
2018-05-23T10:39:55.491Z ERROR log/harvester.go:243 Read line error: decoding docker JSON: json: cannot unmarshal number into Go value of type reader.dockerLog; File: %!(EXTRA string=/var/lib/docker/containers/f6fde94852bb9b1d
cfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log)
2018-05-23T10:39:55.491Z ERROR log/harvester.go:243 Read line error: decoding docker JSON: invalid character 'y' looking for beginning of value; File: %!(EXTRA string=/var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813
c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log)
Last indexed line is :
{"log":"5.8.18.92 - [5.8.18.92] - - [23/May/2018:10:34:44 +0000] \"\\x03\\x00\\x00+\u0026\\xE0\\x00\\x00\\x00\\x00\\x00Cookie: mstshash=hello\" 400 174 \"-\" \"-\" 0 0.094 [] - - - -\n","stream":"stdout","time":"2018-05-23T10:34:44.014458055Z"}
Next line is :
{"log":"00.000.00.000 - [00.000.00.000] - - [23/May/2018:10:46:01 +0000] \"GET /subpath/version-mobile HTTP/2.0\" 401 51 \"-\" \"Mozilla/5.0 (Linux; Android 6.0; MYA-L11 Build/HUAWEIMYA-L11; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/66.0.3359.158 Mobile Safari/537.36\" 1385 0.006 [default-myservice-8080] 10.32.0.31:8080 51 0.008 401\n","stream":"stdout","time":"2018-05-23T10:46:01.463045569Z"}
exekias
(Carlos Pérez Aradros)
May 23, 2018, 12:25pm
5
Could you please share the contents of this file: /var/lib/docker/containers/f6fde94852bb9b1d cfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
? I'm only interested on the lines around this one: 5.8.18.92 - [5.8.18.92] - - [23/May/2018:10:34:44 +0000]...
orgoz
(Olivier Boudet)
May 23, 2018, 1:51pm
6
Here are the lines around the last indexed one :
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:10:31:25 +0000] \"POST /api/metrics/vis/data HTTP/1.1\" 200 486 \"http://kibana.mydomain.com/app/kibana\" \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/66.0.3359.139 Chrome/66.0.3359.139 Safari/537.36\" 1828 0.020 [monitoring-kibana-5601] 10.32.0.17:5601 486 0.020 200\n","stream":"stdout","time":"2018-05-23T10:31:25.790025711Z"}
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:10:31:25 +0000] \"GET /api/metrics/fields?index=filebeat-* HTTP/1.1\" 200 37 \"http://kibana.mydomain.com/app/kibana\" \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/66.0.3359.139 Chrome/66.0.3359.139 Safari/537.36\" 458 0.004 [monitoring-kibana-5601] 10.32.0.17:5601 37 0.004 200\n","stream":"stdout","time":"2018-05-23T10:31:25.981819376Z"}
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:10:31:27 +0000] \"GET /ui/favicons/favicon-32x32.png HTTP/1.1\" 304 0 \"http://kibana.mydomain.com/app/kibana\" \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/66.0.3359.139 Chrome/66.0.3359.139 Safari/537.36\" 547 0.002 [monitoring-kibana-5601] 10.32.0.17:5601 0 0.004 304\n","stream":"stdout","time":"2018-05-23T10:31:27.847171961Z"}
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:10:31:28 +0000] \"GET /ui/favicons/favicon-16x16.png HTTP/1.1\" 304 0 \"http://kibana.mydomain.com/app/kibana\" \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/66.0.3359.139 Chrome/66.0.3359.139 Safari/537.36\" 547 0.002 [monitoring-kibana-5601] 10.32.0.17:5601 0 0.004 304\n","stream":"stdout","time":"2018-05-23T10:31:28.024105236Z"}
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:10:31:28 +0000] \"POST /api/metrics/vis/data HTTP/1.1\" 200 751 \"http://kibana.mydomain.com/app/kibana\" \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/66.0.3359.139 Chrome/66.0.3359.139 Safari/537.36\" 1828 0.100 [monitoring-kibana-5601] 10.32.0.17:5601 751 0.100 200\n","stream":"stdout","time":"2018-05-23T10:31:28.037632638Z"}
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:10:31:28 +0000] \"GET /api/metrics/fields?index=filebeat-* HTTP/1.1\" 200 37 \"http://kibana.mydomain.com/app/kibana\" \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/66.0.3359.139 Chrome/66.0.3359.139 Safari/537.36\" 458 0.005 [monitoring-kibana-5601] 10.32.0.17:5601 37 0.008 200\n","stream":"stdout","time":"2018-05-23T10:31:28.218912783Z"}
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:10:32:45 +0000] \"POST /elasticsearch/_msearch HTTP/1.1\" 200 22431 \"http://kibana.mydomain.com/app/kibana\" \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/66.0.3359.139 Chrome/66.0.3359.139 Safari/537.36\" 1356 0.634 [monitoring-kibana-5601] 10.32.0.17:5601 22403 0.632 200\n","stream":"stdout","time":"2018-05-23T10:32:45.885468176Z"}
{"log":"5.8.18.92 - [5.8.18.92] - - [23/May/2018:10:34:44 +0000] \"\\x03\\x00\\x00+\u0026\\xE0\\x00\\x00\\x00\\x00\\x00Cookie: mstshash=hello\" 400 174 \"-\" \"-\" 0 0.094 [] - - - -\n","stream":"stdout","time":"2018-05-23T10:34:44.014458055Z"}
{"log":"00.000.00.000 - [00.000.00.000] - - [23/May/2018:10:46:01 +0000] \"GET /app-ws/version-mobile HTTP/2.0\" 401 51 \"-\" \"Mozilla/5.0 (Linux; Android 6.0; MYA-L11 Build/HUAWEIMYA-L11; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/66.0.3359.158 Mobile Safari/537.36\" 1385 0.006 [default-app-service-8080] 10.32.0.31:8080 51 0.008 401\n","stream":"stdout","time":"2018-05-23T10:46:01.463045569Z"}
{"log":"00.000.00.000 - [00.000.00.000] - - [23/May/2018:10:46:02 +0000] \"POST /app-ws/refresh-token HTTP/2.0\" 200 1401 \"-\" \"Mozilla/5.0 (Linux; Android 6.0; MYA-L11 Build/HUAWEIMYA-L11; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/66.0.3359.158 Mobile Safari/537.36\" 1198 0.532 [default-app-service-8080] 10.32.0.31:8080 2481 0.044 200\n","stream":"stdout","time":"2018-05-23T10:46:02.164214622Z"}
{"log":"00.000.00.000 - [00.000.00.000] - - [23/May/2018:10:46:02 +0000] \"GET /app-ws/version-mobile HTTP/2.0\" 200 5 \"-\" \"Mozilla/5.0 (Linux; Android 6.0; MYA-L11 Build/HUAWEIMYA-L11; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/66.0.3359.158 Mobile Safari/537.36\" 1116 0.005 [default-app-service-8080] 10.32.0.31:8080 5 0.004 200\n","stream":"stdout","time":"2018-05-23T10:46:02.986110301Z"}
{"log":"00.000.00.000 - [00.000.00.000] - - [23/May/2018:10:46:03 +0000] \"GET /app-ws/version-mobile HTTP/2.0\" 200 5 \"-\" \"Mozilla/5.0 (Linux; Android 6.0; MYA-L11 Build/HUAWEIMYA-L11; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/66.0.3359.158 Mobile Safari/537.36\" 39 0.004 [default-app-service-8080] 10.32.0.31:8080 5 0.004 200\n","stream":"stdout","time":"2018-05-23T10:46:03.483461454Z"}
{"log":"00.000.00.000 - [00.000.00.000] - - [23/May/2018:10:46:03 +0000] \"GET /app-sg/data/ HTTP/2.0\" 200 183 \"-\" \"Mozilla/5.0 (Linux; Android 6.0; MYA-L11 Build/HUAWEIMYA-L11; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/66.0.3359.158 Mobile Safari/537.36\" 124 0.004 [default-sync-gateway-service-4984] 10.32.1.28:4984 183 0.004 200\n","stream":"stdout","time":"2018-05-23T10:46:03.790551672Z"}
{"log":"00.000.00.000 - [00.000.00.000] - - [23/May/2018:10:46:04 +0000] \"GET /app-sg/data/ HTTP/2.0\" 200 183 \"-\" \"Mozilla/5.0 (Linux; Android 6.0; MYA-L11 Build/HUAWEIMYA-L11; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/66.0.3359.158 Mobile Safari/537.36\" 69 0.003 [default-sync-gateway-service-4984] 10.32.1.27:4984 183 0.004 200\n","stream":"stdout","time":"2018-05-23T10:46:04.092312269Z"}
orgoz
(Olivier Boudet)
May 23, 2018, 2:08pm
7
I received another error :
2018-05-23T12:51:57.594Z ERROR log/harvester.go:243 Read line error: decoding docker JSON: invalid character 'T' looking for beginning of value; File: %!(EXTRA string=/var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log)
For these lines (the last correctly indexed is the 4th) :
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:12:31:06 +0000] \"GET /ui/favicons/favicon-32x32.png HTTP/1.1\" 304 0 \"http://kibana.mydomain.com/app/kibana\" \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/66.0.3359.139 Chrome/66.0.3359.139 Safari/537.36\" 547 0.006 [monitoring-kibana-5601] 10.32.0.17:5601 0 0.008 304\n","stream":"stdout","time":"2018-05-23T12:31:06.684417017Z"}
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:12:31:07 +0000] \"GET /ui/favicons/favicon-16x16.png HTTP/1.1\" 304 0 \"http://kibana.mydomain.com/app/kibana\" \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/66.0.3359.139 Chrome/66.0.3359.139 Safari/537.36\" 547 0.003 [monitoring-kibana-5601] 10.32.0.17:5601 0 0.004 304\n","stream":"stdout","time":"2018-05-23T12:31:07.072998177Z"}
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:12:31:07 +0000] \"POST /elasticsearch/_msearch HTTP/1.1\" 200 69534 \"http://kibana.mydomain.com/app/kibana\" \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/66.0.3359.139 Chrome/66.0.3359.139 Safari/537.36\" 1255 0.778 [monitoring-kibana-5601] 10.32.0.17:5601 69443 0.780 200\n","stream":"stdout","time":"2018-05-23T12:31:07.883790922Z"}
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:12:31:15 +0000] \"POST /elasticsearch/_msearch HTTP/1.1\" 200 23860 \"http://kibana.mydomain.com/app/kibana\" \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/66.0.3359.139 Chrome/66.0.3359.139 Safari/537.36\" 1588 0.872 [monitoring-kibana-5601] 10.32.0.17:5601 23832 0.868 200\n","stream":"stdout","time":"2018-05-23T12:31:15.280776281Z"}
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:12:46:49 +0000] \"GET /grafana HTTP/1.1\" 404 3558 \"http://devtools.mydomain.com/confluence/display/PROJECT/My+Page\" \"Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/66.0.3359.181 Safari/537.36\" 650 0.017 [monitoring-grafana-3000] 10.32.2.15:3000 14545 0.016 404\n","stream":"stdout","time":"2018-05-23T12:46:49.708948048Z"}
{"log":"00.000.000.000 - [00.000.000.000] - - [23/May/2018:12:46:49 +0000] \"GET /public/build/grafana.dark.css?v5.0.4 HTTP/1.1\" 304 0 \"http://grafana.mydomain.com/grafana\" \"Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/66.0.3359.181 Safari/537.36\" 596 0.002 [monitoring-grafana-3000] 10.32.2.15:3000 0 0.000 304\n","stream":"stdout","time":"2018-05-23T12:46:49.997672894Z"}
Each time, the recreation of the pods reindex correctly the file.
The only repeating scenario I see is that each time an error occurs, the response code is not 200.
exekias
(Carlos Pérez Aradros)
May 23, 2018, 9:02pm
8
This looks pretty similar to this open issue: https://github.com/elastic/beats/issues/7045
So far I had no luck trying to reproduce it will keep trying
Best regards
orgoz
(Olivier Boudet)
May 24, 2018, 8:56am
9
If it can help, here is the harvester debug log when error occured. Is there something else which can help for investigation ?
2018-05-24T08:53:48.283Z INFO log/harvester.go:216 Harvester started for file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
2018-05-24T08:53:48.283Z ERROR log/harvester.go:243 Read line error: decoding docker JSON: invalid character 'o' looking for beginning of value; File: %!(EXTRA string=/var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log)
2018-05-24T08:53:48.283Z DEBUG [harvester] log/harvester.go:468 Stopping harvester for file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
2018-05-24T08:53:48.283Z DEBUG [harvester] log/harvester.go:478 Closing file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
2018-05-24T08:53:48.283Z DEBUG [harvester] log/harvester.go:348 Update state: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log, offset: 451348
2018-05-24T08:53:48.283Z DEBUG [harvester] log/harvester.go:489 harvester cleanup finished for file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
2018-05-24T08:53:48.283Z INFO log/harvester.go:216 Harvester started for file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
2018-05-24T08:53:48.283Z ERROR log/harvester.go:243 Read line error: decoding docker JSON: json: cannot unmarshal number into Go value of type reader.dockerLog; File: %!(EXTRA string=/var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log)
2018-05-24T08:53:48.283Z DEBUG [harvester] log/harvester.go:468 Stopping harvester for file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
2018-05-24T08:53:48.283Z DEBUG [harvester] log/harvester.go:478 Closing file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
2018-05-24T08:53:48.283Z DEBUG [harvester] log/harvester.go:348 Update state: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log, offset: 6618202
2018-05-24T08:53:48.283Z DEBUG [harvester] log/harvester.go:489 harvester cleanup finished for file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
After the creation of a new pod, it index correctly the logs from the start, here is the debug log :
2018-05-24T08:58:02.232Z DEBUG [harvester] log/harvester.go:447 Setting offset for file based on seek: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
2018-05-24T08:58:02.232Z DEBUG [harvester] log/harvester.go:433 Setting offset for file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log. Offset: 0
2018-05-24T08:58:02.232Z DEBUG [harvester] log/harvester.go:348 Update state: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log, offset: 0
2018-05-24T08:58:02.232Z DEBUG [harvester] log/harvester.go:447 Setting offset for file based on seek: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
2018-05-24T08:58:02.232Z DEBUG [harvester] log/harvester.go:433 Setting offset for file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log. Offset: 0
2018-05-24T08:58:02.232Z DEBUG [harvester] log/harvester.go:348 Update state: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log, offset: 0
2018-05-24T08:58:02.232Z INFO log/harvester.go:216 Harvester started for file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
2018-05-24T08:58:02.232Z INFO log/harvester.go:216 Harvester started for file: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log
2018-05-24T08:58:20.355Z DEBUG [harvester] log/log.go:85 End of file reached: /var/lib/docker/containers/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321/f6fde94852bb9b1dcfea85a0f4b2d813c7e0b062fcefce22b4261257d97f3321-json.log; Backoff now.
orgoz
(Olivier Boudet)
May 25, 2018, 8:49am
10
For your information, I've just upgraded to filebeat 6.2.4, same error occurs.
TimWard
(Tim Ward)
June 8, 2018, 1:03pm
11
I'm getting the same error with 6.2.4.
TimWard
(Tim Ward)
June 8, 2018, 1:18pm
12
... and it went away on restarting the Filebeat pods (with a configuration change, but only to do with the output).
orgoz
(Olivier Boudet)
June 8, 2018, 1:46pm
13
system
(system)
Closed
July 6, 2018, 1:47pm
14
This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.