[2019-09-04T01:21:03,232][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{"other"=>[{"thread_id"=>34, "name"=>"[main]<file", "current_call"=>"[...]/vendor/bundle/jruby/2.3.0/gems/logstash-input-file-4.1.5/lib/filewatch/watched_files_collection.rb:80:in `[]='"}], ["Lo
gStash::Filters::Mutate", {"remove_field"=>["message"], "id"=>"894418ee323961c3a2605309bcb5a8b97fdf98c9a3d7b233f8d3ef9d02e018d3"}]=>[{"thread_id"=>26, "name"=>nil, "current_call"=>"[...]/logstash-core/lib/logstash/pipeline.rb:315:in `read_batch'"}, {"thread_id"=>27, "name"=>nil, "current_call"=>"[...]/logstash-core/
lib/logstash/pipeline.rb:315:in `read_batch'"}, {"thread_id"=>28, "name"=>nil, "current_call"=>"[...]/logstash-core/lib/logstash/pipeline.rb:315:in `read_batch'"}, {"thread_id"=>29, "name"=>nil, "current_call"=>"[...]/logstash-core/lib/logstash/pipeline.rb:315:in `read_batch'"}, {"thread_id"=>30, "name"=>nil, "curre
nt_call"=>"[...]/logstash-core/lib/logstash/pipeline.rb:315:in `read_batch'"}, {"thread_id"=>31, "name"=>nil, "current_call"=>"[...]/logstash-core/lib/logstash/pipeline.rb:315:in `read_batch'"}, {"thread_id"=>32, "name"=>nil, "current_call"=>"[...]/logstash-core/lib/logstash/pipeline.rb:315:in `read_batch'"}, {"thre
ad_id"=>33, "name"=>nil, "current_call"=>"[...]/logstash-core/lib/logstash/pipeline.rb:315:in `read_batch'"}]}}
[2019-09-04T01:21:31,236][INFO ][logstash.runner ] Starting Logstash {"logstash.version"=>"6.4.0"}
[2019-09-04T01:21:39,212][WARN ][logstash.outputs.elasticsearch] You are using a deprecated config setting "document_type" set in elasticsearch. Deprecated settings will continue to work, but are scheduled for removal from logstash in the future. Document types are being deprecated in Elasticsearch 6.0, and removed
entirely in 7.0. You should avoid this feature If you have any questions about this, please visit the #logstash channel on freenode irc. {:name=>"document_type", :plugin=><LogStash::Outputs::ElasticSearch index=>"lcch-cisco-%{+YYYY.MM.dd}", manage_template=>false, id=>"d5222d947f76b01bc62c5d5f284541324145e7e29df6698
7f5f8d756dcf5ad65", document_id=>"%{pkid}", hosts=>[//localhost:9200], document_type=>"%{type}", enable_metric=>true, codec=><LogStash::Codecs::Plain id=>"plain_9f87a70c-3780-472a-afcb-5bb43f04f845", enable_metric=>true, charset=>"UTF-8">, workers=>1, template_name=>"logstash", template_overwrite=>false, doc_as_upse
rt=>false, script_type=>"inline", script_lang=>"painless", script_var_name=>"event", scripted_upsert=>false, retry_initial_interval=>2, retry_max_interval=>64, retry_on_conflict=>1, action=>"index", ssl_certificate_verification=>true, sniffing=>false, sniffing_delay=>5, timeout=>60, pool_max=>1000, pool_max_per_rout
e=>100, resurrect_delay=>5, validate_after_inactivity=>10000, http_compression=>false>}
[2019-09-04T01:21:39,280][INFO ][logstash.pipeline ] Starting pipeline {:pipeline_id=>"main", "pipeline.workers"=>8, "pipeline.batch.size"=>125, "pipeline.batch.delay"=>50}
[2019-09-04T01:21:39,833][INFO ][logstash.outputs.elasticsearch] Elasticsearch pool URLs updated {:changes=>{:removed=>[], :added=>[http://localhost:9200/]}}
[2019-09-04T01:21:39,847][INFO ][logstash.outputs.elasticsearch] Running health check to see if an Elasticsearch connection is working {:healthcheck_url=>http://localhost:9200/, :path=>"/"}
[2019-09-04T01:21:40,092][WARN ][logstash.outputs.elasticsearch] Restored connection to ES instance {:url=>"http://localhost:9200/"}
[2019-09-04T01:21:40,176][INFO ][logstash.outputs.elasticsearch] ES Output version determined {:es_version=>6}
[2019-09-04T01:21:40,180][WARN ][logstash.outputs.elasticsearch] Detected a 6.x and above cluster: the `type` event field won't be used to determine the document _type {:es_version=>6}
[2019-09-04T01:21:40,213][INFO ][logstash.outputs.elasticsearch] New Elasticsearch output {:class=>"LogStash::Outputs::ElasticSearch", :hosts=>["//localhost:9200"]}
[2019-09-04T01:21:40,617][INFO ][logstash.inputs.file ] No sincedb_path set, generating one based on the "path" setting {:sincedb_path=>"/var/lib/logstash/plugins/inputs/file/.sincedb_697fd033fdfc9bb6ccfac8a56026cf37", :path=>["/home/ftp/cdr*"]}
[2019-09-04T01:21:40,656][INFO ][logstash.inputs.file ] No sincedb_path set, generating one based on the "path" setting {:sincedb_path=>"/var/lib/logstash/plugins/inputs/file/.sincedb_d0f4e6f86e69bd93fda203a6645e41bd", :path=>["/home/ftp/cmr*"]}
[2019-09-04T01:21:40,684][INFO ][logstash.pipeline ] Pipeline started successfully {:pipeline_id=>"main", :thread=>"#<Thread:0xcdbbfe9 run>"}
[2019-09-04T01:21:40,805][INFO ][filewatch.observingread ] START, creating Discoverer, Watch with file and sincedb collections
[2019-09-04T01:21:40,807][INFO ][filewatch.observingread ] START, creating Discoverer, Watch with file and sincedb collections
[2019-09-04T01:21:40,817][INFO ][logstash.agent ] Pipelines running {:count=>1, :running_pipelines=>[:main], :non_running_pipelines=>[]}
[2019-09-04T01:21:41,258][INFO ][logstash.agent ] Successfully started Logstash API endpoint {:port=>9600}
root@elk:/var/log/logstash#