Hi, I am attempting to use Logstash to import Metricbeat 'file output' files into Elasticsearch. I am using file output on Metricbeat because the servers I'm monitoring are on a different domain than my ES, Logstash, and Kibana installations, and direct connectivity is not possible (this is due to my current corporate network security constraints, and won't be changing soon...). I have the Metricbeat file output configured to roll files every 20 mb, and I'll be using Powershell scripts to migrate those completed output files into a staging directory on my Logstash server.
My question is, how can I configure Logstash to consume these files, create indexes based off the event timestamps, and drop the data into the appropriate indexes? The Metricbeat file output page in the documentation said it was possible to consume those files with Logstash, but offered no guidance on how to configure the pipeline.
I am running 5.5.1 for all ELK components. Any help would be greatly appreciated