I have an ingest pipeline "filebeat-7.2.0-apache-access-default" and others, however when my document doesnt match something it doesnt get indexed. Is there any way that i can define that even if any pipeline fails index the document anyway ?
@spinscale since i have no control over these pipelines (they get created by filebeat etc...) what i am looking for is a way to not have to change them so that the "index even if it fails" mentality would be applied to all pipelines. is there anything of that kind ?
do you care to explain your use-case a little bit further? Are you trying to manually index data or use another filebeat instance? What kind of data are you indexing? I'm curious where this starts to differ from the 'general' filebeat use-case or if there is an general issue on the filebeat/elasticsearch side of things.
@spinscale I'm using a typical ELK stack with beats (all OSS) and adding all the pipelines etc.. via the setup command. beats (filebeat for example) is collecting (via the apache module and others) and sending to logs to Logstash.
What i am verifying is that my apache logs are not passing the pipeline saying that some fields dont exist and therefore are not being indexed. I dont want this to happen, if it passes the pipeline, fine, if not then atleast they should be indexed that i dont lose them.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.