I wondering about my pipeline setup:
- pipeline.id: bal
path.config: "/usr/share/logstash/pipeline/1*.conf"
pipeline.workers: 24
pipeline.batch.size: 2000
- pipeline.id: soul
path.config: "/usr/share/logstash/pipeline/2*.conf"
pipeline.workers: 24
pipeline.batch.size: 2000
- pipeline.id: outputs
path.config: "/usr/share/logstash/pipeline/9*.conf"
pipeline.workers: 8
pipeline.batch.size: 2000
For instance, i have set pipeline workers to 24 for input and parsing, for two independent redis inputs, with a output set to 8 workers. My question starts here. am I creating a bottleneck by having 24 works for input and filtering while output has only 8?