I figure this may be due to the Python library having integration issues with Elasticsearch. By default I'd expect having the mapping "ignore_malformed: true" to keep bulk indexing even when hitting fields with geo_shape errors.
But it seems in the Python library an exception is raised by default, crashing your program, regardless of your mapping. Which is counter intuitive.
I've circumvented this issue by including the parameter: "raise_on_error=False" in my Python code. Runs to completion now even with issue documents.
In the ingest pipeline that ignore malformed means continue processing the ingest pipeline there could be many steps after that it means don't fail on that step And stop the whole processing.
Ingest pipelines are pre-writing if the actual document into the index.
When the document is actually written to the index is where a mapping exception takes place.
My suspicion is what's happening is it's malformed in such a way that when it went to Write the actual document there was a mapping exception which I think you may already have figured out I just thought I would remind anyone reading this that pipelines run before writing and mapping exceptions happen on writing.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.