I now want to ingest a larger number of documents using machine learning and importing a CSV. I have imported the CSV file, set the index name and set the correct mappings. However, when it comes to the Ingest Pipeline field, I am at a loss as to what I need to put in there.
It's evident that it needs some sort of properly formatted query language, but what exactly I do not know. Apologies, I am just starting my ELK journey, and I'm not very familiar with the query language used!
I am guessing you are referring to the CSV upload feature in the data visualization section of ML.
In the pipeline section, you can simply add your enrichment processor as the last processor.
"processors": [ // Might already be present in the pipeline section
... //processors defined automatically by ml (if present)
{// Your enrich processor
"enrich": {
"policy_name": "postal_policy",
"field": "geo_location",
"target_field": "geo_data",
"shape_relation": "INTERSECTS"
}
}
]
I am curious, how were you testing via the kibana console? I am guessing you created your pipeline and used _simulate. If you did, then you should be able to copy paste your processor definition into the pipeline in the CSV Uploader.
I am curious, how were you testing via the kibana console? I am guessing you created your pipeline and used _simulate . If you did, then you should be able to copy paste your processor definition into the pipeline in the CSV Uploader.
I just followed the instructions in the example I posted earlier, which involved using PUT commands to create the geo-shap index, enrichment policy, and then index a document and specifiy the enrichment policy as the ingest pipeline.
When using the console, this sort of error is flagged. I wonder whether it's worthwhile extending this sort of error detection to other places, like the ingest pipeline window as an example.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.