Sure i can explain more
I am using FScrawler for crawling PDF files content and custom metadata into elastic search already created index for fast search. PDF files has too many metadata field so i have used ingest pipeline with rename and remove processors that worked fine.
Now i have the only issue that i have multi value field in my elastic search index with nested type i.e.
"Tags":
{
"type": "nested",
"properties": {
"Tag": { "type": "keyword" }}
}
ingest pipeline mapping
PUT _ingest/pipeline/my_mapping
{
"processors": [
{
"rename": {
"field": "meta.raw.Tags",
"target_field": "Tags",
"ignore_missing": true
}
},
{
"remove":{
"field": ["meta"]
}
}
]
}
FSCrawler generated data
"_source" : {
"content" : "my test content",
"meta" : {
"date" : "2019-12-12T10:09:13.000+0000",
"format" : "application/pdf; version=1.4",
"created" : "2017-01-12T10:03:50.000+0000",
"raw" : {
"date" : "2019-12-12T15:09:13Z",
"pdf:PDFVersion" : "1.4",
"access_permission:can_print_degraded" : "true",
"pdfa:PDFVersion" : "A-1a",
"dc:format" : "application/pdf; version=1.4",
"access_permission:fill_in_form" : "true",
"pdf:encrypted" : "false",
"modified" : "2019-12-12T15:09:13Z",
"Status" : "1",
"SeqNo" : "2",
"created" : "2017-01-12T15:03:50Z",
"access_permission:extract_for_accessibility" : "true",
"Creation-Date" : "2017-01-12T15:03:50Z",
"25107-0208-ComponentReleaseCertificate-20090207-5304-201912110909541826",
"pdfaid:part" : "1",
"OCR" : "1",
"Tags" : "\"Tags\": [{\"Tag\":\"Ford\"}, {\"Tag\":\"BMW\"}, {\"Tag\":\"Fiat\"}]",
}
}
}
}
after running fscrawler an error occured
ElasticsearchException[Elasticsearch exception [type=mapper_parsing_exception, reason=object mapping for [Tags] tried to parse field [Tags] as object, but found a concrete value]]
Please guide me how to solve this problem