How do your documents look like indexed in Elasticsearch?
For each word in a document you can use the Terms aggregation on the field that contains the text of your document. This will work only if the field was analysed by Elasticsearch at index time.
And you want to count a specific word across different documents, you just filter on that word from list that you get from the terms aggregation.
You can check the Shakespeare example in the Kibana Getting Started tutorial: https://www.elastic.co/guide/en/kibana/current/getting-started.html
Initially I have used a python code to to index my pdf file and recently I used fscrawler, both seems to be working and indexing according to the fields, I have provided a screenshot.
The screenshot I provided is the indexing done by fscrawler.
Suppose, from the screenshot if it says like content contains the words in document, when I try to visualize that kibana it doesn't even show in the split series.
For Example, I indexed my file using a python script, title.keyword is the title name and file.keyword is the words in the file. if I select title.keyword for vertical bar graph it shows the names of the different files i use and there count, so in the same way if I select the file.keyword it shows nothing as it takes the whole file.keyword as one thing.
I have tried in multiple ways but didn't succeed, please help!
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.