Using jdbc input plugin in logstash, I have fetched all the data which I have wanted the most from database tables.
Now I want only updated data from all that db tables. Every time I don’t want to index whole data, rather I want to index only updated data from db tables and add that indexes to old index.
Is this possible using elastic search? And If yes kindly guide me how we can achieve this?
I have last modified timestamp column in our database for some tables.
So currently I have fetched data from more than 5 tables and for each sql query there is a separate SQL file. So every time I have to go each and every query and need to change last modified date?And what if my data does not have any column for last modified date? Not all tables has timestamp column in my data.
Not sure I understood. But that may be a question to ask in #logstash instead?
If your goal is to sync a SQL based application with elasticsearch, I'd recommend reading this blog post where I shared most of my thoughts.
Basically, I'd recommend modifying the application layer if possible and send data to elasticsearch in the same "transaction" as you are sending your data to the database.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.