For now, disable the elasticsearch output and insert a stdout { codec => rubydebug } output. Look in the logs. Warning messages are emitted for various kinds of error conditions. If there's nothing interesting there, increase logging verbosity by starting Logstash with --verbose or --debug. Look in the logs again. You should e.g. see every query that's sent for execution.
Its now working successfully 
. My error is, the plug in was not completely updated. Thank you very much Magnus. See you on another thread.
Regards,
Croos.
How to update the existing document by schedule,  when a MySQL table row is updated?
Example:
+---+------+---------+
|id | name | city    |
+---+------+---------+
| 11| croos| chennai |
| 22| nilu | colombo |
+---+------+---------+
+---+------+---------+
|id | name | city    |
+---+------+---------+
| 11| croos| trichy  |
| 22| nilu | colombo |
+---+------+---------+
In this case how to update city on elasticsearch by schedule?
I answered a very similar question in the past couple of days. Set elasticsearch's document_id option to the desired id of the document (perhaps the "id" column) instead of letting Elasticsearch allocate a random id. Then Logstash will be able to update the existing document. You may have to modify the action option too.
Would you please give me some example links..
I had an  example from your previous answer. For now, this is enough to me.  Thanks and catch you soon 
 ..!