Hi
I am using Elasticsearch version 7.8.0 , Logtash version 7.8.0
RDBMS : Oracle
I have configured Logstash to move data from Oracle to Elasticsearch of particular table on continuity basis.
Logstash.conf file -- >
   jdbc {
     jdbc_driver_library => "/mnt/logstash-7.8.0/config/ojdbc10.jar"
     jdbc_driver_class => "Java::oracle.jdbc.driver.OracleDriver"
     jdbc_connection_string => "jdbc:oracle:thin:@10.10.10.1:1521/orcl"
     jdbc_user => "scott"
     jdbc_password => "scott"
     jdbc_validate_connection => true
     jdbc_paging_enabled => true
     use_column_value => true
     tracking_column_type => "timestamp"
     tracking_column => "unix_ts_in_secs"
     clean_run => true
     last_run_metadata_path => "/mnt/logstash-7.8.0/config/employee_b1-sql_last_value.yml"
   # schedule => "*/15 * * * * *"
     statement => "select employee_id,employee_name,data_in_bytes,mime_type,mime_sub_type,created_date,status,UNIX2TS(last_updated_date) as unix_ts_in_secs from employee where last_updated_date > :sql_last_value and last_updated_date < systimestamp order by last_updated_date asc"
   }
 }
 filter {
   mutate {
     copy => { "employee_id" => "[@metadata][_id]"}
     remove_field => ["employee_id", "@version", "unix_ts_in_secs"]
   }
 }
 output {
  # stdout { codec => json_lines }
   elasticsearch {
     hosts => ["http://10.10.10.2:9200"]
     manage_template => false
     index => "employee_b1"
     document_id => "%{[@metadata][_id]}"
     user => "elastic"
     password => "elastic"
   }
 }```
When i run this config  , it runs properly but does not terminated and logstash is running in infinite loops and I am getting below error message :
  [2020-09-10T14:26:02,881][ERROR][logstash.javapipeline    ][main][6a31b22bffc61689e18837f701c826ce6cdd536cc374713273e8a1ad4aaee7a1] A plugin had an unrecoverable error. Will restart this plugin.
           Pipeline_id:main
           Plugin: <LogStash::Inputs::Jdbc jdbc_user=>"scott", last_run_metadata_path=>"/mnt/logstash-7.8.0/config/employee_b1-sql_last_value.yml", 
           jdbc_paging_enabled=>true, tracking_column_type=>"timestamp", use_column_value=>true, tracking_column=>"unix_ts_in_secs", 
           jdbc_validate_connection=>true, jdbc_password=><password>, statement=>"select employee_id,employee_name,data_in_bytes,mime_type,mime_sub_type,created_date,status,
           UNIX2TS(last_updated_date) as unix_ts_in_secs from employee where last_updated_date > :sql_last_value 
           and last_updated_date < systimestamp order by last_updated_date asc", clean_run=>true, jdbc_driver_library=>"/mnt/logstash-7.8.0/config/ojdbc10.jar", 
           jdbc_connection_string=>"jdbc:oracle:thin:@10.10.10.1:1521/orcl", id=>"6a31b22bffc61689e18837f701c826ce6cdd536cc374713273e8a1ad4aaee7a1", 
           jdbc_driver_class=>"Java::oracle.jdbc.driver.OracleDriver", enable_metric=>true, codec=><LogStash::Codecs::Plain id=>"plain_06dd4efa-2fcc-40ae-8dc2-93b1b0b9a98f", 
           enable_metric=>true, charset=>"UTF-8">, jdbc_page_size=>100000, jdbc_validation_timeout=>3600, jdbc_pool_timeout=>5, sql_log_level=>"info", 
           connection_retry_attempts=>1, connection_retry_attempts_wait_time=>0.5, plugin_timezone=>"utc", record_last_run=>true, lowercase_column_names=>true, 
           use_prepared_statements=>false>
           Error: no implicit conversion of BigDecimal into String
           Exception: TypeError
           Stack: uri:classloader:/META-INF/jruby.home/lib/ruby/stdlib/date/format.rb:335:in `_parse'
Can you please help me in understanding the reason of this error so that i can fix it.
Thank You 
Tushar Nemade