Hello,
I using logstash jdbc for pulling data from mongodb and pushing to elasticsearch. I using this pipeline configuration for logstash:
input {
jdbc {
type => "bb_purchaselog"
jdbc_driver_library => "/usr/share/logstash/logstash-core/lib/jars/mongojdbc3.1.jar"
jdbc_driver_class => "com.dbschema.MongoJdbcDriver"
jdbc_connection_string => "jdbc connection string"
jdbc_user => dbuser
jdbc_password => "pass"
# jdbc_paging_enabled => true
# tracking_column => "transaction_time"
last_run_metadata_path => "/usr/share/logstash/lastrun/bb-purchaselog-lastrun.yml"
# use_column_value => true
# tracking_column_type => "numeric"
schedule => "1 */1 * * * *"
statement_filepath => "/usr/share/logstash/query/bb-purchaselog.js"
}
.
.
.
}
My configuration works correctly but when I enabling tracking_column and use_column_value features, logstash jdbc returns some error about "can't finding tracking_column value"
I have three question:
-
can I using tracking_column and use_column_value feature in jdbc input plugin when I using mongojdbc3.1.jar?
-
If I can't using these two features, is there way to set :sql_last_value value in input or filter stage in logstash config?
-
If we havn't any solution for two above questions, can you describing any risks about data loss when we not using use_column_value feature?
Thanks for any help.