Oracle Database monitoring by Elasticsearch, Logstash Pipeline

Hello Experts,

I am trying to monitor Oracle Database (which is on a Linux Server) by Elasticsearch. I have created a Logstash Config file which I am running on my own machine.

I have few questions; Please suggest / answer, It will be a great help

  1. Should I install Logstash on Database Server and run config there ?

  2. In my config file ,

    a. Is 'jdbc_connection_string' correct ?
    Info: "" is IP address of Oracle Database Server , "2222" is port of Database, "XYZ" is Database name

    b. Is 'jdbc_driver_library' correct ?
    Info: I have downloaded "ojdbc7-" file from internet and kept in "logstash-7.16.3-windows-x86_64" folder.

    c. Is 'jdbc_driver_class' correct ?

Config file : (Note: Ignore Output section)

input {
 			jdbc {
   				    jdbc_validate_connection => true
  			     	jdbc_connection_string => "jdbc:oracle:thin:@"
      				jdbc_user => "user_details"
      				jdbc_password => "user_password"
      				jdbc_driver_library => "C:logstash-7.16.3-windows-x86_64\ojdbc7-"
      				jdbc_driver_class => "Java::oracle.jdbc.driver.OracleDriver"
                    statement => "select * from search"
output {
 elasticsearch {
   					hosts => ["http://localhost:9200"]
   					index => "Ora_Elastic"
   					#user => "elastic"
  					#password => "changeme"

Ajit Shinde :pray:

This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.