Elasticsearch/client.go:405 Cannot index event publisher.Event

Hello,

I'm trying to change my index name. I made some modifications to my .yml file including:

  • change index name
  • change index template
  • change index pattern
  • change index ilm related settings.

However, the following error is observed on the logs: "elasticsearch/client.go:405 Cannot index event publisher.Event". Nevertheless, I am able to see the index template and ilm policy name created on_Kibana but not the custom Index name neither custom Index pattern.

As an attachment the filebeat yml file.

###################### Filebeat Configuration Example #########################

# This file is an example configuration file highlighting only the most common
# options. The filebeat.reference.yml file from the same directory contains all the
# supported options with more comments. You can use it as a reference.
#
# You can find the full configuration reference here:
# Filebeat Reference | Elastic

# For more available modules and options, please see the filebeat.reference.yml sample
# configuration file.

# ============================== Filebeat inputs ===============================

filebeat.inputs:

# Each - is an input. Most options can be set at the input level, so
# you can use different inputs for various configurations.
# Below are the input specific configurations.

- type: log

** # Change to true to enable this input configuration.**
** enabled: true**

** # Paths that should be crawled and fetched. Glob based paths.**
** paths:**
** #- /var/log/.log**
** - C:/Users/admin/Desktop/Filebeat_Test_csv/
.csv**


** # Exclude lines. A list of regular expressions to match. It drops the lines that are**
** # matching any regular expression from the list.**
** #exclude_lines: ['^DBG']**

** # Include lines. A list of regular expressions to match. It exports the lines that are**
** # matching any regular expression from the list.**
** #include_lines: ['^ERR', '^WARN']**

** # Exclude files. A list of regular expressions to match. Filebeat drops the files that**
** # are matching any regular expression from the list. By default, no files are dropped.**
** #exclude_files: ['.gz$']**

** # Optional additional fields. These fields can be freely picked**
** # to add additional information to the crawled log files for filtering**
** #fields:**
** # level: debug**
** # review: 1**

** ### Multiline options**

** # Multiline can be used for log messages spanning multiple lines. This is common**
** # for Java Stack Traces or C-Line Continuation**

** # The regexp Pattern that has to be matched. The example pattern matches all lines starting with [**
** #multiline.pattern: ^[**

** # Defines if the pattern set under pattern should be negated or not. Default is false.**
** #multiline.negate: false**

** # Match can be set to "after" or "before". It is used to define if lines should be append to a pattern**
** # that was (not) matched before or after or as long as a pattern is not matched based on negate.**
** # Note: After is the equivalent to previous and before is the equivalent to to next in Logstash**
** #multiline.match: after**

# filestream is an input for collecting log messages from files. It is going to replace log input in the future.
- type: filestream

** # Change to true to enable this input configuration.**
** enabled: false**

** # Paths that should be crawled and fetched. Glob based paths.**
** paths:**
** - /var/log/*.log**


** # Exclude lines. A list of regular expressions to match. It drops the lines that are**
** # matching any regular expression from the list.**
** #exclude_lines: ['^DBG']**

** # Include lines. A list of regular expressions to match. It exports the lines that are**
** # matching any regular expression from the list.**
** #include_lines: ['^ERR', '^WARN']**

** # Exclude files. A list of regular expressions to match. Filebeat drops the files that**
** # are matching any regular expression from the list. By default, no files are dropped.**
** #prospector.scanner.exclude_files: ['.gz$']**

** # Optional additional fields. These fields can be freely picked**
** # to add additional information to the crawled log files for filtering**
** #fields:**
** # level: debug**
** # review: 1**

# ============================== Filebeat modules ==============================

filebeat.config.modules:
** # Glob pattern for configuration loading**
** path: ${path.config}/modules.d/*.yml**

** # Set to true to enable config reloading**
** reload.enabled: false**

** # Period on which files under path should be checked for changes**
** #reload.period: 10s**

# ======================= Elasticsearch template setting =======================

setup.template.enabled: false
setup.template.name: "rooms"
setup.template.pattern: "rooms-*"
setup.template.overwrite: true
setup.ilm.enabled: false
setup.ilm.policy_name: "rooms"
setup.ilm.rollover_alias: "rooms-%{[agent.version]}"
setup.template.settings:
** index.number_of_shards: 1**
** #index.codec: best_compression**
** #_source.enabled: false**

# ================================== General ===================================

# The name of the shipper that publishes the network data. It can be used to group
# all the transactions sent by a single shipper in the web interface.
name: Rooms

# The tags of the shipper are included in their own field with each
# transaction published.
tags: ["Rooms", "Mitel",]

# Optional fields that you can specify to add additional information to the
# output.
#fields:
# env: staging

# ================================= Dashboards =================================
# These settings control loading the sample dashboards to the Kibana index. Loading
# the dashboards is disabled by default and can be enabled either by setting the
# options here or by using the setup command.
#setup.dashboards.enabled: false

# The URL from where to download the dashboards archive. By default this URL
# has a value which is computed based on the Beat name and version. For released
# versions, this URL points to the dashboard archive on the artifacts.elastic.co
# website.
#setup.dashboards.url:

# =================================== Kibana ===================================

# Starting with Beats version 6.0.0, the dashboards are loaded via the Kibana API.
# This requires a Kibana endpoint configuration.
setup.kibana:

** # Kibana Host**
** # Scheme and port can be left out and will be set to the default (http and 5601)**
** # In case you specify and additional path, the scheme is required: http://localhost:5601/path**
** # IPv6 addresses should always be defined as: https://[2001:db8::1]:5601**
** #host: "localhost:5601"**

** # Kibana Space ID**
** # ID of the Kibana Space into which the dashboards should be loaded. By default,**
** # the Default Space will be used.**
** #space.id:**

# =============================== Elastic Cloud ================================

# These settings simplify using Filebeat with the Elastic Cloud (https://cloud.elastic.co/).

# The cloud.id setting overwrites the output.elasticsearch.hosts and
# setup.kibana.host options.
# You can find the cloud.id in the Elastic Cloud web UI.
#cloud.id:

# The cloud.auth setting overwrites the output.elasticsearch.username and
# output.elasticsearch.password settings. The format is <user>:<pass>.
#cloud.auth:

# ================================== Outputs ===================================

# Configure what output to use when sending the data collected by the beat.

# ---------------------------- Elasticsearch Output ----------------------------
output.elasticsearch:
** hosts: ["myhost:9200"]**
** index: "rooms-%{+yyyy.MM.dd}"**
** # Protocol - either http (default) or https.**
** protocol: "https"**

** # Authentication credentials - either API key or username/password.**
** #api_key: "id:api_key"**
** username: "test"**
** password: test**

# ------------------------------ Logstash Output -------------------------------
#output.logstash:
** # The Logstash hosts**
** #hosts: ["localhost:5044"]**

** # Optional SSL. By default is off.**
** # List of root certificates for HTTPS server verifications**
** #ssl.certificate_authorities: ["/etc/pki/root/ca.pem"]**

** # Certificate for SSL client authentication**
** #ssl.certificate: "/etc/pki/client/cert.pem"**

** # Client Certificate Key**
** #ssl.key: "/etc/pki/client/cert.key"**

# ================================= Processors =================================
processors:
** - add_host_metadata:**
** when.not.contains.tags: forwarded**
** - add_cloud_metadata: ~**
** - add_docker_metadata: ~**
** - add_kubernetes_metadata: ~**

# ================================== Logging ===================================

# Sets log level. The default log level is info.
# Available log levels are: error, warning, info, debug
#logging.level: debug

# At debug level, you can selectively enable logging only for some components.
# To enable all selectors use ["*"]. Examples of other selectors are "beat",
# "publisher", "service".
#logging.selectors: ["*"]

# ============================= X-Pack Monitoring ==============================
# Filebeat can export internal metrics to a central Elasticsearch monitoring
# cluster. This requires xpack monitoring to be enabled in Elasticsearch. The
# reporting is disabled by default.

# Set to true to enable the monitoring reporter.
#monitoring.enabled: false

# Sets the UUID of the Elasticsearch cluster under which monitoring data for this
# Filebeat instance will appear in the Stack Monitoring UI. If output.elasticsearch
# is enabled, the UUID is derived from the Elasticsearch cluster referenced by output.elasticsearch.
#monitoring.cluster_uuid:

# Uncomment to send the metrics to Elasticsearch. Most settings from the
# Elasticsearch output are accepted here as well.
# Note that the settings should point to your Elasticsearch monitoring cluster.
# Any setting that is not set is automatically inherited from the Elasticsearch
# output configuration, so if you have the Elasticsearch output configured such
# that it is pointing to your Elasticsearch monitoring cluster, you can simply
# uncomment the following line.
#monitoring.elasticsearch:

# ============================== Instrumentation ===============================

# Instrumentation support for the filebeat.
#instrumentation:
** # Set to true to enable instrumentation of filebeat.**
** #enabled: false**

** # Environment in which filebeat is running on (eg: staging, production, etc.)**
** #environment: ""**

** # APM Server hosts to report instrumentation results to.**
** #hosts:**
** # - http://localhost:8200**

** # API Key for the APM Server(s).**
** # If api_key is set then secret_token will be ignored.**
** #api_key:**

** # Secret token for the APM Server(s).**
** #secret_token:**

# ================================= Migration ==================================

# This allows to enable 6.7 migration aliases
#migration.6_to_7.enabled: true


This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.