In amonaly detection I get an actual value of "75.0" when I go to a visualization of the raw data I cant find that actual value, I was thinking that maybe the agregation was blurring the value, but I choose a smaller interval in the visualization to get the raw data and still is far from the actual, the max value shown in the visualization is 35
values over 0.7 are from november and I got the alert with timestamp "2020-12-04T12:30:00.000Z"
another weird thing is that when I click on the link of the watcher mail to go to the anomaly explorer, it doesnt match the top records that comes in the mail
Hi Rich, Thats one of the problems the actual value is not there when I search for it on Discover, but when execute the watch I get the value 0.75 which later in the script I tranform to 75 multiplying it by 100
The timestamp of 1606115700 is from November 23rd, not Dec 4th as you're showing in Discover.
Are you sure you're looking back that far in Discover? Also keep in mind that timestamp is in GMT and you may need to adjust for your timezone given that by default Kibana will show you times in your local time zone.
Thanks for your answer Rich, but I dont understand, that is the range that the query returns in the bucket result, that should be the range on where the bucket anomaly was found, or Im wrong?
All that I was saying is that you showed me a screenshot of the .ml-anomalies-* index and it showed a record from Nov 23rd. Your screenshot of Discover showed Dec 4th. These are obviously two different dates and I wanted to make sure you realized that if you were looking for this specific anomaly record that you had the wrong date.
In your latest screenshot, you're showing a bucket result (result_type:bucket) with a timestamp of 1607085000 which corresponds to Friday, December 4, 2020 12:30:00 PM GMT, but bucket results don't list actual values, only result_type:record results do.
I think I would need to see your entire Watcher configuration to know exactly what you're attempting to do, but I would think that putting in a range query for the last 30 days is not philosophically compatible with real-time alerting. Why look back so far in time?
Normally, you'd run the Watch every X minutes and look back in time roughly for a similar span of time (i.e. about X minutes as well, to avoid picking up the same alert conditions over and over again).
Setting the frequency of the Watch execution with respect to the look-back window of time is a little trickier with ML job results because of the fact that ML job results are written with a timestamp resolution of bucket_span and the value of the timestamp is the leading edge of the bucket. Therefore, there's a general rule that the look-back window of the watch should be about 2*bucket_span of the ML job so that no results are missed.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.