%3CLINGO-SUB%20id%3D%22lingo-sub-749190%22%20slang%3D%22en-US%22%3EExperiencing%20Alerting%20failure%20and%20Latency%20in%20Detecting%20Log%20Alerts%20for%20Log%20Search%20Alerts%20-%20Resolved%3C%2FLINGO-SUB%3E%3CLINGO-BODY%20id%3D%22lingo-body-749190%22%20slang%3D%22en-US%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EFinal%20Update%3C%2FU%3E%3A%20Thursday%2C%2011%20July%202019%2005%3A06%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe've%20confirmed%20that%20all%20systems%20are%20back%20to%20normal%20with%20no%20customer%20impact%20as%20of%2007%2F11%2C%2004%3A30%20UTC.%20Our%20logs%20show%20the%20incident%20started%20on%2007%2F11%2C%2001%3A30%20UTC%20and%20that%20during%20the%203%20hours%20that%20it%20took%20to%20resolve%20the%20issue%2080%25%20of%20customers%20using%20Log%20alerts%20which%20are%20configured%20in%20East%20Japan%20region%20would%20have%20experienced%20latency%20in%20detecting%20log%20alerts%20and%20execution%20of%20Log%20Search%20queries%20might%20have%20timed%20out.%3CBR%20%2F%3E%3CUL%3E%0A%20%3CLI%3E%3CU%3ERoot%20Cause%3C%2FU%3E%3A%26nbsp%3BThe%20failure%20was%20due%20to%20issues%20with%20one%20of%20our%20backend%20services%20which%20went%20in%20to%20unhealthy%20state.%3C%2FLI%3E%0A%20%3CLI%3E%3CU%3EIncident%20Timeline%3C%2FU%3E%3A%203%20Hours%20-%2007%2F11%2C%2001%3A30%20UTC%20through%2007%2F11%2C%2004%3A30%20UTC%3C%2FLI%3E%0A%3C%2FUL%3EWe%20understand%20that%20customers%20rely%20on%20Log%20Search%20Alerts%20as%20a%20critical%20service%20and%20apologize%20for%20any%20impact%20this%20incident%20caused.%3CBR%20%2F%3E%3CBR%20%2F%3E-Rama%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EInitial%20Update%3C%2FU%3E%3A%20Thursday%2C%2011%20July%202019%2004%3A12%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe%20are%20aware%20of%20issues%20within%20Log%20Search%20Alerts%20and%20are%20actively%20investigating.%20Some%20customers%20may%20experience%20Alerting%20failures%20and%20data%20latency%20issues.%3C%2FDIV%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CBR%20%2F%3E%3C%2FDIV%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CUL%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2007%2F11%2007%3A30%20UTC%3C%2FLI%3E%3C%2FUL%3EWe%20are%20working%20hard%20to%20resolve%20this%20issue%20and%20apologize%20for%20any%20inconvenience.%3CBR%20%2F%3E-Leela%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FLINGO-BODY%3E%3CLINGO-LABS%20id%3D%22lingo-labs-749190%22%20slang%3D%22en-US%22%3E%3CLINGO-LABEL%3ELog%20Search%20Alerts%3C%2FLINGO-LABEL%3E%3C%2FLINGO-LABS%3E
Final Update: Thursday, 11 July 2019 05:06 UTC

We've confirmed that all systems are back to normal with no customer impact as of 07/11, 04:30 UTC. Our logs show the incident started on 07/11, 01:30 UTC and that during the 3 hours that it took to resolve the issue 80% of customers using Log alerts which are configured in East Japan region would have experienced latency in detecting log alerts and execution of Log Search queries might have timed out.
  • Root Cause: The failure was due to issues with one of our backend services which went in to unhealthy state.
  • Incident Timeline: 3 Hours - 07/11, 01:30 UTC through 07/11, 04:30 UTC
We understand that customers rely on Log Search Alerts as a critical service and apologize for any impact this incident caused.

-Rama

Initial Update: Thursday, 11 July 2019 04:12 UTC

We are aware of issues within Log Search Alerts and are actively investigating. Some customers may experience Alerting failures and data latency issues.

  • Next Update: Before 07/11 07:30 UTC
We are working hard to resolve this issue and apologize for any inconvenience.
-Leela