%3CLINGO-SUB%20id%3D%22lingo-sub-1405719%22%20slang%3D%22en-US%22%3EExperiencing%20Delayed%20alerts%20issue%20in%20Azure%20Portal%20in%20East%20US2%20and%20Central%20US%20-%2005%2F20%20-%20Resolved%3C%2FLINGO-SUB%3E%3CLINGO-BODY%20id%3D%22lingo-body-1405719%22%20slang%3D%22en-US%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EFinal%20Update%3C%2FU%3E%3A%20Wednesday%2C%2020%20May%202020%2013%3A39%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe've%20confirmed%20that%20all%20systems%20are%20back%20to%20normal%20with%20no%20customer%20impact%20as%20of%205%2F18%2C%2011%3A55%20UTC.%20Our%20logs%20show%20the%20incident%20started%20on%205%2F20%2C%2010%3A35%20UTC%20and%20that%20during%20the%201%20hours%20and%2020%20minutes%20that%20it%20took%20to%20resolve%20the%20issue%20some%20of%20customers%20may%20have%20experienced%20issues%20with%20delayed%20Metric%20Alerts%20and%20Activity%20Log%20Alerts%20in%20East%20US2%20and%20Central%20US%20regions.%3CBR%20%2F%3E%3CUL%3E%3CLI%3E%3CU%3ERoot%20Cause%3C%2FU%3E%3A%20The%20failure%20was%20due%20to%20an%20issue%20in%20one%20of%20our%20dependent%20services.%3C%2FLI%3E%3CLI%3E%3CU%3EIncident%20Timeline%3C%2FU%3E%3A%201%20Hours%20%26amp%3B%2020%20minutes%20-%205%2F20%2C%2010%3A35%20UTC%20through%205%2F20%2C%2011%3A55%20UTC%3C%2FLI%3E%3C%2FUL%3EWe%20understand%20that%20customers%20rely%20on%20Azure%20Monitor%20Alerts%20as%20a%20critical%20service%20and%20apologize%20for%20any%20impact%20this%20incident%20caused.%3CBR%20%2F%3E%3CBR%20%2F%3E-Madhuri%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FLINGO-BODY%3E%3CLINGO-LABS%20id%3D%22lingo-labs-1405719%22%20slang%3D%22en-US%22%3E%3CLINGO-LABEL%3EActivity%20Log%20Alerts%3C%2FLINGO-LABEL%3E%3CLINGO-LABEL%3EMetric%20Alerts%3C%2FLINGO-LABEL%3E%3C%2FLINGO-LABS%3E
Final Update: Wednesday, 20 May 2020 13:39 UTC

We've confirmed that all systems are back to normal with no customer impact as of 5/18, 11:55 UTC. Our logs show the incident started on 5/20, 10:35 UTC and that during the 1 hours and 20 minutes that it took to resolve the issue some of customers may have experienced issues with delayed Metric Alerts and Activity Log Alerts in East US2 and Central US regions.
  • Root Cause: The failure was due to an issue in one of our dependent services.
  • Incident Timeline: 1 Hours & 20 minutes - 5/20, 10:35 UTC through 5/20, 11:55 UTC
We understand that customers rely on Azure Monitor Alerts as a critical service and apologize for any impact this incident caused.

-Madhuri