We've confirmed that all systems are back to normal with no customer impact as of 4/19, 23:50 UTC. Our logs show the incident started on 4/19, 19:00 UTC.
Root Cause: The failure was due to a service entering an unhealthy state due to a unanticipated spike in traffic to the service.
Incident Timeline: 4 Hours & 50 minutes, 4/19 19:00 UTC to 4/19 23:50 UTC.
We understand that customers rely on Log Search Alerts as a critical service and apologize for any impact this incident caused.
Initial Update: Friday, 19 April 2019 22:39 UTC We are aware of issues within Log Search Alerts and are actively investigating. Customers who created Log Search Alerts in the East US region after 2019-04-19 19:00 UTC or in the West Central US region starting from 2019-04-19 20:00 UTC may not see these tests begin to run for a much longer period than normal.
Work Around: Log Search Alerts from other regions are not impacted.
Next Update: Before 04/20 01:00 UTC
We are working hard to resolve this issue and apologize for any inconvenience. -Matthew Cosner