%3CLINGO-SUB%20id%3D%22lingo-sub-863717%22%20slang%3D%22en-US%22%3EExperiencing%20Metric%20Alerts%20Issue%20for%20Log%20Analytics%20in%20East%20US%20region%20-%2009%2F19%20-%20Resolved%3C%2FLINGO-SUB%3E%3CLINGO-BODY%20id%3D%22lingo-body-863717%22%20slang%3D%22en-US%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EFinal%20Update%3C%2FU%3E%3A%20Friday%2C%2020%20September%202019%2010%3A33%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe've%20confirmed%20that%20all%20systems%20are%20back%20to%20normal%20with%20no%20customer%20impact%20as%20of%2009%2F20%2C%2010%3A20%20UTC.%20Our%20logs%20show%20the%20incident%20started%20on%2009%2F19%2C%2009%3A00%20UTC%20and%20that%20during%20the%2025%20hours%2020%20minutes%20that%20it%20took%20to%20resolve%20the%20issue%20some%20customers%20in%20East%20US%20region%20might%20have%20experienced%20alerting%20delays%20and%20would%20not%20had%20received%20metric%20alerts.%3CBR%20%2F%3E%3CUL%3E%3CLI%3E%3CU%3ERoot%20Cause%3C%2FU%3E%3A%20The%20failure%20was%20due%20to%20an%20issue%20in%20one%20of%20our%20dependent%20service.%3C%2FLI%3E%3CLI%3E%3CU%3EIncident%20Timeline%3C%2FU%3E%3A%2025%20Hours%20%26amp%3B%2020%20minutes%20-%2009%2F19%2C%2009%3A00%20UTC%20through%2009%2F20%2C%2010%3A20%20UTC%3C%2FLI%3E%3C%2FUL%3EWe%20understand%20that%20customers%20rely%20on%20Azure%20Log%20Analytics%20as%20a%20critical%20service%20and%20apologize%20for%20any%20impact%20this%20incident%20caused.%3CBR%20%2F%3E%3CBR%20%2F%3E-Monish%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EUpdate%3C%2FU%3E%3A%20Friday%2C%2020%20September%202019%2007%3A41%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe%20continue%20to%20investigate%20issues%20within%20Log%20Analytics.%20Root%20cause%20is%20not%20fully%20understood%20at%20this%20time.%20Some%20customers%20in%20East%20Us%20Region%20continue%20to%20experience%20delayed%20or%20might%20not%20receive%20metric%20alerts.We%20currently%20have%20no%20estimate%20for%20resolution.%3CBR%20%2F%3E%3CUL%3E%3CLI%3E%3CU%3EWork%20Around%3C%2FU%3E%3A%3CNONE%20or%3D%22%22%20details%3D%22%22%3E%3C%2FNONE%3E%3C%2FLI%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2009%2F20%2012%3A00%20UTC%3C%2FLI%3E%3C%2FUL%3E-Monish%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EUpdate%3C%2FU%3E%3A%20Friday%2C%2020%20September%202019%2002%3A27%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe%20continue%20to%20investigate%20issues%20within%20Log%20Analytics%20in%20East%20US%20region.%20Root%20cause%20is%20not%20fully%20understood%20at%20this%20time%20and%20we%20are%20working%20on%20resolution.%20Some%20customers%20continue%20to%20experience%20delay%20or%20may%20not%20be%20receiving%20metric%20alerts%20in%20East%20US.%20We%20currently%20have%20no%20estimate%20for%20resolution.%3CUL%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2009%2F20%2007%3A30%20UTC%3C%2FLI%3E%3C%2FUL%3E-Jayadev%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EUpdate%3C%2FU%3E%3A%20Thursday%2C%2019%20September%202019%2022%3A12%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe%20continue%20to%20investigate%20issues%20within%20Log%20Analytics%20in%20East%20US%20region%2C%20we%20encountered%20another%20issue%20along%20the%20process%20of%20mitigation%20and%20we%20are%20working%20on%20resolution.%20Some%20customers%20continue%20to%20experience%20%3CSPAN%20style%3D%22display%3A%20inline%20!important%3B%20float%3A%20none%3B%20background-color%3A%20rgb(255%2C%20255%2C%20255)%3B%20color%3A%20rgb(0%2C%200%2C%200)%3B%20font-family%3A%20%26quot%3BHelvetica%20Neue%26quot%3B%2CHelvetica%2CArial%2Csans-serif%3B%20font-size%3A%2014px%3B%20font-style%3A%20normal%3B%20font-variant%3A%20normal%3B%20font-weight%3A%20400%3B%20letter-spacing%3A%20normal%3B%20orphans%3A%202%3B%20text-align%3A%20left%3B%20text-decoration%3A%20none%3B%20text-indent%3A%200px%3B%20text-transform%3A%20none%3B%20-webkit-text-stroke-width%3A%200px%3B%20white-space%3A%20normal%3B%20word-spacing%3A%200px%3B%22%3Edelay%20or%20may%20not%20be%20receiving%20alerts%3C%2FSPAN%3E.%20We%20are%20working%20to%20establish%20the%20start%20time%20for%20the%20issue%2C%20initial%20findings%20indicate%20that%20the%20problem%20began%20at%2009%2F19%20~09%3A00%20UTC.%20We%20currently%20have%20no%20estimate%20for%20resolution.%3CUL%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2009%2F20%2002%3A30%20UTC%3C%2FLI%3E%3C%2FUL%3E-Jayadev%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EUpdate%3C%2FU%3E%3A%20Thursday%2C%2019%20September%202019%2017%3A57%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3ERoot%20cause%20has%20been%20isolated%20to%20failure%20in%20deployment%20which%20was%20impacting%20metric%20alerts%20to%20be%20triggered%20for%20Log%20Analytics%20in%20East%20US%20region.%20To%20address%20this%20issue%20we%20scaled%20out%20our%20service%20for%20faster%20processing%20of%20data%20that%20was%20stuck.%20Data%20flow%20is%20now%20working%20as%20expected.%20Some%20customers%20may%20experience%20either%20delay%20or%20may%20not%20be%20receiving%20alerts%20and%20we%20estimate%20~%203%20hours%20before%20all%20metric%20alerts%20start%20working%20as%20expected.%3CUL%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2009%2F19%2022%3A00%20UTC%3C%2FLI%3E%3C%2FUL%3E-Jayadev%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FLINGO-BODY%3E%3CLINGO-LABS%20id%3D%22lingo-labs-863717%22%20slang%3D%22en-US%22%3E%3CLINGO-LABEL%3EAzure%20Log%20Analytics%3C%2FLINGO-LABEL%3E%3C%2FLINGO-LABS%3E
Final Update: Friday, 20 September 2019 10:33 UTC

We've confirmed that all systems are back to normal with no customer impact as of 09/20, 10:20 UTC. Our logs show the incident started on 09/19, 09:00 UTC and that during the 25 hours 20 minutes that it took to resolve the issue some customers in East US region might have experienced alerting delays and would not had received metric alerts.
  • Root Cause: The failure was due to an issue in one of our dependent service.
  • Incident Timeline: 25 Hours & 20 minutes - 09/19, 09:00 UTC through 09/20, 10:20 UTC
We understand that customers rely on Azure Log Analytics as a critical service and apologize for any impact this incident caused.

-Monish

Update: Friday, 20 September 2019 07:41 UTC

We continue to investigate issues within Log Analytics. Root cause is not fully understood at this time. Some customers in East Us Region continue to experience delayed or might not receive metric alerts.We currently have no estimate for resolution.
  • Work Around:
  • Next Update: Before 09/20 12:00 UTC
-Monish

Update: Friday, 20 September 2019 02:27 UTC

We continue to investigate issues within Log Analytics in East US region. Root cause is not fully understood at this time and we are working on resolution. Some customers continue to experience delay or may not be receiving metric alerts in East US. We currently have no estimate for resolution.
  • Next Update: Before 09/20 07:30 UTC
-Jayadev

Update: Thursday, 19 September 2019 22:12 UTC

We continue to investigate issues within Log Analytics in East US region, we encountered another issue along the process of mitigation and we are working on resolution. Some customers continue to experience delay or may not be receiving alerts. We are working to establish the start time for the issue, initial findings indicate that the problem began at 09/19 ~09:00 UTC. We currently have no estimate for resolution.
  • Next Update: Before 09/20 02:30 UTC
-Jayadev

Update: Thursday, 19 September 2019 17:57 UTC

Root cause has been isolated to failure in deployment which was impacting metric alerts to be triggered for Log Analytics in East US region. To address this issue we scaled out our service for faster processing of data that was stuck. Data flow is now working as expected. Some customers may experience either delay or may not be receiving alerts and we estimate ~ 3 hours before all metric alerts start working as expected.
  • Next Update: Before 09/19 22:00 UTC
-Jayadev