%3CLINGO-SUB%20id%3D%22lingo-sub-1431184%22%20slang%3D%22en-US%22%3EExperiencing%20Data%20Access%20Issue%20in%20Azure%20portal%20for%20Log%20Analytics%20-%2006%2F01%20-%20Resolved%3C%2FLINGO-SUB%3E%3CLINGO-BODY%20id%3D%22lingo-body-1431184%22%20slang%3D%22en-US%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EFinal%20Update%3C%2FU%3E%3A%20Monday%2C%2001%20June%202020%2022%3A43%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe've%20confirmed%20that%20all%20systems%20are%20back%20to%20normal%20with%20no%20customer%20impact%20as%20of%206%2F1%2C%2021%3A50%20UTC.%20Our%20logs%20show%20the%20incident%20started%20on%206%2F1%2C%2012%3A05%20UTC%20and%20that%20during%20the%209%20hours%20and%2045%20minutes%20that%20it%20took%20to%20resolve%20the%20issue%2033%25%20of%20customers%20experienced%20possible%20query%20execution%20errors%2C%20as%20well%20as%20the%20misbehavior%20or%20alerts%20configured%20in%20Log%20Search%20alerts%2C%20and%20Smart%20Detection%20alerts.%3CBR%20%2F%3E%3CUL%3E%3CLI%3E%3CU%3ERoot%20Cause%3C%2FU%3E%3A%20The%20failure%20was%20due%20to%20a%20backend%20service%20getting%20into%20a%20bad%20state.%3C%2FLI%3E%3CLI%3E%3CU%3EIncident%20Timeline%3C%2FU%3E%3A%209%20Hours%20%26amp%3B%2045%20minutes%20-%206%2F1%2C%2012%3A05%20UTC%20through%206%2F1%2C%2021%3A50%20UTC%3C%2FLI%3E%3C%2FUL%3EWe%20understand%20that%20customers%20rely%20on%20Azure%20Log%20Analytics%20as%20a%20critical%20service%20and%20apologize%20for%20any%20impact%20this%20incident%20caused.%3CBR%20%2F%3E%3CBR%20%2F%3E-Jeff%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EUpdate%3C%2FU%3E%3A%20Monday%2C%2001%20June%202020%2022%3A13%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3ERoot%20cause%20has%20been%20isolated%20to%20a%20backend%20service%20that%20is%20in%20a%20bad%20state%26nbsp%3Bwhich%20was%20impacting%20query%20execution%2C%20Log%20Search%20alerts%2C%20and%20Smart%20Detection%20alerts.%20To%20address%20this%20issue%20we%20updated%20backend%20service%20code%20and%20it%20is%20now%20working%20as%20expected.%26nbsp%3B%3CBR%20%2F%3E%3CUL%3E%3CLI%3E%3CU%3EWork%20Around%3C%2FU%3E%3A%20None%3C%2FLI%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2006%2F01%2023%3A30%20UTC%3C%2FLI%3E%3C%2FUL%3E-Jeff%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EUpdate%3C%2FU%3E%3A%20Monday%2C%2001%20June%202020%2019%3A49%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe%20continue%20to%20investigate%20issues%20within%20Log%20Analytics.%20Root%20cause%20has%20been%20narrowed%20to%20a%20backend%20service%20that%20is%20in%20a%20bad%20state.%20Some%20customers%20continue%20to%20experience%20timeouts%20and%20failures%20when%20running%20queries%20and%20issues%20with%20missed%20or%20delayed%20Log%20Search%20and%20Smart%20Detection%20alerts%20hosted%20in%20multiple%20region.%20Initial%20findings%20indicate%20that%20the%20problem%20began%20at%2006%2F01%2012%3A22%20UTC.%20We%20currently%20have%20no%20estimate%20for%20resolution.%3CBR%20%2F%3E%3CUL%3E%3CLI%3E%3CU%3EWork%20Around%3C%2FU%3E%3A%20None%3C%2FLI%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2006%2F02%2000%3A00%20UTC%3C%2FLI%3E%3C%2FUL%3E-Jeff%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EUpdate%3C%2FU%3E%3A%20Monday%2C%2001%20June%202020%2016%3A37%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3E%3CDIV%3EWe%20continue%20to%20investigate%20issues%20within%20Log%20Analytics.%20Root%20cause%20is%20not%20fully%20understood%20at%20this%20time.%20Customer%20using%20Azure%20Monitor%20who%20may%20have%20continue%20to%20experience%20timeouts%20and%20failures%20when%20running%20queries%20and%20issues%20with%20missed%20or%20delayed%20Log%20Search%20and%20Smart%20Detection%20alerts%20hosted%20in%20multiple%20region.%20We%20currently%20have%20no%20estimate%20for%20resolution.%3C%2FDIV%3E%3CUL%3E%3CLI%3E%3CU%3EWork%20Around%3C%2FU%3E%3A%20None%3C%2FLI%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2006%2F01%2019%3A00%20UTC%3C%2FLI%3E%3C%2FUL%3E-Anmol%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EInitial%20Update%3C%2FU%3E%3A%20Monday%2C%2001%20June%202020%2014%3A29%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe%20are%20aware%20of%20issues%20within%20Log%20Analytics%20and%20are%20actively%20investigating.%20Some%20customers%20may%26nbsp%3Bexperience%20timeouts%20and%20failures%20when%20running%20queries%20and%20issues%20with%20missed%20or%20delayed%20Log%20Search%20and%20Smart%20Detection%20alerts%20hosted%20in%26nbsp%3BEast%20US%20region.%3CBR%20%2F%3E%3CUL%3E%3CLI%3E%3CU%3EWork%20Around%3C%2FU%3E%3A%20None%3C%2FLI%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2006%2F01%2016%3A30%20UTC%3C%2FLI%3E%3C%2FUL%3EWe%20are%20working%20hard%20to%20resolve%20this%20issue%20and%20apologize%20for%20any%20inconvenience.%3CBR%20%2F%3E-Anmol%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FLINGO-BODY%3E%3CLINGO-LABS%20id%3D%22lingo-labs-1431184%22%20slang%3D%22en-US%22%3E%3CLINGO-LABEL%3EAzure%20Log%20Analytics%3C%2FLINGO-LABEL%3E%3CLINGO-LABEL%3ELog%20Search%20Alerts%3C%2FLINGO-LABEL%3E%3C%2FLINGO-LABS%3E
Final Update: Monday, 01 June 2020 22:43 UTC

We've confirmed that all systems are back to normal with no customer impact as of 6/1, 21:50 UTC. Our logs show the incident started on 6/1, 12:05 UTC and that during the 9 hours and 45 minutes that it took to resolve the issue 33% of customers experienced possible query execution errors, as well as the misbehavior or alerts configured in Log Search alerts, and Smart Detection alerts.
  • Root Cause: The failure was due to a backend service getting into a bad state.
  • Incident Timeline: 9 Hours & 45 minutes - 6/1, 12:05 UTC through 6/1, 21:50 UTC
We understand that customers rely on Azure Log Analytics as a critical service and apologize for any impact this incident caused.

-Jeff

Update: Monday, 01 June 2020 22:13 UTC

Root cause has been isolated to a backend service that is in a bad state which was impacting query execution, Log Search alerts, and Smart Detection alerts. To address this issue we updated backend service code and it is now working as expected. 
  • Work Around: None
  • Next Update: Before 06/01 23:30 UTC
-Jeff

Update: Monday, 01 June 2020 19:49 UTC

We continue to investigate issues within Log Analytics. Root cause has been narrowed to a backend service that is in a bad state. Some customers continue to experience timeouts and failures when running queries and issues with missed or delayed Log Search and Smart Detection alerts hosted in multiple region. Initial findings indicate that the problem began at 06/01 12:22 UTC. We currently have no estimate for resolution.
  • Work Around: None
  • Next Update: Before 06/02 00:00 UTC
-Jeff

Update: Monday, 01 June 2020 16:37 UTC

We continue to investigate issues within Log Analytics. Root cause is not fully understood at this time. Customer using Azure Monitor who may have continue to experience timeouts and failures when running queries and issues with missed or delayed Log Search and Smart Detection alerts hosted in multiple region. We currently have no estimate for resolution.
  • Work Around: None
  • Next Update: Before 06/01 19:00 UTC
-Anmol

Initial Update: Monday, 01 June 2020 14:29 UTC

We are aware of issues within Log Analytics and are actively investigating. Some customers may experience timeouts and failures when running queries and issues with missed or delayed Log Search and Smart Detection alerts hosted in East US region.
  • Work Around: None
  • Next Update: Before 06/01 16:30 UTC
We are working hard to resolve this issue and apologize for any inconvenience.
-Anmol