%3CLINGO-SUB%20id%3D%22lingo-sub-2140740%22%20slang%3D%22en-US%22%3EExperiencing%20Data%20latency%20Issue%20in%20Azure%20portal%20for%20Log%20Analytics%20EUS2%20-%2002%2F16%20-%20Resolved%3C%2FLINGO-SUB%3E%3CLINGO-BODY%20id%3D%22lingo-body-2140740%22%20slang%3D%22en-US%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EFinal%20Update%3C%2FU%3E%3A%20Wednesday%2C%2017%20February%202021%2000%3A40%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe've%20confirmed%20that%20all%20systems%20are%20back%20to%20normal%20with%20no%20customer%20impact%20as%20of%202%2F17%2C%2000%3A20%20UTC.%20Our%20logs%20show%20the%20incident%20started%20on%202%2F16%2C%2019%3A45%20UTC%20and%20that%20during%20the%20~4%20hours%2025%20min%20hours%20that%20it%20took%20to%20resolve%20the%20issue%2C%20customers%20in%20East%20US2%20experienced%20data%20gaps%20and%20incorrect%20alert%20activation.%3CBR%20%2F%3E%3CUL%3E%3CLI%3E%3CU%3ERoot%20Cause%3C%2FU%3E%3A%20The%20failure%20was%20due%20to%20one%20of%20the%20backend%20services%20becoming%20unhealthy%20due%20to%20incorrect%20configuration%20which%20was%20impacting%20telemetry%20ingestion%20into%20Log%20analytics%20workspaces.%3C%2FLI%3E%3CLI%3E%3CU%3EIncident%20Timeline%3C%2FU%3E%3A%204%20Hours%20%26amp%3B%2025%20minutes%20-%202%2F16%2C%2019%3A45%20UTC%20through%202%2F17%2C%2000%3A20%20UTC%3C%2FLI%3E%3C%2FUL%3EWe%20understand%20that%20customers%20rely%20on%20Azure%20Log%20Analytics%20as%20a%20critical%20service%20and%20apologize%20for%20any%20impact%20this%20incident%20caused.%3CBR%20%2F%3E%3CBR%20%2F%3E-Anupama%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EUpdate%3C%2FU%3E%3A%20Tuesday%2C%2016%20February%202021%2023%3A39%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3ERoot%20cause%20has%20been%20isolated%20to%20one%20of%20the%20backend%20services%20becoming%20unhealthy%20due%20to%20incorrect%20configuration%20which%20was%20impacting%20telemetry%20ingestion%20into%20Log%20analytics%20workspaces.%20To%20address%20this%20issue%20backend%20service%20was%20upgraded%20to%20use%20correct%20configuration.%20Some%20customers%20may%20experience%20data%20gaps%20and%20incorrect%20alert%20activation.%20Currently%20the%20latent%20data%20is%20being%20recovered.%3CBR%20%2F%3E%3CUL%3E%3CLI%3E%3CU%3EWork%20Around%3C%2FU%3E%3A%26nbsp%3B%20None%3C%2FLI%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2002%2F17%2002%3A00%20UTC%3C%2FLI%3E%3C%2FUL%3E-Anupama%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EUpdate%3C%2FU%3E%3A%20Tuesday%2C%2016%20February%202021%2022%3A13%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3E%3CDIV%20style%3D%22font-family%3A%20%22%20segoe%3D%22%22%20ui%3D%22%22%3EWe%20continue%20to%20investigate%20issues%20within%20Log%20Analytics%20in%20East%20US2%20region.%20Root%20cause%20is%20not%20fully%20understood%20at%20this%20time.%20Some%20customers%20ingesting%20telemetry%20in%20their%20Log%20Analytics%20workspaces%20in%20East%20US%202%20may%20experience%20intermittent%20data%20latency%2C%20data%20gaps%20and%20incorrect%20alert%20activation.%20We%20are%20working%20to%20establish%20the%20start%20time%20for%20the%20issue%2C%20initial%20findings%20indicate%20that%20the%20problem%20began%20at%202020-02-16%2019%3A45%20UTC.%20We%20currently%20have%20no%20estimate%20for%20resolution.%3C%2FDIV%3E%3CUL%3E%3CLI%3E%3CU%3EWork%20Around%3C%2FU%3E%3A%20None%3C%2FLI%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2002%2F17%2000%3A30%20UTC%3C%2FLI%3E%3C%2FUL%3E-Anupama%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FLINGO-BODY%3E%3CLINGO-LABS%20id%3D%22lingo-labs-2140740%22%20slang%3D%22en-US%22%3E%3CLINGO-LABEL%3EAzure%20Log%20Analytics%3C%2FLINGO-LABEL%3E%3C%2FLINGO-LABS%3E
Final Update: Wednesday, 17 February 2021 00:40 UTC

We've confirmed that all systems are back to normal with no customer impact as of 2/17, 00:20 UTC. Our logs show the incident started on 2/16, 19:45 UTC and that during the ~4 hours 25 min hours that it took to resolve the issue, customers in East US2 experienced data gaps and incorrect alert activation.
  • Root Cause: The failure was due to one of the backend services becoming unhealthy due to incorrect configuration which was impacting telemetry ingestion into Log analytics workspaces.
  • Incident Timeline: 4 Hours & 25 minutes - 2/16, 19:45 UTC through 2/17, 00:20 UTC
We understand that customers rely on Azure Log Analytics as a critical service and apologize for any impact this incident caused.

-Anupama

Update: Tuesday, 16 February 2021 23:39 UTC

Root cause has been isolated to one of the backend services becoming unhealthy due to incorrect configuration which was impacting telemetry ingestion into Log analytics workspaces. To address this issue backend service was upgraded to use correct configuration. Some customers may experience data gaps and incorrect alert activation. Currently the latent data is being recovered.
  • Work Around:  None
  • Next Update: Before 02/17 02:00 UTC
-Anupama

Update: Tuesday, 16 February 2021 22:13 UTC

We continue to investigate issues within Log Analytics in East US2 region. Root cause is not fully understood at this time. Some customers ingesting telemetry in their Log Analytics workspaces in East US 2 may experience intermittent data latency, data gaps and incorrect alert activation. We are working to establish the start time for the issue, initial findings indicate that the problem began at 2020-02-16 19:45 UTC. We currently have no estimate for resolution.
  • Work Around: None
  • Next Update: Before 02/17 00:30 UTC
-Anupama