%3CLINGO-SUB%20id%3D%22lingo-sub-2137878%22%20slang%3D%22en-US%22%3EExperiencing%20Latency%20and%20Data%20Loss%20issue%20in%20Azure%20Portal%20for%20Many%20Data%20Types%20-%2002%2F16%20-%20Resolved%3C%2FLINGO-SUB%3E%3CLINGO-BODY%20id%3D%22lingo-body-2137878%22%20slang%3D%22en-US%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EFinal%20Update%3C%2FU%3E%3A%20Tuesday%2C%2016%20February%202021%2002%3A26%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe've%20confirmed%20that%20all%20systems%20are%20back%20to%20normal%20with%20no%20customer%20impact%20as%20of%2002%2F16%2C%2002%3A05%20UTC.%20Our%20logs%20show%20the%20incident%20started%20on%2002%2F16%2C%2000%3A15%20UTC%20and%20that%20during%20the%201%20hour%20and%2050%20minutes%20that%20it%20took%20to%20resolve%20the%20issue%20customers%20experienced%20intermittent%20data%20gaps%20of%20up%20to%2010%25%20of%20data%20and%20incorrect%20alert%20activation.%3CBR%20%2F%3E%3CUL%3E%3CLI%3E%3CU%3ERoot%20Cause%3C%2FU%3E%3A%20The%20failure%20was%20due%20to%20a%20specific%20instance%20of%20the%20service%20processing%20backend%20that%20became%20unhealthy.%3C%2FLI%3E%3CLI%3E%3CU%3EIncident%20Timeline%3C%2FU%3E%3A%201%20Hours%20%26amp%3B%2050%20minutes%20-%2002%2F16%2C%2000%3A15%20UTC%20through%2002%2F16%2C%2002%3A05%20UTC%3CBR%20%2F%3E%3C%2FLI%3E%3C%2FUL%3EWe%20understand%20that%20customers%20rely%20on%20Application%20Insights%20as%20a%20critical%20service%20and%20apologize%20for%20any%20impact%20this%20incident%20caused.%3CBR%20%2F%3E%3CBR%20%2F%3E-Jeff%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EUpdate%3C%2FU%3E%3A%20Tuesday%2C%2016%20February%202021%2002%3A12%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3ERoot%20cause%20has%20been%20isolated%20to%20a%20specific%20instance%20of%20the%20service%20processing%20backend%20that%20became%20unhealthy%20which%20was%20impacting%20the%20ingestion%20pipeline.%20To%20address%20this%20issue%20we%20restarted%20the%20affected%20instance%20and%20retrieved%20instance%20data%20for%20analysis.%3CBR%20%2F%3E%3CUL%3E%3CLI%3E%3CU%3EWork%20Around%3C%2FU%3E%3A%20None%3C%2FLI%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2002%2F16%2004%3A30%20UTC%3C%2FLI%3E%3C%2FUL%3E-Jeff%3CBR%20%2F%3E%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CU%3EInitial%20Update%3C%2FU%3E%3A%20Tuesday%2C%2016%20February%202021%2001%3A40%20UTC%3CBR%20%2F%3E%3CBR%20%2F%3EWe%20are%20aware%20of%20issues%20within%20Application%20Insights%20and%20are%20actively%20investigating.%20Some%20customers%20in%20Switzerland%20West%20may%20experience%20intermittent%20data%20gaps%20of%20up%20to%2010%25%20of%20data%20and%20incorrect%20alert%20activation%20starting%20at%202021-02-15%2000%3A15%20UTC.%3C%2FDIV%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CBR%20%2F%3E%3C%2FDIV%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CBR%20%2F%3E%3C%2FDIV%3E%3CDIV%20style%3D%22font-size%3A14px%3B%22%3E%3CUL%3E%3CLI%3E%3CU%3ENext%20Update%3C%2FU%3E%3A%20Before%2002%2F16%2004%3A00%20UTC%3C%2FLI%3E%3C%2FUL%3EWe%20are%20working%20hard%20to%20resolve%20this%20issue%20and%20apologize%20for%20any%20inconvenience.%3CBR%20%2F%3E-Jeff%3C%2FDIV%3E%3CHR%20style%3D%22border-top-color%3Alightgray%22%20%2F%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FDIV%3E%3C%2FLINGO-BODY%3E%3CLINGO-LABS%20id%3D%22lingo-labs-2137878%22%20slang%3D%22en-US%22%3E%3CLINGO-LABEL%3EApplication%20Insights%3C%2FLINGO-LABEL%3E%3C%2FLINGO-LABS%3E
Final Update: Tuesday, 16 February 2021 02:26 UTC

We've confirmed that all systems are back to normal with no customer impact as of 02/16, 02:05 UTC. Our logs show the incident started on 02/16, 00:15 UTC and that during the 1 hour and 50 minutes that it took to resolve the issue customers experienced intermittent data gaps of up to 10% of data and incorrect alert activation.
  • Root Cause: The failure was due to a specific instance of the service processing backend that became unhealthy.
  • Incident Timeline: 1 Hours & 50 minutes - 02/16, 00:15 UTC through 02/16, 02:05 UTC
We understand that customers rely on Application Insights as a critical service and apologize for any impact this incident caused.

-Jeff

Update: Tuesday, 16 February 2021 02:12 UTC

Root cause has been isolated to a specific instance of the service processing backend that became unhealthy which was impacting the ingestion pipeline. To address this issue we restarted the affected instance and retrieved instance data for analysis.
  • Work Around: None
  • Next Update: Before 02/16 04:30 UTC
-Jeff

Initial Update: Tuesday, 16 February 2021 01:40 UTC

We are aware of issues within Application Insights and are actively investigating. Some customers in Switzerland West may experience intermittent data gaps of up to 10% of data and incorrect alert activation starting at 2021-02-15 00:15 UTC.


  • Next Update: Before 02/16 04:00 UTC
We are working hard to resolve this issue and apologize for any inconvenience.
-Jeff