Forum Discussion
OMS Agent on Azure Sentinel Log forwarder not receiving and forwarding logs to sentinel workspace
Hello,
We have observed that we no longer are receiving Syslog and CEF logs from the Azure Sentinel Log forwarder that is deployed on client premise. I have performed the following steps:
netstat -an | grep 514
Status: Listening or established (which is fine)
netstat -an | grep 25226
Status: Listening or established (which is fine)
sudo tcpdump -A -ni any port 514 -vv
Status: receive logs from the data sources (which is fine)
sudo tcpdump -A -ni any port 514 -vv | grep (Zscaler IP)
Status: receive logs from the Zscaler data source, the logs showed Palo Alto name in the CEF messages which means Zscaler traffic was routed through the firewall (which is fine, as confirmed by client)
sudo tcpdump -A -ni any port 25226 -vv
Status: No logs were received (Issue Identified)
sudo tcpdump -A -ni any port 25226 -vv | grep (Zscaler IP)
Status: No logs were received (Issue Identified)
Restarted the Rsyslog Service:
service rsyslog restart (After service restart, Azure Sentinel Started receiving the syslog. The Syslog data source came up and working fine)
Restarted the OMSAgent Service
/opt/microsoft/omsagent/bin/service_control restart {workspace ID}
Status: There was no status message and prompt came, assuming it restarted in the background (Please confirm if this is the normal, not prompting or showing any message)
After OMS Agent restart, ran tcpdump again on OMS Agent to see if it starts receiving the logs but no luck.
I followed the following link: https://docs.microsoft.com/en-us/azure/sentinel/troubleshooting-cef-syslog?tabs=rsyslog
Can any one guide what probably be the cause of this issue????? Any help will be much appreciated. Thanks in advance.
- GaryBusheyBronze ContributorI would also check with your Azure admins to see if they may have added any firewall rules that could have blocked the data. It seems that there is one command from the document to check for this:
sudo wget -O cef_troubleshoot.py https://raw.githubusercontent.com/Azure/Azure-Sentinel/master/DataConnectors/CEF/cef_troubleshoot.py&&sudo python cef_troubleshoot.py [WorkspaceID]- FahadAhmedBrass Contributor
GaryBushey Thank you for the prompt response. We have got the issue resolved through microsoft support. Apparently for some reason the OMI Agent was in a zombie/stuck state. Restarting the agent didnt work, had to manually kill the process and start the agent again.
Killing and starting the agent again resolved the issue. As per MS teams , one of the possibility of this behavior may be that the disk space got full earlier at some point in time, which was then resolved however may be that disk space issue might have caused the agent to go into such a state.
Anyways, the issue has been resolved, we are still monitoring to see if it remains stable or not.
Sharing the above for the benefit of all.
Thanks once again for your support, much appreciated.Fahad.
- alazargCopper ContributorFahadAhmed - I am having the same issue here. how did you prevent it from getting disk full again?