Authored by: Alex Hocking and Mason Cusack, Applied Data Scientists and Karol Zak and Shane Peckham, Software Engineers in the ML Platform team in Microsoft's Commercial Software Engineering organization. @shanepeckham
Azure Video Analyzer for Media (AVAM) is an Azure service designed to extract deep insights from video and audio files, including items such as transcripts, identifying faces and people, recognizing brands and image captioning amongst others. The output of Azure Video Analyzer for Media can then be indexed by Azure Cognitive Search (ACS) to provide a rich search experience to allow users to quickly find items of interest within the videos.
The above scenario works well but what if customers want to search on items of interest that are not extracted by Azure Video Analyzer for Media. Or what if they want to search on custom terms that are not found in the extracted insights? This solution will
enable exactly these scenarios by using an Azure Machine Learning (AML) custom machine learning model trained using the new Data Labelling AutoML Vision solution.
Figure 1 – Search results showing snippets of video scenes that match the query "Doberman"
Solution Overview
The following describes the high-level process for this solution:
Figure 2 – High-Level flow of the solution
Next steps
The full code and a step-by-step guide to deploy the results can be found here for building a custom video search experience using Azure Video Indexer, Azure Machine Learning and Azure Cognitive Search. Need to be able to search your video for custom objects using your own custom terminology? Now you have a ready-made solution for that, go ahead and give it a try!
Have questions or feedback? We would love to hear from you! Use our UserVoice page to help us prioritize features or use
Azure Video Analyzer for Media's Stackoverflow page for any questions you have around Azure Video Analyzer for Media.
.
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.