Saved video conferences are processed by a few Machine Learning algorithms: #NLP transcribes speech to text, next algorithm performs #FaceRecognition, other identifies speakers activity by lip movement to bind phrases to appropriate speakers. Collected information is structured to metadata related to video and stored in the database.
User self-service portal allows users to sort and filter videos by participants, topics, video name, date, and length. Portal's internal search engine helps users find specific moments inside videos and navigate directly to required video fragments by clicking the item in search results.
The system can evolve by adding new features: gesture recognition for bookmarking or other purposes, specific object recognition or detection, etc. There are many directions for systems growth. For example, we can add the functionality of automatic e-mailing meeting minutes with commitments and meeting conclusions to meeting participants.