Multimodal Active Speaker Detection And Virtual Cinematography For Video Conferencing
Ross Cutler, Ramin Mehran, Sam Johnson, Oliver Whyte, Adarsh Kowdle, Cha Zhang, Adam Kirk
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 19:34
Active speaker detection (ASD) and virtual cinematography (VC) can significantly improve the experience of a video conference by automatically panning, tilting and zooming of a camera: subjectively users rate an expert video cinematographer significantly higher than the unedited video. We describe a new automated ASD and VC that performs within 0.3 MOS of an expert cinematographer based on subjective ratings with a 1-5 scale. This system uses a 4K wide-FOV camera, a depth camera, and a microphone array, extracts features from each modality and trains an ASD using an AdaBoost machine learning system that is very efficient and runs in real-time. A VC is similarly trained using machine learning. To avoid distracting the room participants the system has no moving parts â the VC works by cropping and zooming the 4K wide-FOV video stream. The system was tuned and evaluated using extensive crowdsourcing techniques and evaluated on a system with N=100 meetings, each 2-5 minutes in length.