Roy, Sujit and Gaur, Vishal and Raza, Haider and Jameel, Shoaib (2023) CLEFT: Contextualised Unified Learning of User Engagement in Video Lectures with Feedback. IEEE Access, 11. pp. 17707-17720. DOI https://doi.org/10.1109/ACCESS.2023.3245982
Roy, Sujit and Gaur, Vishal and Raza, Haider and Jameel, Shoaib (2023) CLEFT: Contextualised Unified Learning of User Engagement in Video Lectures with Feedback. IEEE Access, 11. pp. 17707-17720. DOI https://doi.org/10.1109/ACCESS.2023.3245982
Roy, Sujit and Gaur, Vishal and Raza, Haider and Jameel, Shoaib (2023) CLEFT: Contextualised Unified Learning of User Engagement in Video Lectures with Feedback. IEEE Access, 11. pp. 17707-17720. DOI https://doi.org/10.1109/ACCESS.2023.3245982
Abstract
Predicting contextualised engagement in videos is a long-standing problem that has been popularly attempted by exploiting the number of views or likes using different computational methods. The recent decade has seen a boom in online learning resources, and during the pandemic, there has been an exponential rise of online teaching videos without much quality control. As a result, we are faced with two key challenges. First, how to decide which lecture videos are engaging to intrigue the listener and increase productivity, and second, how to automatically provide feedback to the content creator using which they could improve the content. The quality of the content could be improved if the creators could automatically get constructive feedback on their content. On the other hand, there has been a steep rise in developing computational methods to predict a user engagement score. In this paper, we have proposed a new unified model, CLEFT, that means “Contextualised unified Learning of user Engagement in video lectures with Feedback” that learns from the features extracted from freely available public online teaching videos and provides feedback on the video along with the user engagement score. Given the complexity of the task, our unified framework employs different pre-trained models working together as an ensemble of classifiers. Our model exploits a range of multi-modal features to model the complexity of language, context agnostic information, textual emotion of the delivered content, animation, speaker’s pitch, and speech emotions. Our results support hypothesis that proposed model can detect engagement reliably and the feedback component gives useful insights to the content creator to further help improve the content.
Item Type: | Article |
---|---|
Uncontrolled Keywords: | NLP; emotions; video engagement; contextual language models; text-based emotions; BERT |
Divisions: | Faculty of Science and Health Faculty of Science and Health > Computer Science and Electronic Engineering, School of |
SWORD Depositor: | Unnamed user with email elements@essex.ac.uk |
Depositing User: | Unnamed user with email elements@essex.ac.uk |
Date Deposited: | 27 Feb 2023 21:53 |
Last Modified: | 16 May 2024 21:42 |
URI: | http://repository.essex.ac.uk/id/eprint/34904 |
Available files
Filename: CLEFT_Contextualised_Unified_Learning_of_User_Engagement_in_Video_Lectures_With_Feedback.pdf
Licence: Creative Commons: Attribution-Noncommercial-No Derivative Works 4.0