Institutions | About Us | Help | Gaeilge
rian logo

Go Back
Utilising high-level features in summarisation of academic presentations
Curtis, Keith; Jones, Gareth J.F.; Campbell, Nick
We present a novel method for the generation of automatic video summaries of academic presentations. We base our investigation on a corpus of multimodal academic conference presentations combining transcripts with paralinguistic multimodal features. We First generate summaries based on keywords by using transcripts created using automatic speech recognition (ASR). Start and end times for each spoken phrase are identiFied from the ASR transcript, then a value for each phrase created. Spoken phrases are then augmented by incorporating scores for human annotation of paralinguistic features. These features measure audience engagement, comprehension and speaker emphasis. We evaluate the effectiveness of summaries generated for individual presentations, created using speech transcripts and paralinguistic multimodal features, by performing eye-tracking evaluation of participants as they watch summaries and full presentations, and by questionnaire of participants upon completion of eye-tracking studies. Summaries were also evaluated for effectiveness by performing comparisons with an enhanced digital video browser
Keyword(s): Evalueation; Video summarisation; classification; evaluation; eye-tracking
Publication Date:
Type: Other
Peer-Reviewed: Unknown
Language(s): English
Institution: Dublin City University
Citation(s): Curtis, Keith, Jones, Gareth J.F. ORCID: 0000-0003-2923-8365 <> and Campbell, Nick (2017) Utilising high-level features in summarisation of academic presentations. In: ICMR’17 International Conference on Multimedia Retrieval, 6–9 June 2017, Bucharest, Romania. ISBN 978-1-4503-4701-3
Publisher(s): Association for Computing Machinery (ACM)
File Format(s): application/pdf
Related Link(s):,
First Indexed: 2019-06-06 06:05:07 Last Updated: 2019-06-06 06:05:07