Expanding the point-automatic enlargement of presentation video elements

Qiyam Tung, Ranjini Swaminathan, Alon Efrat, Jacobus J Barnard

Research output: Chapter in Book/Report/Conference proceedingConference contribution

3 Citations (Scopus)

Abstract

We present a system that assists users in viewing videos of lectures on small screen devices, such as cell phones. It automatically identifies semantic units on the slides, such as bullets, groups of bullets, and images. As the participant views the lecture, the system magnifies the appropriate semantic unit while it is the focus of the discussion. The system makes this decision based on cues from laser pointer gestures and spoken words that are read off the slide. It then magnifies the semantic element using the slide image and the homography between the slide image and the video frame. Experiments suggest that the semantic units of laser-based events identified by our algorithm closely match those identified by humans. In the case of identifying bullets through spoken words, results are more limited but are a good starting point for more complex methods. Finally, we show that this kind of magnification has potential for improving learning of technical content from video lectures when the resolution of the video is limited, such as when being viewed on hand held devices.

Original languageEnglish (US)
Title of host publicationMM'11 - Proceedings of the 2011 ACM Multimedia Conference and Co-Located Workshops
Pages961-964
Number of pages4
DOIs
StatePublished - 2011
Event19th ACM International Conference on Multimedia ACM Multimedia 2011, MM'11 - Scottsdale, AZ, United States
Duration: Nov 28 2011Dec 1 2011

Other

Other19th ACM International Conference on Multimedia ACM Multimedia 2011, MM'11
CountryUnited States
CityScottsdale, AZ
Period11/28/1112/1/11

Fingerprint

Semantics
Lasers
Experiments

Keywords

  • Laser
  • Learning
  • Lecture
  • Magnification
  • Mobile
  • Speech
  • Video

ASJC Scopus subject areas

  • Computer Graphics and Computer-Aided Design
  • Human-Computer Interaction

Cite this

Tung, Q., Swaminathan, R., Efrat, A., & Barnard, J. J. (2011). Expanding the point-automatic enlargement of presentation video elements. In MM'11 - Proceedings of the 2011 ACM Multimedia Conference and Co-Located Workshops (pp. 961-964) https://doi.org/10.1145/2072298.2071913

Expanding the point-automatic enlargement of presentation video elements. / Tung, Qiyam; Swaminathan, Ranjini; Efrat, Alon; Barnard, Jacobus J.

MM'11 - Proceedings of the 2011 ACM Multimedia Conference and Co-Located Workshops. 2011. p. 961-964.

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Tung, Q, Swaminathan, R, Efrat, A & Barnard, JJ 2011, Expanding the point-automatic enlargement of presentation video elements. in MM'11 - Proceedings of the 2011 ACM Multimedia Conference and Co-Located Workshops. pp. 961-964, 19th ACM International Conference on Multimedia ACM Multimedia 2011, MM'11, Scottsdale, AZ, United States, 11/28/11. https://doi.org/10.1145/2072298.2071913
Tung Q, Swaminathan R, Efrat A, Barnard JJ. Expanding the point-automatic enlargement of presentation video elements. In MM'11 - Proceedings of the 2011 ACM Multimedia Conference and Co-Located Workshops. 2011. p. 961-964 https://doi.org/10.1145/2072298.2071913
Tung, Qiyam ; Swaminathan, Ranjini ; Efrat, Alon ; Barnard, Jacobus J. / Expanding the point-automatic enlargement of presentation video elements. MM'11 - Proceedings of the 2011 ACM Multimedia Conference and Co-Located Workshops. 2011. pp. 961-964
@inproceedings{112fbdb37e784e44b512f7aa60cbf886,
title = "Expanding the point-automatic enlargement of presentation video elements",
abstract = "We present a system that assists users in viewing videos of lectures on small screen devices, such as cell phones. It automatically identifies semantic units on the slides, such as bullets, groups of bullets, and images. As the participant views the lecture, the system magnifies the appropriate semantic unit while it is the focus of the discussion. The system makes this decision based on cues from laser pointer gestures and spoken words that are read off the slide. It then magnifies the semantic element using the slide image and the homography between the slide image and the video frame. Experiments suggest that the semantic units of laser-based events identified by our algorithm closely match those identified by humans. In the case of identifying bullets through spoken words, results are more limited but are a good starting point for more complex methods. Finally, we show that this kind of magnification has potential for improving learning of technical content from video lectures when the resolution of the video is limited, such as when being viewed on hand held devices.",
keywords = "Laser, Learning, Lecture, Magnification, Mobile, Speech, Video",
author = "Qiyam Tung and Ranjini Swaminathan and Alon Efrat and Barnard, {Jacobus J}",
year = "2011",
doi = "10.1145/2072298.2071913",
language = "English (US)",
isbn = "9781450306164",
pages = "961--964",
booktitle = "MM'11 - Proceedings of the 2011 ACM Multimedia Conference and Co-Located Workshops",

}

TY - GEN

T1 - Expanding the point-automatic enlargement of presentation video elements

AU - Tung, Qiyam

AU - Swaminathan, Ranjini

AU - Efrat, Alon

AU - Barnard, Jacobus J

PY - 2011

Y1 - 2011

N2 - We present a system that assists users in viewing videos of lectures on small screen devices, such as cell phones. It automatically identifies semantic units on the slides, such as bullets, groups of bullets, and images. As the participant views the lecture, the system magnifies the appropriate semantic unit while it is the focus of the discussion. The system makes this decision based on cues from laser pointer gestures and spoken words that are read off the slide. It then magnifies the semantic element using the slide image and the homography between the slide image and the video frame. Experiments suggest that the semantic units of laser-based events identified by our algorithm closely match those identified by humans. In the case of identifying bullets through spoken words, results are more limited but are a good starting point for more complex methods. Finally, we show that this kind of magnification has potential for improving learning of technical content from video lectures when the resolution of the video is limited, such as when being viewed on hand held devices.

AB - We present a system that assists users in viewing videos of lectures on small screen devices, such as cell phones. It automatically identifies semantic units on the slides, such as bullets, groups of bullets, and images. As the participant views the lecture, the system magnifies the appropriate semantic unit while it is the focus of the discussion. The system makes this decision based on cues from laser pointer gestures and spoken words that are read off the slide. It then magnifies the semantic element using the slide image and the homography between the slide image and the video frame. Experiments suggest that the semantic units of laser-based events identified by our algorithm closely match those identified by humans. In the case of identifying bullets through spoken words, results are more limited but are a good starting point for more complex methods. Finally, we show that this kind of magnification has potential for improving learning of technical content from video lectures when the resolution of the video is limited, such as when being viewed on hand held devices.

KW - Laser

KW - Learning

KW - Lecture

KW - Magnification

KW - Mobile

KW - Speech

KW - Video

UR - http://www.scopus.com/inward/record.url?scp=84455188418&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=84455188418&partnerID=8YFLogxK

U2 - 10.1145/2072298.2071913

DO - 10.1145/2072298.2071913

M3 - Conference contribution

SN - 9781450306164

SP - 961

EP - 964

BT - MM'11 - Proceedings of the 2011 ACM Multimedia Conference and Co-Located Workshops

ER -