skip to main content
10.1145/3145690.3145735acmconferencesArticle/Chapter ViewAbstractPublication Pagessiggraph-asiaConference Proceedingsconference-collections
poster

Subtitle positioning for e-learning videos based on rough gaze estimation and saliency detection

Published: 27 November 2017 Publication History

Abstract

Subtitle is very common shown in a variety categories of videos, especially useful as translated subtitles for native speakers. Traditional subtitle is placed at the bottom of videos in order to prevent from occluding essential video contents. However, traversing between important video contents and subtitle frequently will have a negative impact on focusing watching video itself. Recently, some research work try more flexible subtitle positioning strategy. However, these methods are effective with restrictions on the video content and devices adopted. In this work, we propose a novel subtitle content organization and placement framework based on rough gaze estimation and saliency detection.

References

[1]
Hongli Chen, Mengzhen Yan, Sijiang Liu, and Bo Jiang. 2017. Gaze inspired subtitle position evaluation for MOOCs videos. (2017), 1044318-1044318-5 pages.
[2]
M. M. Cheng, J. Warrell, W. Y. Lin, S. Zheng, V. Vineet, and N. Crook. 2013. Efficient Salient Region Detection with Soft Image Abstraction. In 2013 IEEE International Conference on Computer Vision. 1529--1536.
[3]
Face++ 2017. Face Landmarks-Locate keypoints of face components. (2017). Retrieved Aug 15, 2017 from https://www.faceplusplus.com/landmarks/
[4]
Yongtao Hu, Jan Kautz, Yizhou Yu, and Wenping Wang. 2015. Speaker-Following Video Subtitles. ACM Trans. Multimedia Comput. Commun. Appl. 11, 2, Article 32 (Jan. 2015), 17 pages.
[5]
Qiong Huang, Ashok Veeraraghavan, and Ashutosh Sabharwal. 2017. TabletGaze: dataset and analysis for unconstrained appearance-based gaze estimation in mobile tablets. Machine Vision and Applications 28, 5 (01 Aug 2017), 445--461.
[6]
Kuno Kurzhals, Emine Cetinkaya, Yongtao Hu, Wenping Wang, and Daniel Weiskopf. 2017. Close to the Action: Eye-Tracking Evaluation of Speaker-Following Subtitles. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (CHI '17). ACM, New York, NY, USA, 6559--6568.
[7]
Congyi Wang, Fuhao Shi, Shihong Xia, and Jinxiang Chai. 2016. Realtime 3D Eye Gaze Animation Using a Single RGB Camera. ACM Trans. Graph. 35, 4, Article 118 (July 2016), 14 pages.

Cited By

View all
  • (2024)Unspoken Sound: Identifying Trends in Non-Speech Audio Captioning on YouTubeProceedings of the CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642162(1-19)Online publication date: 11-May-2024
  • (2022)Watch It, Don’t Imagine It: Creating a Better Caption-Occlusion Metric by Collecting More Ecologically Valid Judgments from DHH ViewersProceedings of the 2022 CHI Conference on Human Factors in Computing Systems10.1145/3491102.3517681(1-14)Online publication date: 29-Apr-2022
  • (2021)Caption-occlusion severity judgments across live-television genres from deaf and hard-of-hearing viewersProceedings of the 18th International Web for All Conference10.1145/3430263.3452429(1-12)Online publication date: 19-Apr-2021
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
SA '17: SIGGRAPH Asia 2017 Posters
November 2017
114 pages
ISBN:9781450354059
DOI:10.1145/3145690
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 27 November 2017

Check for updates

Author Tags

  1. gaze estimation
  2. saliency detection
  3. subtitle positioning

Qualifiers

  • Poster

Funding Sources

  • the Open Project Program of the State Key Lab of CAD&CG, Zhejiang University
  • NUPTSF

Conference

SA '17
Sponsor:
SA '17: SIGGRAPH Asia 2017
November 27 - 30, 2017
Bangkok, Thailand

Acceptance Rates

Overall Acceptance Rate 178 of 869 submissions, 20%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)13
  • Downloads (Last 6 weeks)1
Reflects downloads up to 08 Mar 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Unspoken Sound: Identifying Trends in Non-Speech Audio Captioning on YouTubeProceedings of the CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642162(1-19)Online publication date: 11-May-2024
  • (2022)Watch It, Don’t Imagine It: Creating a Better Caption-Occlusion Metric by Collecting More Ecologically Valid Judgments from DHH ViewersProceedings of the 2022 CHI Conference on Human Factors in Computing Systems10.1145/3491102.3517681(1-14)Online publication date: 29-Apr-2022
  • (2021)Caption-occlusion severity judgments across live-television genres from deaf and hard-of-hearing viewersProceedings of the 18th International Web for All Conference10.1145/3430263.3452429(1-12)Online publication date: 19-Apr-2021
  • (2021)Preferences of Deaf or Hard of Hearing Users for Live-TV Caption AppearanceUniversal Access in Human-Computer Interaction. Access to Media, Learning and Assistive Environments10.1007/978-3-030-78095-1_15(189-201)Online publication date: 24-Jul-2021

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media