TY - GEN
T1 - Automatic Keyword Extraction for Viewport Prediction of 360-degree Virtual Tourism Video
AU - Doan, Long
AU - Duc, Tho Nguyen
AU - Jing, Chuanzhe
AU - Kamioka, Eiji
AU - Tan, Phan Xuan
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - In 360-degree streaming videos, viewport prediction can reduce the bandwidth needed during the stream while still maintaining a quality experience for the users by streaming only the area that is visible to the user. Existing research in viewport prediction aims to predict the user's viewport with data from the user's head movement trajectory, video saliency, and subtitles of the video. While these subtitles can contain much information necessary for viewport prediction, previous studies can only extract these information manually, which requires in-depth knowledge about the topic of the video. Moreover, extracting these information by hand can still miss some important keywords from the subtitles and limit the accuracy of the viewport prediction. In this paper, we focus on automate this extraction process by proposing three types of automatic keyword extraction methods, namely Adverb, NER (Named entity recognition) and Adverb+NER. We provide an analysis to demonstrate the effectiveness of our automatic methods compared to extracting important keywords by hand. We also incorporate our methods into an existing viewport prediction model to improve prediction accuracy. The experimental results show that the model with our automatic keyword extraction methods outperforms baseline methods which only use manually extracted information.
AB - In 360-degree streaming videos, viewport prediction can reduce the bandwidth needed during the stream while still maintaining a quality experience for the users by streaming only the area that is visible to the user. Existing research in viewport prediction aims to predict the user's viewport with data from the user's head movement trajectory, video saliency, and subtitles of the video. While these subtitles can contain much information necessary for viewport prediction, previous studies can only extract these information manually, which requires in-depth knowledge about the topic of the video. Moreover, extracting these information by hand can still miss some important keywords from the subtitles and limit the accuracy of the viewport prediction. In this paper, we focus on automate this extraction process by proposing three types of automatic keyword extraction methods, namely Adverb, NER (Named entity recognition) and Adverb+NER. We provide an analysis to demonstrate the effectiveness of our automatic methods compared to extracting important keywords by hand. We also incorporate our methods into an existing viewport prediction model to improve prediction accuracy. The experimental results show that the model with our automatic keyword extraction methods outperforms baseline methods which only use manually extracted information.
KW - 360-degree video
KW - video subtitles
KW - viewport prediction
KW - virtual reality
KW - virtual tourism videos
UR - http://www.scopus.com/inward/record.url?scp=85148452473&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85148452473&partnerID=8YFLogxK
U2 - 10.1109/ICOCO56118.2022.10032026
DO - 10.1109/ICOCO56118.2022.10032026
M3 - Conference contribution
AN - SCOPUS:85148452473
T3 - 2022 IEEE International Conference on Computing, ICOCO 2022
SP - 386
EP - 391
BT - 2022 IEEE International Conference on Computing, ICOCO 2022
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2022 IEEE International Conference on Computing, ICOCO 2022
Y2 - 14 November 2022 through 16 November 2022
ER -