TY - GEN
T1 - Subtitle-based Viewport Prediction for 360-degree Virtual Tourism Video
AU - Jing, Chuanzhe
AU - Duc, Tho Nguyen
AU - Tan, Phan Xuan
AU - Kamioka, Eiji
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - 360-degree streaming videos can provide a rich immersive experiences to the users. However, it requires an extremely high bandwidth network. One of the common solutions for saving bandwidth consumption is to stream only a portion of video covered by the user's viewport. To do that, the user's viewpoint prediction is indispensable. In existing viewport prediction methods, they mainly concentrate on the user's head movement trajectory and video saliency. None of them consider navigation information contained in the video, which can turn the attention of the user to specific regions in the video with high probability. Such information can be included in video subtitles, especially the one in 360-degree virtual tourism videos. This fact reveals the potential contribution of video subtitles to viewport prediction. Therefore, in this paper, a subtitle-based viewport prediction model for 360-degree virtual tourism videos is proposed. This model leverages the navigation information in the video subtitles in addition to head movement trajectory and video saliency, to improve the prediction accuracy. The experimental results demonstrate that the proposed model outperforms baseline methods which only use head movement trajectory and video saliency for viewport prediction.
AB - 360-degree streaming videos can provide a rich immersive experiences to the users. However, it requires an extremely high bandwidth network. One of the common solutions for saving bandwidth consumption is to stream only a portion of video covered by the user's viewport. To do that, the user's viewpoint prediction is indispensable. In existing viewport prediction methods, they mainly concentrate on the user's head movement trajectory and video saliency. None of them consider navigation information contained in the video, which can turn the attention of the user to specific regions in the video with high probability. Such information can be included in video subtitles, especially the one in 360-degree virtual tourism videos. This fact reveals the potential contribution of video subtitles to viewport prediction. Therefore, in this paper, a subtitle-based viewport prediction model for 360-degree virtual tourism videos is proposed. This model leverages the navigation information in the video subtitles in addition to head movement trajectory and video saliency, to improve the prediction accuracy. The experimental results demonstrate that the proposed model outperforms baseline methods which only use head movement trajectory and video saliency for viewport prediction.
KW - 360-degree video
KW - video subtitles
KW - viewport prediction
KW - virtual reality
KW - virtual tourism videos
UR - http://www.scopus.com/inward/record.url?scp=85141027452&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85141027452&partnerID=8YFLogxK
U2 - 10.1109/IISA56318.2022.9904420
DO - 10.1109/IISA56318.2022.9904420
M3 - Conference contribution
AN - SCOPUS:85141027452
T3 - 13th International Conference on Information, Intelligence, Systems and Applications, IISA 2022
BT - 13th International Conference on Information, Intelligence, Systems and Applications, IISA 2022
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 13th International Conference on Information, Intelligence, Systems and Applications, IISA 2022
Y2 - 18 July 2022 through 20 July 2022
ER -