您的位置:首页 > Web前端

Video captioning with recurrent networks based on frame- and video-level features and visual content

2017-06-11 10:16 405 查看


Video captioning with recurrent networks based on frame- and video-level features and visual content classification

Rakshith ShettyJorma Laaksonen

(Submitted on 9 Dec 2015)

In this paper, we describe the system for generating textual descriptions of short video clips using recurrent neural networks (RNN), which we used while participating in the Large Scale Movie Description Challenge 2015 in ICCV 2015. Our work builds on static
image captioning systems with RNN based language models and extends this framework to videos utilizing both static image features and video-specific features. In addition, we study the usefulness of visual content classifiers as a source of additional information
for caption generation. With experimental results we show that utilizing keyframe based features, dense trajectory video features and content classifier outputs together gives better performance than any one of them individually.

Subjects:Computer Vision and Pattern Recognition (cs.CV)
Cite as:arXiv:1512.02949 [cs.CV]
 (or arXiv:1512.02949v1 [cs.CV] for this version)

Submission history

From: Rakshith Shetty [view email
[v1] Wed, 9 Dec 2015 17:17:29 GMT (86kb,D)
内容来自用户分享和网络整理,不保证内容的准确性,如有侵权内容,可联系管理员处理 点击这里给我发消息
标签: 
相关文章推荐