A dataset and evaluation methodology for visual saliency in video | |
Jia, Li ; Yonghong, Tian ; Tiejun, Huang ; Wen, Gao | |
2009 | |
英文摘要 | Recently, visual saliency has drawn great research interest in the field of computer vision and multimedia. Various approaches aiming at calculating visual saliency have been proposed. To evaluate these approaches, several datasets have been presented for visual saliency in images. However, there are few datasets to capture spatiotemporal visual saliency in video. Intuitively, visual saliency in video is strongly affected by temporal context and might vary significantly even in visually similar frames. In this paper, we present an extensive dataset with 7.5-hour videos to capture spatiotemporal visual saliency. The salient regions in frames sequentially sampled from these videos are manually labeled by 23 subjects and then averaged to generate the ground-truth saliency maps. We also present three metrics to evaluate competing approaches. Several typical algorithms were evaluated on the dataset. The experimental results show that this dataset is very suitable for evaluating visual saliency. We also discover some interesting findings that would be addressed in future research. Currently, the dataset is freely available online together with the source code for evaluation. ?2009 IEEE.; EI; 0 |
语种 | 英语 |
DOI标识 | 10.1109/ICME.2009.5202529 |
内容类型 | 其他 |
源URL | [http://ir.pku.edu.cn/handle/20.500.11897/263230] ![]() |
专题 | 信息科学技术学院 |
推荐引用方式 GB/T 7714 | Jia, Li,Yonghong, Tian,Tiejun, Huang,et al. A dataset and evaluation methodology for visual saliency in video. 2009-01-01. |
个性服务 |
查看访问统计 |
相关权益政策 |
暂无数据 |
收藏/分享 |
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。
修改评论