請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/5386
完整後設資料紀錄
DC 欄位 | 值 | 語言 |
---|---|---|
dc.contributor.advisor | 貝蘇章 | |
dc.contributor.author | Wen-Wen Chang | en |
dc.contributor.author | 張雯雯 | zh_TW |
dc.date.accessioned | 2021-05-15T17:57:30Z | - |
dc.date.available | 2016-07-22 | |
dc.date.available | 2021-05-15T17:57:30Z | - |
dc.date.copyright | 2014-07-22 | |
dc.date.issued | 2014 | |
dc.date.submitted | 2014-06-04 | |
dc.identifier.citation | [1] Xiaofeng Ren and J. Malik, 'Learning a Classification Model for Segmentation,' in ICCV, vol. 1, pp. 10-17, 2003.
[2] R. Achanta, A. Shaji, K. Smith, A. Lucchi, P. Fua, and S. Susstrunk, 'Slic superpixels compared to state-of-the-art superpixel methods,' in IEEE T-IP, vol. 34(11), pp. 2274 – 2282, May 2012. [3] W. T. Freeman, E. C. Pasztor, and O. T. Carmichael, 'Learning low-level vision,' in IJCV, vol. 40(1), pp. 25 – 47, October 2000. [4] L. Xu, Q. Yan, Y. Xia, and J. Jia, 'Structure extraction from texture via relative total variation,' in ACM SIGGRAPH Asia, vol. 31(6), no. 139, 2012. [5] C.-T. Shen, W.-L. Hwang, and S.-C. Pei, 'Spatially varying out-of-focus image deblurring with L1-2 optimization and a guided blur map,' in ICASSP, 2012, pp. 1069 – 1072. [6] K. He, J. Sun, and X. Tang, 'Guided image filtering,' in IEEE T-PAMI, 2013, pp. 1397 – 1409. [7] Y. Wei, F. Wen, W. Zhu, and J. Sun, 'Geodesic saliency using background priors, ' in ECCV, 2012, pp. 29 – 42. [8] R. Achanta, S. Hemami, F. Estrada, and S. Susstrunk, 'Frequency-tuned salient region detection,' in CVPR, 2009, pp. 1597 – 1604. [9] V. Movahedi and J. Elder, 'Design and perceptual validation of performance measures for salient object segmentation,' in IEEE Workshop on Perceptual Organization in Computer Vision, 2010. [10] C. Koch and S. Ullman, 'Shifts in selective visual attention: towards the underlying neural circuitry,' in Human neurobiology, vol. 4, pp. 219–227, 1985. [11] L. Itti, C. Koch, and E. Niebur, 'A model of saliency based visual attention for rapid scene analysis,' in IEEE T-PAMI, vol. 20(11), pp. 1254 – 1259, 1998. [12] X. Hou and L. Zhang, 'Saliency detection: A spectral residual approach,' in CVPR, 2007, pp. 1 – 8. [13] S. Goferman, L. Zelnik-Manor, and A. Tal, 'Context aware saliency detection,' in CVPR, 2010, pp. 2376 – 2383. [14] M. Cheng, G. Zhang, N. J. Mitra, X. Huang, and S. Hu, 'Global contrast based salient region detection,' in CVPR, 2011, pp. 409 – 416. [15] R. Achanta and S. Susstrunk, 'Saliency detection using maximum symmetric surround,' in ICIP, 2010, pp. 2653 – 2656. [16] Y. Xie, H. Lu, and M.-H. Yang, 'Bayesian saliency via low and mid level cues,' in IEEE T-IP, pp. 1689 – 1698, 2013. [17] S. Alpert, M. Galun, R. Basri, and A. Brandt, 'Image segmentation by probabilistic bottom-up aggregation and cue integration,' in CVPR, 2007, pp. 1 – 8. [18] Y. Zhai and M. Shah, 'Visual attention detection in video sequences using spatiotemporal cues,' in ACM Multimedia, 2006, pp. 815 – 824. [19] M. Cheng, J. Warrell, W. Lin, S. Zheng, V. Vineet, and N. Crook, 'Efficient salient region detection with soft image abstraction,' in ICCV, 2013. [20] H.-H. Yeh, K.-H. Liu, and C.-S. Chen, 'Salient object detection via local saliency estimation and global homogeneity refinement,' in Pattern Recognition, vol. 47(4), pp. 1740–1750, April 2014, to appear. [21] J. Shuai, L. Qing, J. Miao, Z. Ma, and X. Chen, 'Salient region detection via texture-suppressed background contrast,' in ICIP, 2013. [22] D. Parkhurst, K. Law, and E. Neibur, 'Modeling the role of salience in the allocation of overt visual attention,' in Vision Research, vol. 42, pp. 107–123, 2002. [23] C.-L. Guo, Q. Ma, and L.-M. Zhang, 'Spatio-temporal Saliency detection using phase spectrum of quaternion fourier transform,' in CVPR, pp.1–8, 2008 [24] C.-L. Guo and L.-M. Zhang, 'A Novel Multiresolution Spatiotemporal Saliency Detection Model and Its Applications in Image and Video Compression,' in IEEE T-IP, vol.19, no.1, pp.185–198, 2010 [25] J. Li, M. D. Levine, X.-J. An, X. Xu, and H.-G. He, 'Visual Saliency Based on Scale-Space Analysis in the Frequency Domain,' in IEEE T-PAMI, vol.35, no.4, pp.996–1010, 2013 [26] D. Ruderman. 'The Statistics of Natural Images,' in Network: Computation in Neural Systems, vol.5, no.4, pp. 517–548, 1994. [27] A. Srivastava, A. Lee, E. Simoncelli, and S. Zhu, 'Advances in Statistical Modeling of Natural Images,' in Journal of Mathematical Imaging and Vision, vol.18, no.1, pp. 17–33, 2003. [28] K.-C. Lien and Y.-C. F. Wang, 'Automatic Object Extraction in Single-Concept Videos,' in ICME, 2011. [29] W.-T. Li, H.-T. Chang, H. S. Lyu, and Y.-C. F. Wang, 'Automatic Saliency Inspired Foreground Object Extraction from Videos,' in ICIP, 2012. [30] W.-T. Li, H.-S. Chang, K.-C. Lien, H.-T. Chang, and Y.-C. F. Wang, 'Exploring Visual and Motion Saliency for Automatic Video Object Extraction,' in IEEE T-IP, 2013. [31] C. Liu, 'Beyond Pixels: Exploring New Representations and Applications for Motion Analysis,' Doctoral Thesis, Massachusetts Institute of Technology, 2009. [32] T. Brox, A. Bruhn, N. Papenberg, and J.Weickert, 'High accuracy optical flow estimation based on a theory for warping,' in ECCV, pages 25–36, 2004. [33] A. Bruhn, J.Weickert and C. Schnorr, 'Lucas/Kanade meets Horn/Schunk: combining local and global optical flow methods,' in IJCV, vol.61, no.3, pp. 211–231, 2005. [34] B. K. P. Horn and B. G. Schunck, 'Determing optical flow,' in Artificial Intelligence, vol. 17, pp. 185–203, 1981. [35] M. J. Black and P. Anandan, 'The robust estimation of multiple motions: parametric and piecewise-smooth flow fields,' in Computer Vision and Image Understanding, vol. 63, no.1, pp. 75–104, 1996. [36] B. Zhou, X. Hou, and L. Zhang, 'A Phase Discrepancy Analysis of Object Motion,' in ACCV, vol. 3, pp. 225–238, 2010. [37] X. Cui, Q. Liu , and D. Metaxas, 'Temporal spectral residual: fast motion saliency detection,' in ACM, pp. 617–620, 2009. [38] B. J. Frey and D. MacKay, 'A revolution: belief propagation in graphs with cycles, ' in Proc. NIPS, pp. 479–485, 1998. [39] J. Besag, 'On the statistical analysis of dirty pictures (with discussion),' In Journal of the Royal Statistical Society, Series B, vol. 48, no. 3,pp. 259–302, 1986. [40] J. Pearl, 'Reverend Bayes on inference engines: A distributed hierarchical approach,' in Proceedings of the Second National Conference on Artificial Intelligence, pp. 133–136, 1982. [41] J. Pearl, 'Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference,' Morgan Kaufmann Publishers, San Mateo, California, 1988. [42] Y.-W. Tai and M. Brown, 'Single image defocus map estimation using local constrast prior,' in ICIP, pp. 1797–1800, 2009. [43] S. Zhao and T. Sim, 'Defocus map estimation from a single image,' in Pattern Recognition, vol. 44, no.9 , pp. 1852–1858, 2011. [44] P. F. Felzenszwalb, and D. P. Huttenlocher, 'Efficient belief propagation for early vision,' in CVPR, vol. 1, pp. 261–268, 2004. [45] S. Avidan and A. Shamir, 'Seam carving for content aware image resizing, ' in ACM SIGGRAPH, vol. 26(3), no. 10, 2007. [46] M. Rubinstein, A. Shamir, and S. Avidan, 'Improved seam carving for video retargeting,' in ACM SIGGRAPH, vol. 27(3), no. 16, 2008. [47] H. Hadizadeh and I.V. Bajic, 'Saliency-aware video compression,' in IEEE T-IP, pp. 19 – 33, 2014. [48] C. Siagian and L. Itti, 'Rapid biologically-inspired scene classification using features shared with visual attention,' in IEEE T-PAMI, vol. 29(2), pp. 300 – 312, 2007. [49] N. Jacobson and T.Q. Nguyen, 'Scale-aware saliency for application to frame rate up conversion,' in IEEE T-IP, vol. 21(4), pp. 2198 – 2206, 2012. [50] C. Lee,W. Shaw, H. Liao, S. Yeh, and H.-H. Chen, 'Local dimming of liquid crystal display using visual attention prediction model,' in ICCCN, 2011. [51] Y. Zhai and M. Shah, 'Visual attention detection in video sequences using spatiotemporal cues,' in ACM Multimedia, pp. 815 – 824, 2006. [52] Y. Xie, H. Lu, and M.-H. Yang, 'Bayesian saliency via low and mid level cues,' in IEEE T-IP, pp. 1689 – 1698, 2013. | |
dc.identifier.uri | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/5386 | - |
dc.description.abstract | 顯著性偵測(人類視覺注意力偵測),指的是人類觀測者第一眼會注意到的區域,這些區域和其周圍區域通常有著顯著的差異。顯著性偵測對很多電腦視覺上應用上有幫助,近年來很多研究致力於提升顯著性偵測的準確性。經由觀察發現大部分現有的偵測方法,都難以消去複雜的背景區域並偵測出完整的顯著物體。因此我們運用將影像邊緣區域視作背景的假設及考慮各區域在空間上的關連性以提升準確度。
在本篇論文中,我們提出一個新的影像顯著性偵測方法,並將其延伸至影片顯著性偵測。此方法是基於假設影像邊緣為背景,和一個以超像素為節點單位的馬可夫隨機場模型。首先將影像分割成超像素,再取出每個超像素內的色彩、紋理、聚焦程度的特徵值。然後建立一個馬可夫隨機場模型來描述空間和時間上節點之間的關連性。最後再將以超像素為單位的顯著圖轉換為以像素為單位的顯著圖。實驗結果證實我們的方法可以準確的提取出影像和影片中的顯著性物體,並優於大部分現有的方法。 | zh_TW |
dc.description.abstract | Saliency, also known as visual attention, refers to the areas distinct from its surroundings that human observer would focus at a glance. Saliency detection benefits many computer vision tasks, and extensive efforts have been devoted to achieving better saliency detection performance. We observe that most of the previous works are hard to deal with the non-homogeneous color distribution within an object. Motivated by this observation, we consider the spatial structure between image regions to obtain better results.
In this thesis, a proposed approach for image saliency detection and its extension for video saliency detection are introduced. The approach is based on background prior and superpixel-level Markov Random Field (MRF) model. First, we separate the image into middle-level superpixels and extract low-level features (color, texture energy, and defocus level) within each superpixel. Then, we build up a Markov-Random-Field (MRF) on the superpixels and adopt simplified propagation technique to optimize the superpixel saliency. Afterward, we refine this superpixel-level solution to pixel-level saliency map. Experimental results demonstrate that our proposed method is promising as compared to the state-of-the-art methods in two public available datasets. | en |
dc.description.provenance | Made available in DSpace on 2021-05-15T17:57:30Z (GMT). No. of bitstreams: 1 ntu-103-R01942035-1.pdf: 6145698 bytes, checksum: 8ea7a98e6d1320828becc35707bbb069 (MD5) Previous issue date: 2014 | en |
dc.description.tableofcontents | 口試委員會審定書
誌謝 i 中文摘要 ii ABSTRACT iii CONTENTS iv LIST OF FIGURES vii Chapter 1 Introduction 1 1.1 What Is Saliency 1 1.2 Applications 2 1.3 Organization 3 Chapter 2 Overview of Previous Works on Image Saliency Detection 4 2.1 Biologically Based Approach 4 2.1.1 L. Itti’s Model 4 2.1.2 Simulation and Discussion 6 2.2 Frequency Domain Approaches 8 2.2.1 Spectral Residual (SR) 8 2.2.2 Phase Spectrum of Quaternion Fourier Transform (PQFT) 9 2.2.3 Hypercomplex Fourier Transform (HFT) 12 2.2.4 Simulation and Discussion 13 2.3 Context-Aware Saliency 15 2.3.1 Context-Aware Saliency 15 2.3.2 Simulation and Discussion 18 2.4 Global Contrast Approach 19 2.4.1 Histogram Based Contrast (HC) 19 2.4.2 Region Based Contrast (RC) 21 2.4.3 Simulation Results and Discussion 23 Chapter 3 Proposed Method for Image Saliency 24 3.1 Introduction 25 3.2 Superpixel Segmentation 27 3.3 Background Contrast 29 3.4 Probabilistic Model 32 3.4.1 Markov Random Field Model 32 3.4.2 Data Term Energy and Smoothness Term Energy 34 3.4.3 Optimization Using Belief Propagation 38 3.4.4 Optimization Using a Simpler Method 41 3.4.5 Methodology Evaluation 46 3.5 Refinement using Guided Filter 48 3.6 Performance Evaluation 51 3.6.1 Database 51 3.6.2 Performance Evaluation Methods 51 3.6.3 Quantitative Results 52 3.6.4 Qualitative Results 53 3.7 Applications on Image Retargeting 55 3.7.1 Image Retargeting Algorithm 55 3.7.2 Performance 56 Chapter 4 Overview of Previous Works on Video Saliency Detection 58 4.1 Phase Discrepancy 58 4.1.1 Algorithm 58 4.1.2 Simulation and Discussion 61 4.2 Spatial Temporal Spectral 62 4.2.1 Algorithm 62 4.2.2 Simulation and Discussion 65 4.3 Saliency-Based Video Object Extraction Using CRF Model 66 Chapter 5 Proposed Method for Video Saliency 70 5.1 Introduction 70 5.2 Video Saliency Detection Using Proposed Image Saliency Detection Algorithm 72 5.3 Superpixel Segmentation 74 5.4 Motion Feature: Optical Flow 75 5.5 Background Contrast 77 5.6 Probabilistic Model 82 5.6.1 Spatial Temporal Markov Random Field Model 82 5.6.2 Superpixel Tracking 84 5.6.3 Optimization 85 5.6.4 Methodology Evaluation 86 5.7 Results of Our Proposed Approach 87 Chapter 6 Conclusion and Future Work 90 6.1 Conclusion 90 6.2 Future Work 91 REFERENCE 92 | |
dc.language.iso | en | |
dc.title | 影像與影片之顯著性偵測及一個利用超像素之馬可夫隨機場模型的方法 | zh_TW |
dc.title | Saliency Detection of Image and Video and a Proposed Approach using Superpixel-Level Markov Random Field Model | en |
dc.type | Thesis | |
dc.date.schoolyear | 102-2 | |
dc.description.degree | 碩士 | |
dc.contributor.oralexamcommittee | 吳家麟,祁忠勇,鍾國亮,林康平 | |
dc.subject.keyword | 顯著性偵測,視覺注意力,超像素,馬可夫隨機場, | zh_TW |
dc.subject.keyword | saliency detection,visual attention,superpixel,markov random field, | en |
dc.relation.page | 96 | |
dc.rights.note | 同意授權(全球公開) | |
dc.date.accepted | 2014-06-04 | |
dc.contributor.author-college | 電機資訊學院 | zh_TW |
dc.contributor.author-dept | 電信工程學研究所 | zh_TW |
顯示於系所單位: | 電信工程學研究所 |
文件中的檔案:
檔案 | 大小 | 格式 | |
---|---|---|---|
ntu-103-1.pdf | 6 MB | Adobe PDF | 檢視/開啟 |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。