請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/33619
完整後設資料紀錄
DC 欄位 | 值 | 語言 |
---|---|---|
dc.contributor.advisor | 陳炳宇(Bing-Yu Chen) | |
dc.contributor.author | Kai-Yin Cheng | en |
dc.contributor.author | 鄭鎧尹 | zh_TW |
dc.date.accessioned | 2021-06-13T04:50:52Z | - |
dc.date.available | 2006-07-18 | |
dc.date.copyright | 2006-07-18 | |
dc.date.issued | 2006 | |
dc.date.submitted | 2006-07-14 | |
dc.identifier.citation | [1] von Ahn, L., Ruoran Liu, Manuel Blum. Peekaboom: a game for locating objects in images. In ACM Conference on Human Factors in Computing Systems (CHI), 2004, pp. 55–64.
[2] von Ahn, L., and Dabbish, L. Labeling Images with a Computer Game. In ACM Conference on Human Factors in Computing Systems (CHI), 2006, pp. 319–326. [3] Montage-a-google http://grant.robinson.name/projects/montage-a-google/ [4] Guess-a-google http://grant.robinson.name/projects/guess-the-google/ [5] Xing Xie, Hao Liu, Simon Goumaz, Wei-Ying Ma. Learning User Interest for Image Browsing on Small-form-factor Devices. In ACM Conference on Human Factors in Computing Systems (CHI), 2005, pp. 671–680. [6] Temperature Color Visualization http://www.philiplaven.com/p19.html [7] Mikkel, A. Digital Photography, Random House, 1992, pp.9. [8] C. Christopoulos, A. Skodras, and T. Ebrahimi. The JPEG2000 still image coding system: an overview. In IEEE Trans. on Consumer Electronics, Vol. 46, No. 4, 2000, pp. 1103–1127. [9] E.C. Chang, S. Mallat, and C. Yap. Wavlet foveation. Vol. 9, No. 3, Oct. 2000, pp. 312–335. [10] R. Mohan, J.R. Smith, and C.S. Li. Adapting multimedia Internet content for Universal Access. In IEEE Trans. On Multimedia, Vol. 1, No. 1, 1999, pp. 104–114. [11] J.R. Smith, R. Mohan, and C.S. Li. Content-based transcoding of images in the Internet. In Proc. of Int. Conf. on Image Processing 1998, Vol. 3, Chicago, USA, Oct. 1998, pp. 7–11. [12] Lee K, Chang HS, Chun SS, Choi L, Sull S. Perceptionbased image transcoding for universal multimedia access. In: Proceedings of the 8th international conference on image processing (ICIP-2001), Thessaloniki, Greece, October 2001, pp. 2:475–478 [13] ACD System. http://www.acdsystems.com [14] Resco. http://www.resco-net.com [15] Liu, H., Xie, X., Ma, W.Y., and Zhang, H.J. Automatic browsing of large pictures on mobile devices. In ACM Multimedia 2003, Berkeley, CA, USA, Nov. 2003, pp. 148–155 [16] Joula, J.F., Ward, N.J. and MacNamara, T. Visual search and reading of rapid serial presentations of letter strings, words and text. In J. Exper. Psychol.: General, 1982, 111, pp. 208–227. [17] Oscar de Bruijn, Robert Spence: Rapid Serial Visual Presentation: A space-timed trade-off in information presentation. In Advanced Visual Interfaces 2000, pp. 189–192 [18] Paul Viola and Michael J. Jones. Rapid Object Detection using a Boosted Cascade of Simple Features. In IEEE Proc. Computer Vision and Pattern Recognition, Kauai, Hawaii, Dec. 2001, pp. 511–518. [19] Bongwon Suh, Haibin Ling, Benjamin B. Bederson, and David W. Jacobs. Automatic thumbnail cropping and its effectiveness. In Proceedings UIST ’03, , 2003, pp. 95–104. [20] Wang, M.Y., Xie, X., Ma, W.Y., and Zhang, H.J. MobiPicture - browsing pictures on mobile devices. In ACM Multimedia 2003 demo, Berkeley, CA, USA, Nov. 2003, pp. 106–107. [21] Xian-Sheng HUA, Lie LU, and Hong-Jiang ZHANG.. Automatically Converting Photographic Series into Video. 12th ACM International Conference on Multimedia, New York City, USA, Oct. 2004. [22] Microsoft Plus! Digital Media Edition. http://www.microsoft.com [23] F. W. M. Stentiford. An estimator for visual attention through competitive novelty with application to image compression. Picture Coding Symposium, Seoul, April, 2001, pp. 24-27. [24] Wooding, D.S. Fixation maps: quantifying eye movement traces. Eye Tracking Research and Applications Symposium (ETRA 2002), New Orleans, LA, USA, Mar. 2002, pp. 31-36. [25] K Oyekoya, F W M Stentiford. Eye Tracking as a New Interface for Image Retrieval. BT Technology Journal, Vol 22 No 3, July 2004. [26] U. Rutishauser, D. Walther, C. Koch, and P. Perona. Is bottom-up attention useful for object recognition? In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, volume 2, pp. 37–44, Washington, DC, USA, July 2004. [27] U. Rutishauser, D. Walther, C. Koch, and P. Perona. On the usefulness of attention for object recognition. In 2nd International Workshop on Attention and Performance in Computational Vision 2004, pp. 96–103, Prague, Czech Republic, May 2004. [28] D. Walther, U. Rutishauser, C. Koch, and P. Perona. Selective visual attention enables learning and recognition of multiple objects in cluttered scenes. Computer Vision and Image Understanding, pp. 745–770, to be published 2005. [29] A. Bamidele, F. W. Stentiford, and J. Morphett. An attention-based approach to content based image retrieval. British Telecommunications Advanced Research Technology Journal on Intelligent Spaces (Pervasive Computing), 22(3), July 2004. [30] X.-J. Wang, W.-Y. Ma, and X. Li. Data-driven approach for bridging the cognitive gap in image retrieval. In Proceedings of the 2004 IEEE International Conference on Multimedia and Expo, volume 3, pp. 2231–2234, Taibei, Taiwan, June 2004. [31] C. Koch and S. Ullman, Shifts in selective visual attention: Towards the underlying neural circuitry. Hum. Neurobiol. 4, pp.219–297, 916 ~1985!. [32] A. M. Treisman and G. Gelade. A feature-integration theory of attention. Cogn. Psychol. 12, pp.97–136 ~1980!. [33] J. P. Gottlieb, M. Kusunoki, and M. E. Goldberg, The representation of visual salience in monkey parietal cortex. Nature (London) 391, pp.481–4 ~1998!. [34] R. Rarasuraman, The Attentive Brain. MIT Press, Cambridge, MA~1998!. [35] Itti L,Koch C, NieburE. Amodel of saliency-based visual attention for rapid scene analysis. IEEE Trans Pattern Analysis Mach Intell 20(11): pp.1254–1259, 1998. [36] Itti L. and Koch C. A comparison of feature combination strategies for saliency-based visual attention systems. In Proceedings of SPIE Human Vision and Electronic Imaging IV (HVEI’99), volume 3644, pp. 473–482, San Jose, CA, January 1999. [37] A. Treisman, S. Gormican. Feature analysis in early vision: evidence from search asymmetries. Psychology Review 95: pp.15-48, 1988. [38] A. Treisman. Perception of features and objects. In Visual Attention, Oxford University Press, New York, 1998. [39] Ma, Y.F., and Zhang, H.J. Contrast-based image attention analysis by using fuzzy growing. ACM Multimedia 2003, Berkeley, CA, USA, Nov. 200 [40] Y. Deng, C. Kenney, et al., Peer group filtering and perceptual color image quantization, Proc. of IEEE International Symposium on Circuits and Systems, Vol.4, pp.21–24, 1999. [41] L.A. Zadeh. Probability measures of fuzzy events. J. Math. Anal. Appl. 23, pp.421–427, 1968. [42] P.K. Sahoo, D.W. Slaaf, T.A. Albert. Threshold selection using a minimal histogram entropy difference. Optical Engineering 36(7), 1976-1981. [43] Yiqun Hu, Deepu Rajan, Liang-Tien Chia. Robust subspace analysis for detecting visual attention regions in images. ACM Multimedia 2005: pp.716–724 [44] R. Vidal, Y. Ma, and S. Sastry. Generalized principal component analysis (gpca). In Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, volume 1, pp. 621–628, Madison, Wisconsin, USA, June 2003. [45] R. Vidal. Generalized Principal Component Analysis (GPCA): an Algebraic Geometric Approach to Subspace Clustering and Motion Segmentation. PhD thesis, School of Electrical Engineering and Computer Sciences, University of California at Berkeley, August 2003. [46] Chen, L.Q., Xie, X., Fan, X., Ma, W.Y., Zhang, H.J., and Zhou, H.Q. A visual attention model for adapting images on small displays. ACM Multimedia Systems Journal, Vol. 9, No. 4, Oct. 2003. [47] Song Liu, Liang-Tien Chia, Deepu Rajan. Attention region selection with information from professional digital camera. ACM Multimedia 2005: pp. 391–394 [48] S. Kullback and R. A. Leibler. On information and sufficiency. Annals of Mathematical Statistics, 22(1): pp.79–86, March 1951. | |
dc.identifier.uri | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/33619 | - |
dc.description.abstract | 在這篇論文裡面我們利用JAVA設計了一款小遊戲 ─ PhotoShoot。我們利用遊戲的方式,來大量蒐集ROI資料。當使用者玩我們的遊戲的時候,就順便幫我們標記了一張圖片上他們認為比較重要的區域。這些比較重要的區域,我們稱之為region-of-interest (ROI)。許多年來,很多的學者嘗試著發展出許多能自動在圖片中找到ROI的演算法。然而他們所各自發展出來的演算法卻難以互相比較彼此的正確率,這乃是因為缺乏了一個公正的比較基礎。本遊戲裡的照片,因為經過數千位玩家所標記,因此可以視為一個比較客觀而且公正的比較資料庫。有了這些公正的資料後,我們便可以客觀地比較不同演算法之間的效能差異,以及研究各種不同的ROI模型。而這些由玩家標記出來的結果,不但可以幫助電腦專家發展自動找尋ROI的演算法,也可以幫助心理學家及生理學家研究人類究竟是如何觀看照片,以及如何選取照片中有興趣的地方。 | zh_TW |
dc.description.abstract | In this thesis, we have developed a web game ─ PhotoShoot. When people play our game, they also help us to locate important areas in photos. These important areas are often called region-of-interest (ROI). Researchers have studied ROI for many years, and tried to retrieve ROI from an image using automatic methods. Lots of algorithms have been proposed, but it is very hard to compare their performance, since there is no common benchmark for comparison. Because our game has already been played by thousands of players, the results can help us to build a ROI ground truth model for each photo. With this ground truth database, we can easily compare these algorithms’ performance. Moreover, by observing the calculated ROI models, we also also draw some conclusions on the ROI prperties. We hope that our database and visualization results can benefit a lot to researchers of ROI, including computer scientists, psychologists and psychophysicists. | en |
dc.description.provenance | Made available in DSpace on 2021-06-13T04:50:52Z (GMT). No. of bitstreams: 1 ntu-95-R93725053-1.pdf: 7081877 bytes, checksum: 3660e66e834c4f78fe38e7e84974d2fa (MD5) Previous issue date: 2006 | en |
dc.description.tableofcontents | 1. Introduction 1
1.1. What is ROI? 1 1.2. What can ROI do? 2 1.3. Contributions 5 1.4. Orgnization 6 2. Related work 7 2.1. Non-automatic approach 7 2.2. Automatic approach 10 3. Game design 15 3.1. Goal 15 3.2. Inspiration 16 3.3. System architecture 17 3.4. Rules of the game 19 3.4.1. Targeting mode 20 3.4.2. Watching mode 21 3.4.3. Shooting mode 21 3.4.4. Game result 22 3.4.5. Scoring 23 3.4.6. Levels 24 3.4.7. High score ranking 25 3.4.8. Personal profile 26 3.5. Observations & some mechanism 26 3.5.1. Robot 26 3.5.2. Cheating 28 3.5.3. UI issues 29 3.5.4. Players’ feedback 30 4. ROI retrieving method 33 4.1. Build a candidate model 34 4.2. Vote the candidate 37 4.3. Temperature color visualization 41 4.4. ROI model refinement 43 4.5. Visualization web site 45 5. Results 47 5.1. ROI consistency rate 47 5.2. User strategy for photos without a clear ROI 50 5.3. Photo with no consistent ROI and center area 52 5.4. High contrast area 54 5.5. Text 55 5.6. Perspective view point 57 5.7. Photo with high level semantic meaning 59 5.8. Conclusions 61 6. Conclusion & future work 63 7. References 65 | |
dc.language.iso | en | |
dc.title | PhotoShoot: 使用者輔助ROI標示之線上遊戲 | zh_TW |
dc.title | PhotoShoot: A Web-Game for User Assisted ROI Labeling | en |
dc.type | Thesis | |
dc.date.schoolyear | 94-2 | |
dc.description.degree | 碩士 | |
dc.contributor.coadvisor | 莊永裕(Yung-Yu Chuang) | |
dc.contributor.oralexamcommittee | 吳家麟(Ja-Ling Wu),朱浩華(Hao-Hua Chu) | |
dc.subject.keyword | ROI,遊戲,JAVA,人工智慧,人類注意模型,人類感知,電腦視覺,Region-of-Interest,標記,網路, | zh_TW |
dc.subject.keyword | ROI,Game,JAVA,Artificial Intelligence,Attention Model,Human Perception,Computer Vision,Region-of-Interest,Labeing,Web, | en |
dc.relation.page | 69 | |
dc.rights.note | 有償授權 | |
dc.date.accepted | 2006-07-17 | |
dc.contributor.author-college | 管理學院 | zh_TW |
dc.contributor.author-dept | 資訊管理學研究所 | zh_TW |
顯示於系所單位: | 資訊管理學系 |
文件中的檔案:
檔案 | 大小 | 格式 | |
---|---|---|---|
ntu-95-1.pdf 目前未授權公開取用 | 6.92 MB | Adobe PDF |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。