請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/33454完整後設資料紀錄
| DC 欄位 | 值 | 語言 |
|---|---|---|
| dc.contributor.advisor | 鄭士康(Shyh-Kang Jeng) | |
| dc.contributor.author | Kuo-Chuan Chiu | en |
| dc.contributor.author | 邱國權 | zh_TW |
| dc.date.accessioned | 2021-06-13T04:41:25Z | - |
| dc.date.available | 2006-07-21 | |
| dc.date.copyright | 2006-07-21 | |
| dc.date.issued | 2006 | |
| dc.date.submitted | 2006-07-18 | |
| dc.identifier.citation | [1] T. Beier, S. Neely, ”Feature-based image metamorphosis,” Proceedings of the 19th annual conference on Computer graphics and interactive techniques, pp. 35-42, July 1992.
[2] D. Ruprecht, H. Muller, “Image warping with scattered data interpolation,” IEEE Computer Graphics and Applications, vol. 15, issue 2, pp. 37- 43, March 1995. [3] S. Y. Lee , K. Y. Chwa , S. Y. Shin, “Image metamorphosis using snakes and free-form deformations,” Proceedings of the 22nd annual conference on Computer graphics and interactive techniques, pp. 439-448, September 1995 [4] V Zanella, O. Fuentes, “An Approach to Automatic Morphing of Face Images in Frontal View,” Proceedings of Mexican International Conference on Artificial Intelligence (MICAI), Lecture Notes in Artificial Intelligence 2972, pp. 679-687, 26-30 April 2004. [5] S. Karungaru, M. Fukumi, N. Akamatsu, “Morphing face images using automatically specified features,” Proceedings of the 46th IEEE International Midwest Symposium on vol. 2, pp. 741-744, 27-30 December 2003. [6] T.D. Bui, M. Poel, D. Heylen, A. Nijholt, “Automatic Face Morphing for Transferring Facial Animation,” Proceedings of the 6th IASTED International Conference on Computers, Graphics, and Imaging, pp.19-24, 13-16 August 2003. [7] S. Lee, G. Wolberg, S. Y. Shin, “Polymorph: morphing among multiple images,” IEEE Computer Graphics and Applications, vol.18, issue 1, pp. 58 – 71, January 1998. [8] G. Wolberg, “Image Morphing: A Survey,” The Visual Computer, vol. 14, pp. 360-372, 1998. [9] T. Ezzat, T. Poggio, “Visual speech synthesis by morphing visemes,” International Journal of Computer Vision,, vol. 38, no.1, pp. 45-57, 2000. [10] J. D. Edge, S. Maddock, “Image-based talking heads using radial basis functions,” Proceedings of IEEE Theory and Practice of Computer Graphics, pp. 74–80, June 2003. [11] S. A. King, R. E. Parent, “Lip synchronization for song,” Proceedings of IEEE Computer Animation, pp.233-239, 19-21 June 2002. [12] S. A. King, R. E. Parent, “Animating song,” Journal of Visualization and Computer Animation, vol. 15, no.1, pp. 53-61, March 2004. [13] E. M. Caldognetto, P. Cosi, C. Drioli, G. Tisato, F. Cavicchio, 'Coproduction of speech and emotions: visual and acoustic modifications of some phonetic labial targets,' Proceedings of AVSP, Audio Visual Speech Processing, ISCA Workshop, St Jorioz, France, pp. 209-214, 4-7 September 2003. [14] E. Bevacqua, M. Mancini, C. Pelachaud, “Speaking with Emotions,” Proceedings of the AISB Symposium on Motion, Emotion and Cognition, 2004 [15] I. C. Lin, J. S. Yeh, M. Ouhyoung, 'Realistic 3D Facial Animation Parameters from Mirror-Reflected Multi-View Video,' in IEEE Computer Animation 2001 Conference Proceedings, pp. 2-11, 2001. [16] I. C. Lin, J. S. Yeh, M. Ouhyoung, 'Extracting 3D facial animation parameters from multiview video clips,' IEEE Computer Graphics and Applications, vol. 22, no. 6, pp. 72-80, 2002. [17] I. C. Lin, M. Ouhyoung, 'Mirror MoCap: Automatic and Efficient Capture of Dense 3D Facial Motion Parameters from Video,' The Visual Computer, vol. 12, no. 6, pp. 355-372, 2005 [18] S. Kshirsagar, S. Garchery, G. Sannier, N. M. Thalmann, “Synthetic faces: Analysis and applications,” International Journal of Imaging Systems and Technology, Vol. 13, no. 1, pp. 65–73, 2003. [19] S. Kshirsagar, T. Molet, and N. M. Thalmann, “Principal components of expressive speech animation,” Computer Graphics International, pp. 38-46, July 2001. [20] W. S. Lee, M Escher, G. Sannier; N. Magnenat-Thalmann, ”MPEG-4 compatible faces from orthogonal photos,” Proceedings of IEEE Computer Animation, pp. 186 – 194, 26-29 May 1999. [21] Y. Cao, W. C. Tien, P. Faloutsos, F. H. Pighin, ”Expressive speech-driven facial animation,” ACM Transactions on Graphics, vol. 24, issue 4, pp. 1283-1302, October 2005. [22] Z. Deng, M. Bulut, U. Neumann, S. S. Narayanan, “Automatic dynamic expression synthesis for speech animation,” Proceedings of IEEE 17th International Conference on Computer Animation and Social Agents (CASA), pp.267-274, July 2004. [23] T.D. Bui, D. Heylen, A. Nijholt, “Combination of facial movements on a 3D talking head,”. Proceedings of Computer Graphics International, pp. 284-290, 2004. [24] T.D. Bui, D. Heylen, A. Nijholt, M. Poel, “On combining the facial movements of a talking head,” Proceedings of Measuring Behavior. 5th International Conference on Methods and Techniques in Behavioral Research, pp.19-22, 2005. [25] I. C. Lin, C. F. Huang, J. C. Wu, M. Ouhyoung, 'A Low Bit-rate Web-enabled Synthetic Head with Speech-driven Facial Animation,' Proceedings of Workshop on Computer Animation and Simulation, pp. 29-40, 2000. [26] M.Eck, “Interpolation Methods for Reconstruction of 3D Surfaces from Sequences of Planar Slices,” CAD und Computergraphik, vol. 13, no. 5, pp. 109-120, February 1991. [27] P. Ekman, W. V. Friesen, Unmasking the face: a guide to recognizing emotions facial cues, Prentice-Hall, 1975. [28] http://www.mmk.ei.tum.de/~waf/fgnet/feedtum.html [29] S.A. King, R.E. Parent, and B.L. Olsafsky. “An anatomically-based 3D parametric lip model to support facial animation and synchronized speech,” Proceedings of Deform, pp. 7-19, November 2000. [30] C. Pelachaud, ”Emotion Expressiveness Embedded in Representation Languages for ECAs,” IUT of Montreuil - University of Paris 8, 28 November 2003. [31] M.M. Cohen, D.W. Massaro, “Modeling coarticulation in synthetic visual speech”, In N. M. Thalmann, and D. Thalmann, editors, Model and Technique in Computer Animation, pp. 139-156, Springer-Verlag, Tokyo, 1993. [32] Z. Liu, Y. Shan, Z. Zhang, “Expressive Expression Mapping with Ratio Images,” Computer Graphics, Annual Conference Series, ACM SIGGRAPH, pp. 271-276, August 2001. [33] P. H. Tu, I. C. Lin, J. S. Yeh, R. H. Liang, M. Ouhyoung, 'Surface Detail Capturing for Realistic Facial Animation,' Journal of Computer Science and Technology, vol. 19, no. 4, pp. 618-625, 2004. [34] Q. Zhang, Z. Liu, B. Guo, H. Shum, “Geometry-driven photorealistic facial expression synthesis,” Proceedings of ACM SIGGRAPH/Eurographics Symposium on Computer animation, 26-27 July 2003. [35] M. Brand, “Voice puppetry,” In Computer Graphics, Annual Conference Series, ACM SIGGRAPH, pp. 22-28, August 1999. [36] C. Bregler, M. Covell, M. Slaney, “Video rewrite: Driving visual speech with audio,” In Computer Graphics, ACM SIGGRAPH, pp. 353-360, August 1997. [37] T.Ezzat, G.Geiger, T. Poggio, “Trainable videorealistic speech animation,” In Computer Graphics Annual Conference Series, ACM SIGGRAPH, pp. 388-398, August 2002. | |
| dc.identifier.uri | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/33454 | - |
| dc.description.abstract | 本論文實作完成一套人臉歌唱動畫系統。此系統呈現出人物跟隨著背景音樂來唱歌,並且同時能有臉部表情的變化。我們使用多張臉部表情圖片作為表情資料庫,來合成出臉部的情緒表現。採用的表情圖為Paul Ekman博士所提出的六大基本情緒:快樂﹙happy﹚、憤怒﹙anger﹚、害怕﹙fear﹚、驚訝﹙surprise﹚、哀傷﹙sad﹚、厭惡﹙disgust﹚。歌唱時的發聲嘴形則藉由文字轉換語音系統,Microsoft Speech SDK,來找出相對應於歌詞的嘴形。接著利用文章中所敘述的方法來將表情圖與發聲嘴形做合成,即完成動畫中關鍵影格﹙key-frame﹚的製作。關鍵影格與關鍵影格之間的連續畫面,使用圖形漸形合成﹙image morphing﹚的技術來產生。將這些畫面串連起來,就是栩栩如生的人臉唱歌動畫。此系統除了能降低製造擬真歌唱動畫的成本,在娛樂性的應用方面也相當具有效果。 | zh_TW |
| dc.description.abstract | In this thesis, we implemented a singing-face animation system. In this system, the character sings a song with background music and can change his/her facial expressions simultaneously. We use several expression images to create facial emotional expressions. The images in emotional database are with six basic emotions: happy、anger、fear、surprise、sad and disgust, proposed by Dr. Paul Ekman. The visemes mapping to lyrics are obtained by a text-to-speech system, Microsoft speech SDK. Then, the approach proposed in this article is adopted to synthesize an expression image and a viseme. After that, generating a key-frame of the animation is finished. Using image morphing technique can generate the frames between two key-frames. These clips form a mimic singing-face animation. The proposed system not only lowers the cost of producing a realistic singing-face animation, but also quite powerful for recreational applications. | en |
| dc.description.provenance | Made available in DSpace on 2021-06-13T04:41:25Z (GMT). No. of bitstreams: 1 ntu-95-R93921095-1.pdf: 2049165 bytes, checksum: c400df7625312e54672e3fb33153f107 (MD5) Previous issue date: 2006 | en |
| dc.description.tableofcontents | 摘要 i
Abstract ii 目錄 iii 圖目錄 iv 表目錄 vi 第一章 緒論 1 1.1 引言 1 1.2 研究背景 2 1.3 系統概述 4 1.4 章節概要 5 第二章 圖形漸變合成 6 2.1 圖形漸變合成概述 6 2.2 場域圖形漸變合成法 10 2.3 徑向基底函數 15 2.4 多圖形漸變合成 17 第三章 臉部模型 23 3.1 臉部表情資料庫 23 3.2 發聲嘴形 24 3.3 人臉圖片合成 27 3.4 情緒與發聲嘴形合成 33 第四章 系統實作 44 4.1 系統整體架構 44 4.2 表情圖片產生 44 4.3 表情與嘴形合成 51 4.4 動畫的生成 53 第五章 結論 56 參考文獻 58 | |
| dc.language.iso | zh-TW | |
| dc.subject | 圖形漸變合成 | zh_TW |
| dc.subject | 人臉動畫 | zh_TW |
| dc.subject | 情感合成 | zh_TW |
| dc.subject | Image Morphing | en |
| dc.subject | Facial Animation | en |
| dc.subject | Expressive Synthesis | en |
| dc.title | 基於臉部表情圖片之唱歌人像動畫系統 | zh_TW |
| dc.title | A Singing-Face Animation System Based on Facial Images with Emotional Expressions | en |
| dc.type | Thesis | |
| dc.date.schoolyear | 94-2 | |
| dc.description.degree | 碩士 | |
| dc.contributor.oralexamcommittee | 陳銘憲(Ming-Syan Chen),歐陽明(Ming Ouhyoung) | |
| dc.subject.keyword | 人臉動畫,圖形漸變合成,情感合成, | zh_TW |
| dc.subject.keyword | Facial Animation,Image Morphing,Expressive Synthesis, | en |
| dc.relation.page | 60 | |
| dc.rights.note | 有償授權 | |
| dc.date.accepted | 2006-07-19 | |
| dc.contributor.author-college | 電機資訊學院 | zh_TW |
| dc.contributor.author-dept | 電機工程學研究所 | zh_TW |
| 顯示於系所單位: | 電機工程學系 | |
文件中的檔案:
| 檔案 | 大小 | 格式 | |
|---|---|---|---|
| ntu-95-1.pdf 未授權公開取用 | 2 MB | Adobe PDF |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。
