請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/45995完整後設資料紀錄
| DC 欄位 | 值 | 語言 |
|---|---|---|
| dc.contributor.advisor | 李琳山 | |
| dc.contributor.author | Hung-Yi Li | en |
| dc.contributor.author | 李宏毅 | zh_TW |
| dc.date.accessioned | 2021-06-15T04:50:47Z | - |
| dc.date.available | 2010-08-17 | |
| dc.date.copyright | 2010-08-17 | |
| dc.date.issued | 2010 | |
| dc.date.submitted | 2010-08-02 | |
| dc.identifier.citation | [1] http://www2.sims.berkeley.edu/research/projects/how-much-info/internet.html.
[2] Lan H.witten and Eibe Frank, “Data mining,” 2000. [3] http://www.youtube.com/. [4] http://www.comscore.com/Press_Events/Press_Releases/2010/1/Global_Search_Market_Grows_46_Percent_in_2009. [5] Yi sheng Fu, Chia yu Wan, and Lin shan Lee, “Latent semantic retrieval of personal photos with sparse user annotation by fused image/speech/text features,” in ICASSP, 2009. [6] Text REtrieval Conference, http://trec.nist.gov/. [7] M. Saraclar and R. Sproat, “Lattice-based search for spoken utterance,” in HLT, 2004. [8] J. Mamou, D. Carmel, and R. Hoory, “Spoken document retrieval from call-center conversations,” in SIGIR, 2006. [9] Robertson SE, “The probability ranking principle in IR,” Journal of Documentation, vol. 33(4), pp. 294–304, 1977. [10] Martin Wollmer, Florian Eyben, Joseph Keshet, Alex Graves, Bjorn Schuller, and Gerhard Rigoll, “Robust discriminative keyword spotting for emotionally colored spontaneous speech using bidirectional lstm networks,” in ICASSP, 2009. [11] Chao hong Meng, Hung yi Lee, and Lin shan Lee, “Improved lattice-based spoken document retrieval by directly learning from the evaluation measures,” in ICASSP, 2009. [12] Peng Yu, Duo Zhang, and Frank Seide, “Maximum entropy based normalization ofword posteriors for phonetic and lvcsr lattice search,” in ICASSP, 2006. [13] Dong Wang, Simon King, Joe Frankel, and Peter Bell, “Term-dependent confidence for out-of-vocabulary term detection,” in INTERSPEECH, 2009. [14] Peng Liu, Frank K. Soong, and Jian-Lai Zhou, “Divergence-based similarity measure for spoken document retrieval,” in ICASSP, 2007. [15] Brett Matthews, Upendra Chaudhari, and Bhuvana Ramabhadran, “Fast audio search using vector space modelling,” in ASRU, 2007. [16] Dong Wang, Javier Tejedor, Joe Frankel, Simon King, and Jose Colas, “Posteriorbased confidence measures for spoken term detection,” in ICASSP, 2009. [17] Javier Tejedor, Dong Wang, Simon King, Joe Frankel, and Jose Colas, “A posterior probability-based system hybridisation and combination for spoken term detection,” in INTERSPEECH, 2009. [18] K. Thambiratnam and S. Sridharan, “Dynamic match phone-lattice searches for very fast and accurate unrestricted vocabulary keyword spotting,” in ICASSP, 2005. [19] Timo Mertens and Daniel Schneider, “Efficient subword lattice retrieval for german spoken term detection,” in ICASSP, 2009. [20] Upendra V. Chaudhari and Michael Picheny, “Improvements in phone based audio search via constrained match with high order confusion estimates,” in ASRU, 2007. [21] Yusuke Yokota Tomoyosi Akiba, “Spoken document retrieval by translating recognition candidates into correct transcriptions,” in INTERSPEECH, 2008. [22] Timo Mertens, Daniel Schneider, and Joachim Kohler, “Merging search spaces for subword spoken term detection,” in INTERSPEECH, 2009. [23] Jean-Manuel Van Thong, Pedro J. Moreno, Beth Logan, Blair Fidler, Katrina Maffey, and Matthew Moores, SPEECHBOT: An Experimental Speech-Based Search Engine for Multimedia Content in the Web, 2001. [24] B. Logan, P. Moreno, J. M. Van Thong, and E. Whittacker, “An experimental study of an audio indexing system for the web,” in ICSLP, 2000. [25] Sha Meng, Peng Yu, Frank Seide, and Jia Liu, “A study of lattice-based spoken term detection for chinese spontaneous speech,” in ASRU, 2007. [26] J. Scott Olsson, Jonathan Wintrode, and Matthew Lee, “Fast unconstrained audio search in numerous human languages,” in ICASSP, 2008. [27] Y.-C. Pan, H.-L. Chang, , and L.-S. Lee, “Subword-based position specific posterior lattices (s-pspl) for indexing speech information,” in INTERSPEECH, 2007. [28] Yi cheng Pan, Hung lin Chang, and Lin shan Lee, “Analytical comparison between position specific posterior lattices and confusion networks based onwords and subword units for spoken document indexing,” in ASRU, 2007. [29] Roy Wallace, Robbie Vogt, and Sridha Sridharan, “A phonetic search approach to the 2006 nist spoken term detection evaluation,” in INTERSPEECH, 2007. [30] Ville T. Turunen, “Reducing the effect of oov query words by using morph-based spoken document retrieval,” in INTERSPEECH, 2008. [31] DongWang, Joe Frankel, Javier Tejedor, and Simon King, “A comparison of phone and grapheme-based spoken term detection,” in ICASSP, 2008. [32] Yoshiaki Itoh, Kohei Iwata, Kazunori Kojima, Masaaki Ishigame, Kazuyo Tanaka, and Shi wook Lee, “An integration method of retrieval results using plural subword models for vocabulary-free spoken document retrieval,” in INTERSPEECH, 2007. [33] Alvin Garcia and Herbert Gish, “Keyword spotting of arbitrary words using minimal speech resources,” in ICASSP, 2006. [34] Shi wook Lee, Kazuyo Tanaka, and Yoshiaki Itoh, “Combining multiple subword representations for open-vocabulary spoken document retrieval,” in ICASSP, 2005. [35] Sha Meng, Peng Yu, Jia Liu, and Frank Seide, “Fusing multiple systems into a compact lattice index for chinese spoken term detection,” in ICASSP, 2008. [36] Corentin Dubois and Delphine Charlet, “Using textual information from lvcsr transcripts for phonetic-based spoken term detection,” in ICASSP, 2008. [37] Dogan Can, Erica Cooper, Abhinav Sethy, Chris White, Bhuvana Ramabhadran, and Murat Saraclar, “Effect of pronunciations on oov queries in spoken term detection,” in ICASSP, 2009. [38] Dong Wang, Simon King, and Joe Frankel, “Stochastic pronunciation modeling for spoken term detection,” in INTERSPEECH, 2009. [39] Jian Shao, Roger Peng Yu, Qingwei Zhao, Yonghong Yan, and Frank Seide, “Towards vocabulary-independent speech indexing for large-scale repositories,” in INTERSPEECH, 2008. [40] Ciprian Chelba and Alex Acero, “Position specific posterior lattices for indexing speech,” in ACL, 2005. [41] Jorge Silva, Ciprian Chelba, and Alex Acero, “Pruning analysis for the position specific posterior lattices for spoken document search,” in ICASSP, 2006. [42] Jorge Silva, Ciprian Chelba, and Alex Acero, “Integration of metadata in spoken document search using position specific posterior latices,” in SLT, 2006. [43] Takaaki Hori, I. Lee Hetherington, Timothy J. Hazen, and James R. Glass, “Openvocabulary spoken utterance retrieval using confusion networks,” in ICASSP, 2007. [44] Zheng-Yu Zhou, Peng Yu, Ciprian Chelba, and Frank Seide, “Towards spokendocument retrieval for the internet: lattice indexing for large-scale web-search architectures,” in Proceedings of the main conference on Human Language Technology Conference of the North American Chapter of the Association of Computational Linguistics, Morristown, NJ, USA, 2006, pp. 415–422, Association for Computational Linguistics. [45] Frank Seide, Peng Yu, and Yu Shi, “Towards spoken-document retrieval for the enterprise: Approximateword-lattice indexing with text indexers,” in ASRU, 2007. [46] Roy Wallace, Robbie Vogt, and Sridha Sridharan, “Spoken term detection using fast phonetic decoding,” in ICASSP, 2009. [47] Peng Yu and Frank Seide, “Fast two-stage vocabulary-independent search in spontaneous speech,” in ICASSP, 2005. [48] Wei-Qiang Zhang and Jia Liu, “Two-stage method for specific audio retrieval,” in ICASSP, 2007. [49] Sha Meng, Jian Shao, Roger Peng Yu, Jia Liu, and Frank Seide, “Addressing the out-of-vocabulary problem for large-scale chinese spoken term detection,” in INTERSPEECH, 2008. [50] Carolina Parada, Abhinav Sethy, and Bhuvana Ramabhadran, “Query-by-example spoken term detection for OOV terms,” in ASRU, 2009. [51] Timothy J. Hazen,Wade Shen, and Christopher White, “Query-by-example spoken term detection using phonetic posteriorgram templates,” in ASRU, 2009. [52] Yaodong Zhang and James R. Glass, “Unsupervised spoken keyword spotting via segmental dtw on gaussian posteriorgrams,” in ASRU, 2009. [53] W. Shen, C. White, and T. Hazen, “A comparison of query-by-example methods for spoken term detection,” in INTERSPEECH, 2009. [54] Hui Lin, Alex Stupakov, and Jeff Bilmes, “Improving multi-lattice alignment based spoken keyword spotting,” in ICASSP, 2009. [55] Hui Lin, Alex Stupakov, and Jeff Bilmes, “Spoken keyword spotting via multilattice alignment,” in INTERSPEECH, 2008. [56] J. Fournier, M. Cord, S. Philipp-Foliguet, and F Cergy pontoise Cedex, “Retin: A content-based image indexing and retrieval system,” 2001. [57] Xuehua Shen, Bin Tan, and ChengXiang Zhai, “Context sensitive information retrieval using implicit feedback,” in SIGIR, 2005. [58] Y. Chen, X. Zhou, and T. S. Huang, “One-class svm for learning in image retrieval,” in Proc. IEEE ICIP, 2002. [59] Simon Tong and Daphne Koller, “Support vector machine active learning with applications to text classification,” 2002, vol. 2, pp. 45–66. [60] S. Tong and E. Chang, “Support vector machine active learning for image retrieval,” in Proc. ACM Multimedia, 2001. [61] K.-S. Goh, E. Y. Chang, and W.-C. Lai, “Multimodal concept-dependent active learning for image retrieval,” in Proc. ACM Multimedia, 2004. [62] J. He, M. Li, H.-J. Zhang, H. Tong, and C. Zhang, “Mean version space: a new active learning method for content-based image retrieval,” in In Proc. MIRWorkshop, ACM Multimedia, 2004. [63] Thorsten Joachims and Filip Radlinski, “Search engines that learn from implicit feedback,” Computer, vol. 40, pp. 34–40, 2007. [64] Diane Kelly and Jaime Teevan, “Implicit feedback for inferring user preference: a bibliography,” 2003, vol. 37, pp. 18–28. [65] Georg Buscher, Andreas Dengel, and Ludger van Elst, “Eye movements as implicit relevance feedback,” in CHI ’08: CHI ’08 extended abstracts on Human factors in computing systems, 2008, pp. 2991–2996. [66] Georg Buscher, Andreas Dengel, and Ludger van Elst, “Query expansion using gaze-based feedback on the subdocument level,” in SIGIR ’08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval, 2008, pp. 387–394. [67] Jarkko Salojarvi, Kai Puolamaki, and Samuel Kaski, “Implicit relevance feedback from eye movements,” in In, 2005, pp. 513–518. [68] Ioannis Arapakis, Joemon M. Jose, and Philip D. Gray, “Affective feedback: an investigation into the role of emotions in the information seeking process,” in SIGIR ’08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval, 2008, pp. 395–402. [69] S.E. Robertson, S. Walker, M.M. Beaulieu, M. Gatford, and A. Payne, “Okapi at trec-4,” 1996. [70] Victor Lavrenko and W. Bruce Croft, “Relevance based language models,” in SIGIR ’01: Proceedings of the 24th annual international ACM SIGIR conference on Research and development in information retrieval, 2001, pp. 120–127. [71] D.L. Yeung, C.L.A. Clarke, G.V. Cormack, T.R. Lynam, and E.L. Terra, “Taskspecific query expansion,” in Proc. 12th Text REtrieval Conference (TREC), 2004. [72] Jinxi Xu and W. Bruce Croft, “Query expansion using local and global document analysis,” in SIGIR ’96: Proceedings of the 19th annual international ACM SIGIR conference on Research and development in information retrieval, 1996, pp. 4–11. [73] Tao Tao and ChengXiang Zhai, “Regularized estimation of mixture models for robust pseudo-relevance feedback,” in SIGIR ’06: Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval, 2006, pp. 162–169. [74] Donald Metzler and W. Bruce Croft, “Latent concept expansion using markov random fields,” in SIGIR ’07: Proceedings of the 30th annual international ACM SIGIR conference on Research and development in information retrieval, 2007, pp. 311–318. [75] J.J. Rocchio, “Relevance feedback in information retrieval in the smart retrieval system,” 1971. [76] G. Salton, A. Wong, and C. S. Yang, “A vector space model for automatic indexing,” Commun. ACM, vol. 18, no. 11, pp. 613–620, 1975. [77] J. J. Rocchio, “Relevance feedback in information retrieval,” in The SMART retrieval system - experiments in automatic document processing, 1971. [78] Jay M. Ponte and W. Bruce Croft, “A language modeling approach to information retrieval,” in SIGIR ’98: Proceedings of the 21st annual international ACM SIGIR conference on Research and development in information retrieval, New York, NY, USA, 1998, pp. 275–281, ACM. [79] Fei Song andW. Bruce Croft, “A general language model for information retrieval,” in CIKM ’99: Proceedings of the eighth international conference on Information and knowledge management, New York, NY, USA, 1999, pp. 316–321, ACM. [80] Chengxiang Zhai and John Lafferty, “Model-based feedback in the language modeling approach to information retrieval,” in CIKM ’01: Proceedings of the tenth international conference on Information and knowledge management, New York, NY, USA, 2001, pp. 403–410, ACM. [81] John Lafferty and Chengxiang Zhai, “Document language models, query models, and risk minimization for information retrieval,” in SIGIR ’01: Proceedings of the 24th annual international ACM SIGIR conference on Research and development in information retrieval, New York, NY, USA, 2001, pp. 111–119, ACM. [82] S. E. Robertson and K. Sparck Jones, “Relevance weighting of search terms,” in Journal of the American Society for Information Science, 1976. [83] T Joachims, “Optimizing search engines using clickthrough data,” in ACM SIGKDD, 2002. [84] Olivier Chapelle and Ya Zhang, “A dynamic bayesian network click model for web search ranking,” in WWW, 2009. [85] X. S. Zhou and T. S. Huang, “Relevance feedback in image retrieval: A comprehensive review,” in Multimedia systems, 2003. [86] Ritendra Datta, Dhiraj Joshi, Jia Li, and James Z. Wang, “Image retrieval: Ideas, influences, and trends of the new age,” 2008, vol. 40, pp. 1–60. [87] P. Jourlin, S. E. Johnson, K. S. Jones, and P. C. Woodland, “General query expansion techniques for spoken document retrieval,” in ISCA, 1999. [88] Steve Renals, Dave Abberley, David Kirby, and Tony Robinson, “Indexing and retrieval of broadcast news,” in Speech Communication, 2000. [89] Amit Singhal and Fernando Pereira, “Document expansion for speech retrieval,” in SIGIR ’99: Proceedings of the 22nd annual international ACM SIGIR conference on Research and development in information retrieval. 1999, pp. 34–41, ACM. [90] S. E. Johnson, P. Jourlin, K. Sparck Jones, and P.C. Woodland, “Spoken document retrieval for trec-9 at cambridge university,” in Proc. TREC-7, 1999. [91] Masataka Goto, Jun Ogata, and Kouichirou Eto, “Podcastle: A web 2.0 approach to speech recognition research,” in INTERSPEECH, 2007. [92] Jun Ogata and Masataka Goto, “Podcastle: Collaborative training of acoustic models on the basis of wisdom of crowds for podcast transcription,” in INTERSPEECH, 2009. [93] I. Ruthven and M. Lalmas, “A survey on the use of relevance feedback for information access systems,” in The Knowledge Engineering Review, 2003. [94] Y. K. Chang, C. Cirillo, and J. Razon, “Evaluation of feedback retrieval using modified freezing, residual collection & test and control groups,” in The SMART retrieval system - experiments in automatic document processing, 1971. [95] G. Salton and C. Buckley, “Improving retrieval performance by relevance feedback,” in Journal of the American Society for Information Science, 1990. [96] Roy Wallace, Robbie Vogt, Brendan Baker, and Sridha Sridharan, “Optimising figure of merit for phonetic spoken term detection,” in ICASSP, 2010. [97] Carolina Parada, Abhinav Sethy, and Bhuvana Ramabhadran, “Balancing false alarms and hits in spoken term detection,” in ICASSP, 2010. [98] Joel Pinto, Hynek Hermansky, Igor Szoke, and S.r.m. Prasanna, “Fast approximate spoken term detection from sequence of phonemes,” in SSCS ’08: Proceedings of the third workshop on Searching spontaneous conversational speech, 2008. [99] Laurens van der Werff and Willemijn Heeren, “Evaluating asr output for information retrieval,” in SSCS, 2007. [100] Roy Wallace, Brendan Baker, Robbie Vogt, and Sridha Sridharan, “The effect of language models on phonetic decoding for spoken term detection,” in SSCS ’09: Proceedings of the third workshop on Searching spontaneous conversational speech, 2009, pp. 31–36. [101] Q. Fu and B.-H. Juang, “Automatic speech recognition based on weighted minimum classification error (W-MCE) training method,” in ASRU, 2007. [102] H. Nanjo, T. Misu, and T. Kawahara, “Minimum bayes-risk decoding considering word significance for information retrieval system,” in INTERSPEECH, 2005. [103] H. Nanjo and T. Kawahara, “A new ASR evaluation measure and minimum bayesrisk decoding for open-domain speech understanding,” in ICASSP, 2005. [104] T. Shichiri, H. Nanjo, and T. Yoshimi, “Minimum bayes-risk decoding with presumed word significance for speech based information retrieval,” in ICASSP, 2008. [105] Hung yi Lee and Lin shan Lee, “Improving retrieval performance by user feedback: a new framework for spoken term detection,” in ICASSP, 2010. [106] Cambridge University Engineering Dept. (CUED), Machine Intelligence Laboratory, ”HTK”, http://htk.eng.cam.ac.uk/. [107] SRI Speech Technology and Research Laboratory, ”SRILM”, http://www.speech.sri.com/projects/srilm/. [108] X. Huang, A. Acero, and H.-W. Hon, “Spoken language processing,” Pearson Education Taiwan Ltd., 2005. [109] B.-Y. Liang, “Acoustic models for continuous mandarin speech recognition,” M.S. thesis, NTU, 1998. [110] S. M. Katz, “Estimation of probabilities from sparse data for other language component of a speech recognizer,” in IEEE Trans. Acoustics, Speech and Signal Processing, 1987. [111] D. Povey and P.C. Woodland, “Minimum phone error and i-smoothing for improved discriminative training,” in ICASSP, 2002. [112] D. Povey, “Discriminative training for large vocabulary speech recognition,” 2003. [113] Chia ping Chen, Hung yi Lee, Ching feng Yeh, and Lin shan Lee, “Improved spoken term detection by feature space pseudo-relevance feedback,” in INTERSPEECH (submitted), 2010. [114] G. Aradilla, J. Vepa, and H. Bourlard, “Using posterior-based features in template matching for speech recognition,” in ICSLP, 2006. | |
| dc.identifier.uri | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/45995 | - |
| dc.description.abstract | 本論文提出了以使用者相關回饋提升檢索效能的新架構。過去在語音檢索的領域,有關使用者相關回饋的研究僅限於套用文件檢索領域的技術來修正檢索模型,而本論文提出了以使用者相關回饋來重估測辨識系統的聲學模型參數之新技術。有別於傳統的聲學模型訓練或調適法,本論文以提升檢索效能做為聲學模型訓練的目標,將檢索系統以排序結果進行評估的特性在聲學模型訓練的過程中加以考慮,以及使用沒有標註的資料防止過度適應的情況發生,初步的實驗結果顯示提出的方法可以有效的提升口述語彙偵測系統的效能。
本論文嘗試將長期情境相關回饋、基於範例的虛擬回饋以及基於範例和模型的短期情境相關回饋進行結合,在系統僅從使用者相關回饋得知 5 個口語片段相關性的情況下,凍結排序平均準確率從 0.4819 進步到 0.5433 ,相對進步率為 12.74% ;在系統擁有 60 個訓練查詢詞的相關回饋歷史紀錄之情況下,凍結排序平均準確率從 0.4819 進步到 0.5514 ,相對進步率為 14.42% 。 | zh_TW |
| dc.description.provenance | Made available in DSpace on 2021-06-15T04:50:47Z (GMT). No. of bitstreams: 1 ntu-99-R97942033-1.pdf: 2997757 bytes, checksum: 342b413e55e3fe29ac3bbcfad12306ca (MD5) Previous issue date: 2010 | en |
| dc.description.tableofcontents | 一、緒論. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 研究背景. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.2 本論文研究方向. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2 1.3 本論文研究貢獻. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.4 本論文內容架構. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 二、背景知識介紹. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 2.1 資訊檢索背景知識介紹. . . . . . . . . . . . . . . . . . . . . . . . . . 6 2.2 語音檢索背景知識介紹. . . . . . . . . . . . . . . . . . . . . . . . . . 7 2.3 使用者相關回饋. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 2.3.1 文件資訊檢索上的相關回饋. . . . . . . . . . . . . . . . . . . 14 2.3.2 圖像資訊檢索上的相關回饋. . . . . . . . . . . . . . . . . . . 15 2.3.3 語音資訊檢索上的相關回饋. . . . . . . . . . . . . . . . . . . 16 2.4 資訊檢索評估機制. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 三、本論文提出的相關回饋技術. . . . . . . . . . . . . . . . . . . . . . . . . . 22 3.1 架構. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 3.1.1 短期情境. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 3.1.2 長期情境. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 3.2 提出方法與傳統方法之比較與優勢. . . . . . . . . . . . . . . . . . . 25 3.2.1 和PodCastle系統比較. . . . . . . . . . . . . . . . . . . . . . 25 3.2.2 和辨識結果增強法比較. . . . . . . . . . . . . . . . . . . . . . 27 3.2.3 和傳統聲學模型訓練法比較. . . . . . . . . . . . . . . . . . . 30 四、語音資料庫與實驗環境設定. . . . . . . . . . . . . . . . . . . . . . . . . . 32 4.1 測試語料. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 4.2 聲學模型訓練語料. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 4.3 工具. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 4.4 前端處理. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 4.5 聲學模型訓練. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 4.6 辭典與語言模型. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34 五、以相關回饋重估測聲學模型. . . . . . . . . . . . . . . . . . . . . . . . . . 35 5.1 相關分數. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 5.2 目標函數. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37 5.3 考慮檢索的特性. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39 5.3.1 加入排序. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39 5.3.2 考慮測試資料. . . . . . . . . . . . . . . . . . . . . . . . . . . 41 5.4 實驗結果. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43 5.4.1 基礎實驗. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43 5.4.2 不同的目標函式. . . . . . . . . . . . . . . . . . . . . . . . . . 43 5.4.3 考慮測試資料. . . . . . . . . . . . . . . . . . . . . . . . . . . 46 5.4.4 決定聲學模型訓練的迭代數目. . . . . . . . . . . . . . . . . . 54 5.5 小結. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63 六、結合基於範例的相關回饋. . . . . . . . . . . . . . . . . . . . . . . . . . . 64 6.1 基於範例的相關回饋. . . . . . . . . . . . . . . . . . . . . . . . . . . 64 6.2 結合基於模型和基於範例的相關回饋. . . . . . . . . . . . . . . . . . 66 6.3 實驗結果. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 6.4 小結. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68 七、結合虛擬回饋. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77 7.1 虛擬回饋. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77 7.2 虛擬回饋與相關回饋的結合. . . . . . . . . . . . . . . . . . . . . . . 78 7.3 實驗結果. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79 7.4 小結. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80 八、結合長期情境相關回饋. . . . . . . . . . . . . . . . . . . . . . . . . . . . 83 8.1 長期情境相關回饋. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83 8.2 實驗結果. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84 8.2.1 長期情境相關回饋. . . . . . . . . . . . . . . . . . . . . . . . 84 8.2.2 結合長期情境相關回饋、虛擬回饋與短期情境相關回饋. . . 86 8.3 小結. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87 九、結論與未來展望. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88 9.1 結論. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88 9.2 未來研究方向. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88 9.2.1 和音素混淆度結合. . . . . . . . . . . . . . . . . . . . . . . . 88 9.2.2 隱藏回饋. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 9.2.3 語意分析. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 參考文獻. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92 | |
| dc.language.iso | zh-TW | |
| dc.subject | 相關回饋 | zh_TW |
| dc.subject | Relevance Feedback | en |
| dc.title | 以使用者相關回饋改進語音資訊檢索之新架構 | zh_TW |
| dc.title | A New Framework of Improving Speech Information Retrieval by User Relevance Feedback | en |
| dc.type | Thesis | |
| dc.date.schoolyear | 98-2 | |
| dc.description.degree | 碩士 | |
| dc.contributor.oralexamcommittee | 陳信宏,王小川,鄭秋豫 | |
| dc.subject.keyword | 相關回饋, | zh_TW |
| dc.subject.keyword | Relevance Feedback, | en |
| dc.relation.page | 104 | |
| dc.rights.note | 有償授權 | |
| dc.date.accepted | 2010-08-02 | |
| dc.contributor.author-college | 電機資訊學院 | zh_TW |
| dc.contributor.author-dept | 電信工程學研究所 | zh_TW |
| 顯示於系所單位: | 電信工程學研究所 | |
文件中的檔案:
| 檔案 | 大小 | 格式 | |
|---|---|---|---|
| ntu-99-1.pdf 未授權公開取用 | 2.93 MB | Adobe PDF |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。
