Skip navigation

DSpace JSPUI

DSpace preserves and enables easy and open access to all types of digital content including text, images, moving images, mpegs and data sets

Learn More
DSpace logo
English
中文
  • Browse
    • Communities
      & Collections
    • Publication Year
    • Author
    • Title
    • Subject
    • Advisor
  • Search TDR
  • Rights Q&A
    • My Page
    • Receive email
      updates
    • Edit Profile
  1. NTU Theses and Dissertations Repository
  2. 電機資訊學院
  3. 資訊網路與多媒體研究所
Please use this identifier to cite or link to this item: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/98519
Title: 多模態大型語言模型之可解釋深度偽造檢測基準
MMIDBench: Multimodal Interpretable Deepfake Detection Benchmark for Multimodal Large Language Models
Authors: 黃康洋
Kang-Yang Huang
Advisor: 鄭文皇
Wen-Huang Cheng
Keyword: 多模態大型語言模型,可解釋性,深度偽造檢測,深度偽造生成,影像編輯,語音合成,擴散模型,生成式人工智慧,
Multimodal Large Language Models,Interpretability,Deepfake Detection,Deepfake Generation,Image Editing,Voice Synthesis,Diffusion Models,Generative Artificial Intelligence,
Publication Year : 2025
Degree: 碩士
Abstract: 生成式人工智慧透過運用多樣化的輸入條件,徹底革新了多媒體內容的創作方式。然而,隨著這些模型日益進步,檢測由人工智慧生成的內容,特別是深度偽造(DeepFake),變得愈發困難。對深偽技術日益增加的關注,使得檢測方法,尤其是多模態大型語言模型在辨識深偽內容方面的效能,成為研究重點。多模態大型語言模型不僅能透過提供決策解釋來提升深偽檢測的透明度,區分真實與合成內容的過程同時也是對其感知與推理能力的嚴格考驗。
為了應對這些挑戰,我們提出了 MMIDBench,一個精心設計、全面評估多模態大型語言模型能力的多模態基準。MMIDBench 涵蓋多種最先進的深偽生成模型,橫跨影像、影片與音訊,包含6種不同的深偽任務。該基準包含10k道題目,涵蓋二元選擇、多選題及開放式問答等多種題型,能夠對多模態大型語言模型進行深入評估。我們利用 MMIDBench 評測了5款閉源多模態大型語言模型,揭示了它們在深偽檢測上的優勢與現階段的侷限。
Generative artificial intelligence has revolutionized how multimedia content is created by utilizing diverse input conditions. However, as these models become more advanced, detecting AI-generated content, particularly DeepFakes, has grown increasingly challenging. Rising concerns over DeepFakes have heightened interest in detection methods, specifically the effectiveness of multimodal large language models (MLLMs) in identifying them. MLLMs not only improve the transparency of DeepFake detection by providing explanations for their decisions, but the process of distinguishing authentic from synthetic content also serves as a robust test of their perceptual and reasoning skills.
To address these challenges, we introduce MMIDBench, a comprehensive multimodal benchmark meticulously crafted to assess the capabilities of MLLMs. MMIDBench features a variety of state-of-the-art DeepFake generative models spanning images, videos, and audio, encompassing 6 distinct DeepFake tasks. The benchmark comprises 10k questions, including binary, multiple-choice, and open-ended formats, enabling an in-depth assessment of MLLMs. We evaluated 5 proprietary MLLMs with MMIDBench, revealing both their strengths and current limitations in DeepFake detection.
URI: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/98519
DOI: 10.6342/NTU202503386
Fulltext Rights: 同意授權(全球公開)
metadata.dc.date.embargo-lift: 2025-08-15
Appears in Collections:資訊網路與多媒體研究所

Files in This Item:
File SizeFormat 
ntu-113-2.pdf12.54 MBAdobe PDFView/Open
Show full item record


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved