Please use this identifier to cite or link to this item:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/51054
Title: | 基於GRU的序列對序列自動編碼器的神經元功能之分析 Exposing the Functionalities of Neurons for Gated Recurrent Unit Based Sequence-to-Sequence Autoencoder |
Authors: | Yi-Ting Lee 李漪莛 |
Advisor: | 林守德(Shou-De Lin) |
Keyword: | GRU,序列對序列模型,自動編碼器,神經元功能, Gated Recurrent Unit,Sequence-to-Sequence Model,Autoencoder,Neurons functionalities, |
Publication Year : | 2020 |
Degree: | 碩士 |
Abstract: | 本文的目的在報告有關Seq2Seq模型的科學發現。眾所周知,由於RNN本質上具有遞歸機制,因此在神經元級別的分析會比分析DNN或CNN模型更具挑戰性。本文旨在提供神經元級的分析,以解釋為什麼基於單純GRU的Seq2Seq模型不需attention的機制即可成功地以很高的正確率、照順序輸出正確的token。我們發現了兩種神經元集合:存儲神經元和倒數神經元,分別存儲token和位置信息,通過分析這兩組神經元在各個時間點如何轉變以及它們的相互作用,我們可以揭開模型如何在正確位置產生正確token的機制。 The goal of this paper is to report certain scientific discoveries about a Seq2Seq model. It is known that analyzing the behavior of RNN-based models at the neuron level is considered a more challenging task than analyzing a DNN or CNN models due to their recursive mechanism in nature. This paper aims to provide neuron-level analysis to explain why a vanilla GRU-based Seq2Seq model without attention can successfully output correct tokens in the correct order with a very high accuracy. We found two types of neurons set, storage neurons and count-down neurons, storing token and position information respectively. By analyzing how these two group of neurons transform through the time step and how they interact, we can uncover the mechanism of how to produce the right tokens in the right positions. |
URI: | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/51054 |
DOI: | 10.6342/NTU202002828 |
Fulltext Rights: | 有償授權 |
Appears in Collections: | 資訊工程學系 |
Files in This Item:
File | Size | Format | |
---|---|---|---|
U0001-1008202016174700.pdf Restricted Access | 5.66 MB | Adobe PDF |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.