Skip navigation

DSpace

機構典藏 DSpace 系統致力於保存各式數位資料(如:文字、圖片、PDF)並使其易於取用。

點此認識 DSpace
DSpace logo
English
中文
  • 瀏覽論文
    • 校院系所
    • 出版年
    • 作者
    • 標題
    • 關鍵字
    • 指導教授
  • 搜尋 TDR
  • 授權 Q&A
    • 我的頁面
    • 接受 E-mail 通知
    • 編輯個人資料
  1. NTU Theses and Dissertations Repository
  2. 電機資訊學院
  3. 電子工程學研究所
請用此 Handle URI 來引用此文件: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/46846
完整後設資料紀錄
DC 欄位值語言
dc.contributor.advisor簡韶逸(Shao-Yi Chien)
dc.contributor.authorGuan-Lin Wuen
dc.contributor.author吳冠林zh_TW
dc.date.accessioned2021-06-15T05:42:10Z-
dc.date.available2012-08-25
dc.date.copyright2010-08-25
dc.date.issued2010
dc.date.submitted2010-08-20
dc.identifier.citation[1] R. L. Hsu, M. A.-M., and A. K. Jain, “Face detection in color images,”
IEEE Trans. Pattern Anal. Machine Intell., vol. 24, no. 5, pp. 696–706,
May 2002.
[2] Y.-W. Huang, T.-C. Chen, C.-H. Tsai, C.-Y. Chen, T.-W. Chen, C.-S. Chen,
C.-F. Shen, S.-Y. Ma, T.-C. Wang, B.-Y. Hsieh, H.-C. Fang, and L.-G.
Chen, “A 1.3 TOPS H.264/AVC single-chip encoder for HDTV applica-
tions,” in IEEE International Solid-State Circuits Conference Digest of
Technical Papers (ISSCC’05), Feb. 2005.
[3] T.-C. Chen, S.-Y. Chien, Y.-W. Huang, C.-H. Tsai, C.-Y. Chen, T.-W. Chen,
and L.-G. Chen, “Analysis and architecture design of an HDTV720p 30
frames/s H.264/AVC encoder,” IEEE Trans. Circuits Syst. Video Techn.,
vol. 16, pp. 673–688, June 2006.
[4] M. Handley, H. Schulzrinne, E. Schooler, and J. Rosenberg, SIP: Session
Initiation Protocol, IETF RFC 2543, Mar. 1997.
[5] H. Schulzrinne, A. Rao, and R. Lanphier, Real Time Streaming Protocol
(RTSP), IETF RFC 2326, Apr. 1998.
[6] H. Schulzrinne, S. Casner, R. Frederick, and V. Jacobson, RTP: A transport
protocol for real-time applications, IETF RFC 1889, Jan. 1996.[7] Information Technology - Coding of Moving Pictures and Associated Audio
for Digital Storage Media up to about 1.5 Mbit/s - Part2: Video, ISO/IEC
11172-2, 1993.
[8] Information Technology - Generic Coding of Moving Pictures and Associ-
ated Audio Information: Video, ISO/IEC 13818-2 and ITU-T Recommen-
dation H.262, 1996.
[9] Video Codec for Audiovisual Services at p × 64 Kbit/s, ITU-T Recommen-
dation H.261, 1993.
[10] Video Coding for Low Bit Rate Communication, ITU-T Recommendation
H.263, 1998.
[11] Information Technology - Coding of Audio-Visual Objects - Part 2: Visual,
ISO/IEC 14496-2, 1999.
[12] V. K. Goyal, “Theoretical foundations of transform coding,” IEEE Signal
Processing Magazine, vol. 18, no. 5, pp. 9–21, Sept. 2001.
[13] M. Vetterli, “Wavelets, approximation, and compression,” IEEE Signal
Processing Magazine, vol. 18, no. 5, pp. 59–73, Sept. 2001.
[14] Joint Video Team, Draft ITU-T Recommendation and Final Draft Inter-
national Standard of Joint Video Specification, ITU-T Recommendation
H.264 and ISO/IEC 14496-10, 2003.
[15] A. Joch, K. Kossentini, H. Schwarz, T.Wiegand, and G. J. Sulivan, “Perfor-
mance comparison of video coding standards using Lagragian coder con-
trol,” in Proc. IEEE 2002 International Conferencing on Image Processing
(ICIP2002), 2002, pp. 501–504.
[16] A. Luthra, G. J. Sullivan, and T. Wiegand, “Introduction to the special
issue on the H.264/AVC video coding standard,” IEEE Trans. Circuits Syst.
Video Techn., vol. 13, no. 7, pp. 557–559, July 2003.
[17] T. Wiegand, G. J. Sullivan, G. Bjontegaard, and A. Luthra, “Overview of
the H.264/AVC video coding standard,” IEEE Trans. Circuits Syst. Video
Technol., vol. 13, no. 7, pp. 560–576, July 2003.
[18] T. Wedi and H. G. Musmann, “Motion- and aliasing-compensated predic-
tion for hybrid video coding,” IEEE Trans. Circuits Syst. Video Techn., vol.
13, no. 7, pp. 577–586, July 2003.
[19] M. Flierl and B. Girod, “Generalized B pictures and the draft H.264/AVC
video compression standard,” IEEE Trans. Circuits Syst. Video Techn., vol.
13, no. 7, pp. 587–597, July 2003.
[20] P. List, A. Joch, Lainema, G. Bjontegaard, and M. Karczewicz, “Adaptive
deblocking filter,” IEEE Trans. Circuits Syst. Video Techn., vol. 13, no. 7,
pp. 604–613, July 2003.
[21] H. S. Malvar, A. Hallapuro, M. Karczewicz, and L. Kerosfsky, “Low-
complexity transform and quantization in H.264/AVC,” IEEE Trans. Cir-
cuits Syst. Video Techn., vol. 13, no. 7, pp. 598–603, July 2003.
[22] D. Marpe, H. Schwarz, and T. Wiegand, “Contex-based adaptive binary
arithmetic coding in the H.264/AVC video compression standard,” IEEE
Trans. Circuits Syst. Video Techn., vol. 13, no. 7, pp. 620–636, July 2003.
[23] D.Wu, Y. T. Hou, and Y. Q. Zhang, “Transporting real-time video over the
Internet: challenges and approaches,” Proc. IEEE, vol. 88, pp. 1855–1875,
Dec. 2000.
[24] D. Wu, Y. T. Hou, W. Zhu, H. J. Lee, T. Chiang, Y. Q. Zhang, and H. J.
Chao, “On end-to-end architecture for transporting MPEG-4 video over
the Internet,” IEEE Trans. Circuits Syst. Video Techn., vol. 10, no. 6, pp.
923–941, Sept. 2000.
[25] T. Weigand, M. Lightstone, D. Mukherjee, T. G. Campbell, and S. K. Mi-
tra, “Rate-distortion optimized mode selection for very low bit-rate video
coding and the emerging H.263 standard,” IEEE Trans. Circuits Syst. Video
Techn., vol. 6, no. 2, pp. 182–190, Apr. 1996.
[26] H. Sun, W. Kwok, M. Chien, and C. H. J. Ju, “MPEG coding performance
improvement by jointly optimizing coding mode decision and rate control,”
IEEE Trans. Circuits Syst. Video Techn., vol. 7, no. 3, pp. 449–458, June
1997.
[27] S. Wenger, “H.264/AVC over IP,” IEEE Trans. Circuits Syst. Video Tech-
nol., vol. 13, no. 7, pp. 645–656, July 2003.
[28] T. Stockhammer, M. M. Hannuksela, and T. Wiegand, “H.264/AVC in
wireless environments,” IEEE Trans. Circuits Syst. Video Technol., vol. 13,
no. 7, pp. 657–673, July 2003.
[29] W. Tan and A. Zakhor, “Video multicast using layered FEC and scalable
compression,” IEEE Trans. Circuits Syst. Video Techn., vol. 11, no. 3, pp.
373–386, Mar. 2001.
[30] P. Subrahmanya and T. Berger, “Multiple descriptions encoding of images,”
in Proc. IEEE Data Compression Conference, 1997, p. 470.
[31] V. A. Vaishampayan, “Design of multiple description scalar quantizers,”
IEEE Trans. Inform. Theory, vol. 39, no. 3, pp. 821–834, May 1993.
[32] Y. Wang and Q. F. Zhu, “Error control and concealment for video commu-
nication: a review,” Proc. IEEE, vol. 86, no. 5, pp. 974–997, May 1998.
[33] M. Podolsky, M. Vetterli, and S. McCanne, “Limited retransmission of
real-time layered multimedia,” in Proc. IEEE Workshop Multimedia Signal
Processing, Dec. 1998, pp. 591–596.
[34] S. Fukunaga, T. Nakai, and H. Inoue, “Error resilient video coding by
dynamic replacing of reference pictures,” in Proc. IEEE Global Telecom-
munications Conference, Nov. 1996, number 3, pp. 1503–1508.
[35] R. J. Safrenek and J. D. Johnson, “A perceptually tuned sub-band image
coder with image dependent quantization and post-quantization data com-
pression,” in Proc. IEEE International Conference Acoustics, Speech, and
Signal Processing, 1989, pp. 1945–1948.
[36] A. J. Ahumada and H. A. Peterson, “Luminance-model-based DCT quan-
tization for color image compression,” in Proc. SPIE Human Vision, Visual
Processing, and Digital Display III, 1992, vol. 1666, pp. 365–374.
[37] A. B. Watson, “DCT quantization matrices visually optimized for individ-
ual images,” in Proc. SPIE Human Vision, Visual Processing, and Digital
Display IV, 1993, vol. 1913, pp. 202–216.
[38] C.-H. Chou and Y.-C. Li, “A perceptually tuned subband image coder based
on the measure of just-noticeable-distortion profile,” IEEE Trans. Circuits
Syst. Video Techn., vol. 5, no. 6, pp. 467–476, Dec. 1995.
[39] C.-H. Chou and C.-W. Chen, “A perceptually optimized 3-D subband codec
for video communication over wireless channels,” IEEE Trans. Circuits
Syst. Video Techn., vol. 6, no. 2, pp. 143–156, Apr. 1996.
[40] A. B. Watson, G. Y. Yang, J. A. Solomon, and J. Villasenor, “Visibility of
wavelet quantization noise,” IEEE Trans. Image Proc., vol. 6, no. 8, pp.
1164–1175, Aug. 1997.
[41] D. Chai and K. N. Ngan, “Face segmentation using skin-color map in
videophone applications,” IEEE Trans. Circuits Syst. Video Techn., vol. 9,
no. 4, pp. 551–564, June 1999.
[42] S. Lee, M. S. Pattichis, and A. C. Bovik, “Foveated video compression
with optimal rate control,” IEEE Trans. Image Proc., vol. 10, no. 7, pp.
977–992, July 2001.
[43] S. Lee, M. S. Pattichis, and A. C. Bovik, “Foveated video quality assess-
ment,” IEEE Trans. Multimedia, vol. 4, no. 1, pp. 129–132, Mar. 2002.
[44] I. Hontsch and L. J. Karam, “Adaptive image coding with perceptual dis-
tortion control,” IEEE Trans. Image Proc., vol. 11, no. 3, pp. 213–222,
Mar. 2002.
[45] Z. Wang, L. Lu, and L. C. Bovik, “Foveation scalable video coding with
automatic fixation selection,” IEEE Trans. Image Proc., vol. 12, no. 2, pp.
243–254, Feb. 2003.
[46] L. Itti, “Automatic foveation for video compression using a neurobiological
model of visual attention,” IEEE Trans. Image Proc., vol. 13, no. 10, pp.
1304–1318, Oct. 2004.
[47] X.K. Yang, W.S. Ling, Z.K. Lu, E.P. Ong, and S.S. Yao, “Just noticeable
distortion model and its applications in video coding,” Signal Processing:
Image Communication, vol. 20, no. 7, pp. 662–680, August 2005.
[48] X. K. Yang, W. S. Ling, Z. K. Lu, E. P. Ong, and S. S. Yao,
“Motion-compensated residue preprocessing in video coding based on just-
noticeable-distortion profile,” IEEE Trans. Circuits Syst. Video Techn., vol.
15, no. 6, pp. 742–752, June 2005.
[49] X. K. Yang, W. S. Li, Z. K. Lu, X. Lin, S. Rahardja, E. P. Ong, and S. S.
Yao, “Rate control for videophone using local perceptual cues,” IEEE
Trans. Circuits Syst. Video Techn., vol. 15, no. 4, pp. 496–507, Apr. 2005.
[50] C.-W. Tang, C.-H. Chen, Y.-H. Yu, and C.-J. Tsai, “Visual sensitivity
guided bit allocation for video coding,” IEEE Trans. Multimedia, vol. 8,
no. 1, pp. 11–18, Feb. 2006.
[51] C.-W. Tang, “Spatiotemporal visual considerations for video coding,” IEEE
Trans. Multimedia, vol. 9, no. 2, pp. 231–238, Feb. 2007.
[52] J. G. Robson, “Spatial and temporal contrast sensitivity functions of the
visual system,” J. Opt. Soc. Amer., vol. 56, pp. 1141–1142, 1966.
[53] D. H. Kelly, “Motion and vision II. Stabilized spatio-temporal threshold
surface,” J. Opt. Soc. Amer. 69, pp. 1340–1349, 1979.
[54] S. J. Daly, “Engineering observations from spatiovelocity and spatiotem-
poral visual models,” Jan. 1998, vol. 3299, pp. 162–166.
[55] Y. Jia, W. Lin, and A. A. Kassim, “Estimating just-noticeable distortion
for video,” IEEE Trans. Circuits Syst. Video Techn., vol. 16, no. 7, pp.
820–829, July 2006.
[56] Z. Wei and K. N. Ngan, “Spatio-temporal just noticeable distortion profile
for grey scale image/video in DCT domain,” IEEE Trans. Circuits Syst.
Video Techn., vol. 19, no. 3, pp. 337–346, Mar. 2009.
[57] L. Itti, C. Koch, and E. Niebur, “A model of saliency-based visual attention
for rapid scene analysis,” IEEE Trans. Pattern Anal. Machine Intell., vol.
20, no. 11, pp. 1254–1259, Nov. 1998.
[58] H.-C. Nothdurft, “Salience from feature contrast: additivity across dimen-
sions,” Vis. Res., vol. 40, no. 10-12, pp. 1183–1021, June 2000.
[59] S. Treue and J. C. M. Trujillo, “Feature-based attention influences motion
processing gain in macaque visual cortex,” Nature, vol. 399, pp. 575–579,
June 1999.
[60] Z. Wang, L. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image qual-
ity assessment: from error visibility to structural similarity,” IEEE Trans.
Image Proc., vol. 13, no. 4, pp. 600–612, Apr. 2004.
[61] E. Niebur and C. Koch, “Computational architectures for attention,” The
Attentive Brain, pp. 163–186, Cambridge, Mass.: MIT Press, 1998.
[62] A.-C. Tsai, J.-F. Wang, J.-F. Yang, and W.-G. Lin, “Effective subblock-
based pixel-based fast direction detections for H.264 intra prediction,”
IEEE Trans. Circuits Syst. Video Techn., vol. 18, no. 7, pp. 975–982, July
2008.
[63] Y. Liu, Z. G. Li, and Y. C. Soh, “A novel rate control scheme for low delay
video communication of H.264/AVC standard,” IEEE Trans. Circuits Syst.
Video Techn., vol. 17, no. 1, pp. 68–78, Jan. 2007.
[64] W. Lin Z. Lu, X. Yang, E. Ong, and S. Yao, “Modeling visual attention’s
modulatory aftereffects on visual sensitivity and quality evaluation,” IEEE
Trans. Image Proc., vol. 14, no. 11, pp. 1928–1942, Nov. 2005.
[65] J.M. Wolfe and T.S. Horowitz, “What attributes guide the deployment of
visual attention and how do they do it?,” Nature Rev. Neuroscience, vol. 5,
pp. 1–7, 2004.
[66] A. R. Koene and L. Zhaoping, “Feature-specific interactions in salience
from combined feature contrasts: evidence for a bottom-up saliency map
in V1,” Journal of Vision, vol. 7, no. 6, pp. 1–14, July 2007.
[67] A. M. Rohaly, P. Corriveau, J. Libert, A. Webster, V. Baroncini,
J. Beerends, J. L. Blin, L. Contin, T. Hamada, D. Harrison, A. Hekstra,
J. Lubin, Y. Nishida, R. Nishihara, J. Pearson, A. F. Pessoa, N. Pickford,
A. Schertz, M. Visca, A. Watson, and S. Winkler, “Video quality experts
group: current results and future directions,” in Proc. SPIE Visual Commu-
nications and Image Processing, 2000, vol. 4067, pp. 742–753.
[68] Subjective test results for the CfP on scalable video coding technology,
MPEG Meeting Doc. N6383, Mar. 2004.
[69] Subjective test results for the CfP on multi-view video coding, MPEG Meet-
ing Doc. N7779, Jan. 2009.
[70] Methodology for the subjective assessment of the quality of television pic-
tures, ITU-R Recommendation BT.500-11, 2002.
[71] H.264/AVC reference software JM14.0, 2008.
[72] S. Kumar, L. Xu, M. K. Mandal, and S. Panchanathan, “Error resiliency
schemes in H.264/AVC standard,” J. Vis. Commun. Image Represent., vol.
17, no. 2, pp. 425–450, Apr. 2006.
[73] Y.-H. Han and J.-J. Leou, “Detection and correction of transmission errors
in JPEG images,” IEEE Trans. Circuits Syst. Video Technol., vol. 8, no. 2,
pp. 217–231, Apr. 1998.
[74] E. Khan, S. Lehmann, H. Gunji, and M. Ghanbari, “Iterative error detection
and correction of H.263 coded video for wireless networks,” IEEE Trans.
Circuits Syst. Video Technol., vol. 14, no. 12, pp. 1294–1307, Dec. 2004.
[75] S. Gnavi, M. Grangetto, E. Magli, and G. Olmo, “Comparison of rate
allocation strategies for H.264 video transmission over wireless lossy cor-
related networks,” in Proc. IEEE International Conference on Multimedia
and Expo (ICME’03), July 2003.
[76] Y. Chen, K. Yu, J. Li, and S. Li, “An error concealment algorithm for entire
frame loss in video transmission,” in Proc. PCS, 2004.
[77] S. Belfiore, M. Grangetto, E. Magli, and G. Olmo, “Concealment of whole-
frame losses for wireless low bit-rate video based on multiframe optical
flow estimation,” IEEE Trans. Multimedia, vol. 7, no. 2, pp. 316–329, Apr.
2005.
[78] P. Baccichet, D. Bagni, A. Chimienti, L. Pezzoni, and F. Rovati, “Frame
concealment for H.264/AVC decoders,” IEEE Trans. Consumer Electron.,
vol. 51, no. 1, pp. 227–233, Feb. 2005.
[79] Z. Wu and J. M. Boyce, “An error concealment scheme for entire frame
losses based on H.264/JVT/AVC,” in Proc. IEEE International Symposium
on Circuits and Systems (ISCAS’06), May 2006.
[80] H.264/AVC reference software JM10.2, 2005.
[81] D. Agrafiotis, T. K. Chiew, P. Ferre, D. R. Bull, A. R. Nix, A. Doufexi, J. C.
How, and D. Nicholson, “Seamless wireless networking for video surveil-
lance applications,” in Proc. SPIE The International Society for Optical
Engineering, 2005, pp. 39–53.
[82] T. Stockhammer, K. Kontopodis, and T. Wiegand, “Rate-distortion opti-
mization for H.26L video coding in packet loss environment,” in Proc.
Packet Video Workshop, Apr. 2002.
[83] T. Stockhammer and S.Wenger, “Standard-compliant enhancement of JVT
coded video for transmission over fixed and wireless IP,” in Proc. Interna-
tional Tyrrhenian Workshop on Digital Comunications (IWDC 2002), Sept.
2002.
[84] Y. K.Wang, M. M. Hannuksela, V. Varsa, A. Hourunranta, and M. Gabbouj,
“The error concealment feature in the H.26L test model,” in Proc. IEEE
International Conference on Image Processing (ICIP’02), 2002, pp. 729–
736.
[85] S. C. Hsia, S. C. Cheng, and S.W. Chou, “Efficient adaptive error conceal-
ment technique for video decoding system,” IEEE Trans. Multimedia, vol.
7, no. 5, pp. 860–868, Oct. 2005.
[86] W. Kwok and H. Sun, “Multidirectional interpolation for spatial error con-
cealment,” IEEE Trans. Consumer Electron., vol. 39, no. 3, pp. 455–460,
Aug. 1993.
[87] J. W. Suh and Y. S. Ho, “Error concealment based on directional interpo-
lation,” IEEE Trans. Consumer Electron., vol. 3, no. 3, pp. 295–320, Aug.
1997.
[88] S. C. Hsia, “An edge-oriented spatial interpolation for consective block
error concealment,” IEEE Signal Process. Lett., vol. 11, no. 6, pp. 577–
580, June 2004.
[89] Y. Xu and Y. Zhou, “H.264 video communication based refined error con-
cealment schemes,” IEEE Trans. Consumer Electron., vol. 50, no. 4, pp.
1135–1141, Nov. 2004.
[90] D. Agrafiotis, D. R. Bull, and C. N. Canagarajah, “Enhanced error con-
cealment with mode selection,” IEEE Trans. Circuits Syst. Video Technol.,
vol. 16, no. 8, pp. 960–973, Aug. 2006.
[91] S. Aign and K. Fazel, “Temporal and spatial error concealment technique
for hierarchical MPEG-2 video codec,” in Proc. IEEE International Con-
ference on Communications, 1995, pp. 1778–1783.
[92] S. Tsekeridou and I. Pitas, “MPEG-2 error concealment based on block-
matching principles,” IEEE Trans. Circuits Syst. Video Technol., vol. 10,
no. 4, pp. 646–658, June 2000.
[93] J. Zhang, J. F. Arnold, and M. R. Frater, “A cell-loss concealment technique
for MPEG-2 coded video,” IEEE Trans. Circuits Syst. Video Technol., vol.
10, no. 4, pp. 659–665, June 2000.
[94] W. M. Lam and A. R. Reibman, “Recovery of lost or erroneously received
motion vectors,” in Proc. IEEE International Conference on Acoustics,
Speech, and Signal Processing (ICASSP’93), 1993, pp. 417–420.
[95] T. S. Valente, C. Dufour, F. Groliere, and D. Snook, “An efficient error con-
cealment implementation for MPEG4 video streams,” IEEE Trans. Con-
sumer Electron., vol. 47, no. 3, pp. 568–578, Aug. 2001.
[96] J. Y. Pyun, J. S. Lee, J. W. Jeong, J. H. Jeong, and S. J. Ko, “Robust er-
ror concealment for visual communications in burst-packet-loss networks,”
IEEE Trans. Consumer Electron., vol. 49, no. 4, pp. 1013–1019, Nov. 2003.
[97] J. Zheng and L. P. Chau, “Efficient motion vector recovery algorithm for
H.264 based on a polynomial model,” IEEE Trans. Multimedia, vol. 7, no.
3, pp. 507–513, June 2005.
[98] K. Song, T. Chung, Y. Kim, Y. Oh, and C. S. Kim, “Error concealment
of H.264/AVC video frames for mobile video broadcasting,” IEEE Trans.
Consumer Electron., vol. 53, no. 2, pp. 704–711, May 2007.
[99] Y. Xu and Y. Zhou, “Adaptive temporal error concealment scheme for
H.264/AVC video decoder,” IEEE Trans. Consumer Electron., vol. 54, no.
4, pp. 1846–1851, Nov. 2008.
[100] M. J. Chen, L. G. Chen, and R. M. Weng, “Error concealment of lost mo-
tion vectors with overlapped motion compensation,” IEEE Trans. Circuits
Syst. Video Technol., vol. 7, no. 3, pp. 560–563, June 1997.
[101] S. Belfiore, M. Grangetto, E. Magli, and G. Olmo, “Concealment of whole-
frame losses for wireless low bit-rate video based on multiframe optical
flow estimation,” IEEE Trans. Multimedia, vol. 7, no. 2, pp. 316–329, Apr.
2005.
[102] X. Ji, D. Zhao, and W. Gao, “Concealment of whole-picture loss in hierar-
chical B-picture scalable video coding,” IEEE Trans. Multimedia, vol. 11,
no. 1, pp. 11–22, Jan. 2009.
[103] S. C. Hsia and S. W. Chou, “VLSI implementation of high-performance
error concealment processor for TV broadcasting,” IEEE Trans. Circuits
Syst. Video Technol., vol. 17, no. 8, pp. 1054–1064, Aug. 2007.
[104] G. Gennari, G. A. Mian, and L. Celetto, “A H.264 decoder robust to trans-
mission errors,” in Proc. European Signal Processing Conference, vol. 1.
[105] C. C. Liu, J. W. Chen, H. C. Chang, Y. C. Yang, Y. H. Ou Yang, M. C.
Tsai, J. I. Guo, and J. S. Wang, “A 160k gates/4.5kB SRAM H.264 video
decoder for HDTV applications,” IEEE J. Solid-State Circuits, vol. 42, no.
1, pp. 170–182, Jan. 2007.
[106] Z. Liu, Y. Song, M. Shao, S. Li, L. Li, S. Ishiwata, M. Nakagawa, S. Goto,
and T. Ikenaga, “HDTV 1080p H.264/AVC encoder chip design and perfor-
mance analysis,” IEEE J. Solid-State Circuits, vol. 44, no. 2, pp. 594–607,
Feb. 2009.
[107] T.Wiegand, G. J. Sullivan, J. Reichel, H. Schwarz, and M.Wien, Joint Draft
11 of SVC Amendment, Joint Video Team, Doc. JVT-X201, July 2007.
[108] M. Shafique, L. Bauer, and J. Henkel, “Optimizing the H.264/AVC video
encoder application structure for reconfigurable and application-specific
platforms,” Journal of Signal Processing Systems, Nov. 2008.
[109] H. S. Alhichri and M. Kamel, “Image registration using virtual circles and
edge direction,” in Proc. International Conference on Pattern Recognition,
Aug. 2002, pp. 969–972.
[110] J. C.Wang, J. F.Wang, J. F. Yang, and J. T. Chen, “A fast mode decision al-
gorithm and its VLSI design for H.264/AVC intra-prediction,” IEEE Trans.
Circuits Syst. Video Technol., vol. 17, no. 10, pp. 1414–1422, Oct. 2007.
[111] B. La, M. Eom, and Y. Choe, “Dominant edge direction based fast intra
mode decision in the H.264/AVC encoder,” Journal of Zhejiang University
science A, vol. 10, no. 6, pp. 767–777, June 2009.
[112] C. S. Won, D. K. Park, and S. J. Park, “Efficient use of MPEG-7 edge
histogram descriptor,” Electron. Telecommun. Res. Inst. J., vol. 24, no. 1,
pp. 23–30, Feb. 2002.
[113] “MPEG7 visual experimentation model (XM), version 10.0,” SC29/WG11,
Doc. N4063, ISO/ EC/JTC1/, 2002.
[114] S. C. Hsia andW. C. Hsu, “A parallel median filter with pipelined schedul-
ing for real-time 1D and 2D signal processing,” IEICE Trans. Fundamen-
tals, vol. E83-A, no. 7, pp. 1396–1404, July 2000.
[115] T. W. Chen, Y.W. Huang, T. C. Chen, Y. H. Chen, C. Y. Tsai, and
L. G. Chen, “Architecture design of H.264/AVC decoder with hybrid task
pipelining for high definition video,” in Proc. IEEE International Sympo-
sium on Circuits and Systems (ISCAS’05), May 2005, pp. 2931–2934.
dc.identifier.urihttp://tdr.lib.ntu.edu.tw/jspui/handle/123456789/46846-
dc.description.abstract視訊傳輸是一大挑戰,網路頻寬、網路異質性、封包延遲和遺失等問題,讓視訊傳輸系統的設計變的複雜。從視訊編碼系統的觀點來看,編碼效率、抗錯誤能力和具有可調整能力的視訊編碼系統是對於現今的運算與網路環境主要的挑戰。本論文注重在視訊編碼系統的編碼效率和抗錯誤支援能力。為了增加視訊編碼效率,所提出的感知導向的視訊編碼系統考量了人眼感知的特性。另一方面,抗錯誤能力的視訊解碼系統包含了錯誤偵測和錯誤隱蔽能力,可以減輕錯誤傳輸所引起的視訊封包遺失問題。為了能減低視訊封包傳輸的延遲問題,對於所發展的感知模型與錯誤隱蔽演算法都發展了相對應的硬體來達到HDTV及時處理的能力
在所提出的感知視訊編碼系統方面,人類感知特性被考量在所發展的演算法來增加編碼效率。在一個有效率的影像或視訊編碼系統中,除了移除空間性、時間性和統計上的冗餘外,也必須考量人類感知在影像與視訊上的冗餘。所提出的感知模型可以藉著改變在以macroblock為基礎單位的量化參數來幫助視訊編碼系統做更好的位元分配。我們採用結構相似模型(structural similarity model)、視覺注視模型(visual attention model)、可感受差異化失真模型(just-noticeable-difference model)和對比敏感度函數(contrast sensitivity function)並且作適當的結合去得到對每一個macroblock在人類感知上的重要程度參數。對於這些所使用的模型與演算法,我們更進一步的發展改進成適合硬體實作的方式。在外部記憶體頻寬方面,以macroblock為基礎單位的資料重複使用機制被採用來節省達50%的外部記憶體頻寬。此外,每個模型平行處理並共享內部記憶體的硬體架構可以減少硬體面積成本。主觀性實驗結果顯示提出的模型在Qp範圍從24到36可以達到7-41%的位元率節省而沒有視覺上視訊品質上的下降。對所提出的感知模型的硬體實作上,原型晶片利用TSMC 0.18μm技術製成,在100MHz速度時,面積為3.3×3.3 mm2,其功率消耗為83.9 mW,而處理能力則可以達到每秒30張HDTV720P的速度。
對於會造成封包遺失或錯誤的影響上,本論文提出包含錯誤偵測和錯誤隱蔽的抗錯誤視訊解碼系統。提出的錯誤偵測機制考慮了空間性和時間性視訊訊號的特性。此外,適應式閥值決定機制也被發展來讓演算法更適合用於各種不同特性的視訊影片。實驗結果顯示所提的方法可以得到0.5-2.4dB PSNR的改善,並且適用於不同的視訊編碼標準的解碼視訊。
我們也提出了有效率的錯誤隱蔽機制來減輕視訊封包遺失的問題。針對處理因為連續封包遺失而造成的連續畫面遺失的情況,我們提出利用遺失畫面前面跟後面正確解碼的畫面的移動向量場來預測目前要修補的畫面的移動向量場。實驗結果顯示提出的方法比原來只利用遺失畫面的前面畫面的移動向量場來預測目前遺失畫面移動向量的方法,有更好的視訊品質。對於不是連續畫面遺失的情況,本論文提出了利用時間性與空間性的錯誤隱蔽機制,以macroblock為基礎處理單元的機制下,對於空間性錯誤隱蔽演算法,我們考量了在視訊位元流中的內畫面(intra frame)編碼模式的資訊,被使用來作為要選擇雙線性插補或方向性插補的依據。使用這個方法只會有平均0.08dB PSNR的視訊品質下降,但是跟之前傳統的方法比較起來,用一般用途的處理器執行則有40倍處理速度的加速,而且也很適合硬體的實作。對於時間性錯誤隱蔽演算法上,在要修補的區塊的周圍區塊解碼出來的移動向量被拿來當成預測目前區塊移動向量的參考,對於硬體及時處理的支援上,所提出的資料與計算結果重複使用的移動向量估測機制跟傳統作法比較在只有0.18dB PSNR的視訊品質下降,但是可以減少96%的外部記憶體頻寬跟計算量。原型晶片利用UMC 90nm技術製成,在125MHz速度時,其功率消耗為15.77mW,處理能力可以達到每秒30張HDTV1080P畫面的速度。跟以前被提出的錯誤隱蔽硬體比較起來,所提出的錯誤隱蔽硬體,可以達到較高的處理能力和1.81dB PSNR的視訊品質提升。
本論文主要貢獻可分成兩個方向,第一部分為基於人類感知的視訊編碼系統,可以讓編碼效率提升。第二部分是解決因為封包遺失所造成的問題而提出的抗錯誤的視訊解碼系統,其中提出了錯誤偵測與錯誤隱蔽的演算法跟硬體實作。我們由衷希望我們的研究成果可以給人類帶來便利與進步。
zh_TW
dc.description.abstractVideo transmission is a challenging work. Many issues, such as bandwidth, heterogeneity, delay, and loss, make the design of video transmission system complicated. From the viewpoint of source coding layer, coding efficiency, error robust and scalability of video coding system are the main challenges for nowadays computation and network environments. This dissertation focuses on the coding efficiency and error robust support issues for video coding system. To increase compression efficiency in the encoder side, a perception-aware video coding system considering human perception is developed. On the other hand, a robust video decoding system including error detection and error concealment is presented to alleviate the erroneous channel effects. Moreover, hardware architecture design of the proposed error concealment algorithms is also concerned in this dissertation because of the tight timing budget for real-time HDTV video processing.
For the proposed perception-aware video coding system, human perceptual consideration is taken into the traditional video coding system to increase the coding efficiency. In image and video coding field, an effective compression algorithm should remove not only the spatial, temporal and statistical redundancy but also the perceptual
redundancy information from the pictures. The proposed perception model helps to achieve better bit allocation for video coding systems by changing quantization parameters at macroblock level. We adopt and combine the structural similarity model, visual attention models, and just-noticeable-distortion model, and contrast sensitivity function to get the weighting of importance of human eye
perception for each macroblock in video frame via a proper fusion algorithm. The proposed algorithms of the model are further developed and modified to be suitable for hardware implementation. Macroblock-based processing with data reuse scheme is used to save the system bandwidth. Moreover, the architecture of parallel processing for each visual model with sharing the on-chip memory and buffers is developed to reduce the chip area. Subjective experiment results show that the proposed model achieves about 7--41% bit-rate saving in the QP range of 24--36 without visual quality degradation. For the hardware implementation of the
proposed evaluation engine, the chip is taped out using 0.18 um technology. The chip size is about 3.3x3.3 mm^2, and the power consumption is 83.9 mW. The processing capability is HDTV720p.
For the erroneous channel effects, we propose a robust video decoding system which including error detection and error concealment schemes for compressed video transmission. The proposed error detection scheme jointly considers spatial and temporal video characteristics. In addition, adaptive threshold value decision scheme is also exploited to let the proposed algorithm suitable for
different video sequences which have different aracteristics. The simulation results show that with the proposed technique, the image quality improvement of 0.5-2.4dB can be achieved. Furthermore, since the proposed method is applied on the decoded frames, it can be used with any coding standard.
Moveover, this dissertation also presents efficient error
concealment algorithms for video bitstream over error-prone channel suffering from damage. An error concealment algorithm for successive frame losses for H.264/AVC bitstream is developed. It estimates the motion field of a lost frame by forward or backward motion projection from a nearly frame which has correct motion field. Experimental results demonstrate that significant quality improvements can be obtained by the proposed algorithm, both objectively and subjectively. On the other hand, for non-successive
frame losses case, a spatial-temporal error concealment is
presented. For spatial error concealment, a mode selection algorithm considering the reuse of intra mode information embedded in bitstream is developed for the adaptation of bilinear and directional interpolation. It suffers only 0.08 dB video quality drop in average but the speedup measured on a general purpose processor is up to 40 times compared with the conventional methods. It is also more suitable for low cost hardware design. For temporal
error concealment, the decoded motion vectors of the neighboring blocks of the corrupted macroblock are reused to provide hints to estimate the motion vector of the corrupted macroblock. Moreover, hardware architecture design and chip implementation of the proposed error concealment algorithm are also presented. For low cost
hardware implementation, a data and computational results reuse scheme of motion vector estimation is proposed and 96% computation and memory bandwidth can be reduced compared with the conventional methods with 0.18 dB quality drop in average. With UMC 90 nm 1P9M process, the proposed error concealment engine can process HDTV1080P 30 frames-per-second video data and the power consumption is 15.77mW
at 125MHz operation frequency. Compared with the previous hardware design of error concealment engine, the proposed design can achieve higher processing capability and up to 1.81 dB gain in PSNR.
In brief, digital video techniques are contributed in two
directions. Coding efficiency of video coding system can be improved based on the cooperation of the traditional video coding scheme and the proposed perception analysis model and hardware engine. Error robust ability of video decoding system is improved based on the proposed error concealment algorithm and hardware engine. We sincerely hope that our research results could make progress for the convenience of human life.
en
dc.description.provenanceMade available in DSpace on 2021-06-15T05:42:10Z (GMT). No. of bitstreams: 1
ntu-99-D94943009-1.pdf: 3914484 bytes, checksum: 491bf1afef834669f265fbf7d5ab99ae (MD5)
Previous issue date: 2010
en
dc.description.tableofcontentsContents
Abstract xiii
1 Introduction 1
1.1 Video Communication System . . . . . . . . . . . . . . . . . . . 1
1.1.1 General Architecture . . . . . . . . . . . . . . . . . . . . 1
1.1.2 The Trend and Challenges . . . . . . . . . . . . . . . . . 2
1.2 Digital Video and Video Compression . . . . . . . . . . . . . . . 5
1.2.1 Standards . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.2.2 Challenges of Video Coding System for Real-Time Video
Communication . . . . . . . . . . . . . . . . . . . . . . . 7
1.2.3 Existing Approaches of Video Coding System for Real-
Time Video Communication . . . . . . . . . . . . . . . . 9
1.3 Motivation of this Dissertation . . . . . . . . . . . . . . . . . . . 16
1.4 Organization of this Dissertation . . . . . . . . . . . . . . . . . . 18
Part I: Algorithm and Hardware Architecture Design of Perception-Aware
Video Encoding System 19
2 Algorithm and Architecture Design of Perception Engine for Video
Coding Applications 21
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
2.2 Proposed Perceptual Model . . . . . . . . . . . . . . . . . . . . . 23
2.2.1 Overview of the Proposed Algorithm . . . . . . . . . . . 23
2.2.2 Perceptual Model for Encoding Intra Frames . . . . . . . 26
2.2.3 Perceptual Model for Encoding Inter Frames . . . . . . . 27
2.2.4 Fusion of the Perceptual Models . . . . . . . . . . . . . . 39
2.3 Hardware Implementation of the Proposed Perceptual Model . . . 42
2.3.1 Hardware Configuration . . . . . . . . . . . . . . . . . . 43
2.3.2 Color Contrast . . . . . . . . . . . . . . . . . . . . . . . 45
2.3.3 Simplified Skin Color Detection . . . . . . . . . . . . . . 51
2.3.4 Structural Similarity Index . . . . . . . . . . . . . . . . . 51
2.3.5 Just Noticeable Difference (JND) . . . . . . . . . . . . . 55
2.3.6 Chip Design Flow and Specification . . . . . . . . . . . . 55
2.4 Experimental Results of the Proposed Perceptual Model . . . . . . 59
2.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
Part II: Algorithm and Hardware Architecture Design of Error Robust
Video Decoding System 74
3 Spatial-Temporal Error Detection Scheme for Video Transmission over
Noisy Channels 75
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
3.2 Proposed Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . 77
3.2.1 Proposed Error Detection Algorithm . . . . . . . . . . . . 78
3.2.2 Adaptive Threshold Value Decision . . . . . . . . . . . . 84
3.3 Experimental Results . . . . . . . . . . . . . . . . . . . . . . . . 85
3.4 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
4 Error Concealment Algorithm for Successive Frame Losses 93
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
4.2 Optical Flow and Previous Works . . . . . . . . . . . . . . . . . 94
4.3 Motivation and the Proposed Algorithm . . . . . . . . . . . . . . 95
4.3.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . 96
4.3.2 Proposed Algorithm . . . . . . . . . . . . . . . . . . . . 97
4.4 Experimental Results . . . . . . . . . . . . . . . . . . . . . . . . 101
4.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
5 Algorithm and Hardware Architecture Design of Error Concealment
Engine for H.264/AVC 105
5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
5.2 Previous Works and Problem Statements . . . . . . . . . . . . . . 107
5.2.1 Spatial Error Concealment (SEC) . . . . . . . . . . . . . 107
5.2.2 Temporal Error Concealment (TEC) . . . . . . . . . . . . 107
5.2.3 Concealment Mode Selection . . . . . . . . . . . . . . . 109
5.2.4 Problem Statements . . . . . . . . . . . . . . . . . . . . 109
5.3 Proposed Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . 112
5.3.1 Frame-level Scene-Change Detection . . . . . . . . . . . 113
5.3.2 Spatial Error Concealment . . . . . . . . . . . . . . . . . 114
5.3.3 Temporal Error Concealment . . . . . . . . . . . . . . . . 118
5.4 Hardware Implementation . . . . . . . . . . . . . . . . . . . . . 126
5.4.1 Spatial Mode Selection Unit . . . . . . . . . . . . . . . . 127
5.4.2 BI/DI Interpolator . . . . . . . . . . . . . . . . . . . . . 129
5.4.3 Architecture of Prediction-based MVE . . . . . . . . . . 130
5.4.4 Motion Compensation Unit . . . . . . . . . . . . . . . . 137
5.5 Experimental Results . . . . . . . . . . . . . . . . . . . . . . . . 137
5.5.1 SEC Results . . . . . . . . . . . . . . . . . . . . . . . . 138
5.5.2 TEC Results . . . . . . . . . . . . . . . . . . . . . . . . 139
5.5.3 Chip Implementation Results and Comparison . . . . . . 148
5.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
6 Conclusion 155
6.1 Principal Contributions . . . . . . . . . . . . . . . . . . . . . . . 155
6.1.1 Perception-Aware Video Encoding System . . . . . . . . 155
6.1.2 Error Robust Video Decoding System . . . . . . . . . . . 156
6.2 Future Directions . . . . . . . . . . . . . . . . . . . . . . . . . . 157
6.2.1 Algorithm and Architecture of Visual Attention Model . . 157
6.2.2 Perception-Aware Scalable Video Coding . . . . . . . . . 158
6.2.3 Perception-Aware Video Communication Supporting Cross-
Layer Design . . . . . . . . . . . . . . . . . . . . . . . . 158
6.2.4 Algorithm and Architecture of Error Robust Video En-
coding System . . . . . . . . . . . . . . . . . . . . . . . 159
dc.language.isoen
dc.subject錯誤偵測zh_TW
dc.subject視訊傳輸zh_TW
dc.subject視訊硬體zh_TW
dc.subject感知視訊編碼zh_TW
dc.subject錯誤隱蔽zh_TW
dc.subjecterror detectionen
dc.subjecterror concealmenten
dc.subjecterror detectionen
dc.subjectvideo transmissionen
dc.subjectvideo hardwareen
dc.subjectperceptual video codingen
dc.title應用於視訊傳輸上錯誤補償與感知視訊編碼系統之演算法與硬體架構研究zh_TW
dc.titleAlgorithm and Hardware Architecture Design of Error Concealment and Perceptual Video Coding for Video Communicationen
dc.typeThesis
dc.date.schoolyear98-2
dc.description.degree博士
dc.contributor.oralexamcommittee李佩君(Pei-Jun Lee),李鎮宜,廖弘源,唐之瑋,吳安宇,楊家輝,陳良基
dc.subject.keyword錯誤隱蔽,視訊傳輸,視訊硬體,感知視訊編碼,錯誤偵測,zh_TW
dc.subject.keyworderror concealment,error detection,video transmission,video hardware,perceptual video coding,error detection,en
dc.relation.page174
dc.rights.note有償授權
dc.date.accepted2010-08-20
dc.contributor.author-college電機資訊學院zh_TW
dc.contributor.author-dept電子工程學研究所zh_TW
顯示於系所單位:電子工程學研究所

文件中的檔案:
檔案 大小格式 
ntu-99-1.pdf
  未授權公開取用
3.82 MBAdobe PDF
顯示文件簡單紀錄


系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved