Plenary overview session II — Image, video and multimedia processing
Session chair: C.C. Jay Kuo
Title: An overview of image/video compression down the history lane
Speaker: Professor Oscar C. Au
Affiliation: Hong Kong University of Science and Technology
In this plenary overview talk, I will take the audience to take a brief walk down the history lane to review the various developmental stages of image and video compression standards, including MPEG-1, MPEG-2, MPEG-4, H.261, H..263, H.264, and the latest HEVC video standards, and JPEG and JPEG 2000 image standards. Key features of the latest standards will be introduced. Major industrial and commercial applications such as digital TV broadcasting, set-up box, DVD, Blu-Ray, multi-media enabled smart phone, PMP, video conferencing, digital camera, etc. will be highlighted. I will also briefly describe the exciting and challenging game of standard.
Speaker Photo and Bio
Oscar C. Au received his B.A.Sc. from Univ. of Toronto in 1986, his M.A. and Ph.D. from Princeton Univ. in 1988 and 1991 respectively. After being a postdoctoral researcher in Princeton Univ. for one year, he joined the Hong Kong University of Science and Technology (HKUST) as an Assistant Professor in 1992. He is/was a Professor of the Dept. of Electronic and Computer Engineering, Director of Multimedia Technology Research Center (MTrec), and Director of the Computer Engineering (CPEG) Program in HKUST.
His main research contributions are on video and image coding and processing, watermarking and light weight encryption, speech and audio processing. Research topics include fast motion estimation for MPEG-1/2/4, H.261/3/4 andAVS, optimal and fast sub-optimal rate control, mode decision, transcoding, denoising, deinterlacing, post-processing, multi-view coding, view interpolation, depth estimation, 3DTV, scalable video coding, distributed video coding, subpixel rendering, JPEG/JPEG2000, HDR imaging, compressive sensing, halftone image data hiding, GPU-processing, software-hardware co-design, etc. He has published 50+ technical journal papers, 330+ conference papers, and 70+ contributions to international standards. His fast motion estimation algorithms were accepted into the ISO/IEC 14496-7 MPEG-4 international video coding standard and the ChinaAVS-M standard. His light-weight encryption and error resilience algorithms are accepted into the China AVS standard. He was Chair of Screen Content Coding AdHoc Group in the JCTVC for the ITU-T H.265 HEVC video coding standard. He has 18 granted US patents and is applying for 80+ more on his signal processing techniques. He has performed forensic investigation and stood as an expert witness in the Hong Kong courts many times.
Dr. Au is a Fellow of the Institute of Electrical and Electronic Engineering (IEEE) and is a Board 0fGovernor member of the Asia Pacific Signal and Information Processing Association (APSIPA). He is/was Associate Editors of IEEE Trans. On Circuits and Systems for Video Technology (TCSVT), IEEE Trans. on Image Processing (TIP), and IEEE Trans. on Circuits and Systems, Part 1 (TCAS1). He is on the Editorial Boards of Journal of Visual Communication and Image Representation (JVCIR), Journal of Signal Processing Systems (JSPS), APSIPA Trans. On Signal and Information Processing (TSIP), Journal of Multimedia (JMM), and Journal of Franklin Institute (JFI). He is/was Chair of IEEE CAS Technical Committee on Multimedia Systems and Applications (MSATC), Chair of SP TC on Multimedia Signal Processing (MMSP), and Chair of APSIPA TC on Image, Video and Multimedia (IVM). He is a member of CAS TC on Video Signal Processing and Communications (VSPC), CAS TC on Digital Signal Processing (DSP), SP TC on Image, Video and Multidimensional Signal Processing (IVMSP), SP TC on Information Forensics and Security (IFS), and ComSoc TC on Multimedia Communications (MMTC). He served on the Steering Committee of IEEE Trans. On Multimedia (TMM), and IEEE Int. Conf. of Multimedia and Expo (ICME). He also served on the organizing committee of IEEE Int. Symposium on Circuits and Systems (ISCAS) in 1997, IEEE Int. Conf. On Acoustics, Speech and Signal Processing (ICASSP) in 2003, the ISO/IEC MPEG 71st Meeting in 2005, Int. Conf. on Image Processing (ICIP) in 2010, and other conferences. He was General Chair of Pacific-Rim Conference on Multimedia (PCM) in 2007, IEEE Int. Conf. on Multimedia and Expo (ICME) in 2010 and the International Packet Video Workshop (PV) in 2010. He won best paper awards in SiPS 2007, PCM 2007 and MMSP 2012. He is an IEEE Distinguished Lecturer (DLP) in 2009 and 2010, and has been keynote speaker for multiple times.
Title: Radiometric Compensation for Ubiquitous Projection
Speaker: Homer H. Chen
Affiliation: National Taiwan University, Taiwan
Integrating pico-projectors in mobile and wearable devices allows nearby walls, desks, or bodies to be conveniently used as projection surfaces. Ubiquitous projection is no longer a fiction. Although the miniature of projection technology brings about new form of social interaction and augmented reality, ubiquitous projection is not free of challenges because the ad hoc projection surfaces are not as perfect as the regular projection screens. This paper discusses the technical issues of and solutions for mobile projectors, particularly in the area of radiometric compensation. It provides an overview of the calibration techniques for static and moving pico-projectors and the techniques for combating the physical limitations of projectors and cameras in brightness and dynamic range. Suggestions of future research are also provided.
Speaker Photo and Bio
Homer H. Chen(M’86-SM’01-F’03) received the Ph.D. degree in Electrical and Computer Engineering from University of Illinois at Urbana-Champaign.
Dr. Chen is a renowned expert in multimedia signal processing and communications. His professional career has spanned across academia and industry. Since August 2003, he has been with the College of Electrical Engineering and Computer Science, National Taiwan University, where he is Irving T. Ho Chair Professor. Prior to that, he held various R&D management and engineering positions with U.S. companies over a period of 17 years, including AT&T Bell Labs, Rockwell Science Center, iVast, and Digital Island (acquired by Cable & Wireless). He was a U.S. delegate for ISO and ITU standards committees and contributed to the development of many new interactive multimedia technologies that are now part of the MPEG-4 and JPEG-2000 standards. His professional interests lie in the broad area of multimedia signal processing and communications.
Dr. Chen is an IEEE Fellow. As a Technical Program Co-Chair, he helped revamp the IEEE International Conference on Multimedia and Expo in 2010. He was an Associate Editor of IEEE Transactions on Circuits and Systems for Video Technology from 2004 to 2010, IEEE Transactions on Image Processing from 1992 to 1994, and Pattern Recognition from 1989 to 1999. He served as a Guest Editor for IEEE Transactions on Circuits and Systems for Video Technology in 1999 and for IEEE Transactions on Multimedia in 2011.
Title: 3D Video: Some Challenging Topics
Speaker: Professor Hsueh-Ming Hang
Affiliation: National Chiao Tung University, Taiwan
3D films have a history of a few decades, but only recently 3D video consumer products are gradually becoming popular. The next negation of 3D research focus is on the so-called virtual-viewpoint (or free-viewpoint) video system. It is also an on-going standardization item in the international ITU/MPEG Standards. In addition to the movie special effects, one application of virtual-view system is the multi-view auto-stereoscopic (glasses free 3D) display. Typically, a densely arranged camera array is used to acquire input images and a number of virtual view pictures are synthesized at the player using the depth-image based rendering (DIBR) technique. Three essential components are needed for building a virtual-view system: depth estimation, data compression, and view synthesis. Another issue is how human being judges the quality of 3D video. There exist a number of 2D image/video quality assessment models, but the 3D perceptual modeling is largely under development. We will give a quick overview of these challenging topics and summarize their recent progresses.
Speaker Photo and Bio
Hsueh-Ming Hang received the B.S. and M.S. degrees in control engineering and electronics engineering from National Chiao Tung University, Hsinchu, Taiwan, in 1978 and 1980, respectively, and Ph.D. in electrical engineering from Rensselaer Polytechnic Institute, Troy, NY, in 1984.
From 1984 to 1991, he was with AT&T Bell Laboratories, Holmdel, NJ, and then he joined the Electronics Engineering Department of National Chiao Tung University (NCTU), Hsinchu, Taiwan, in December 1991. From 2006 to 2009, he took a leave from NCTU and was appointed as Dean of the EECS College at National Taipei University of Technology (NTUT). He is currently a Distinguished Professor of the EE Dept at NCTU and an associate dean of the ECE College, NCTU. He has been actively involved in the international MPEG standards since 1984 and his current research interests include multimedia compression, image/signal processing algorithms and architectures, and multimedia communication systems.
Dr. Hang holds 13 patents (Taiwan, US and Japan) and has published over 190 technical papers related to image compression, signal processing, and video codec architecture. He was an associate editor (AE) of the IEEE Transactions on Image Processing (TIP, 1992-1994), the IEEE Transactions on Circuits and Systems for Video Technology (1997-1999), and currently an AE of the IEEE TIP again. He is a co-editor and contributor of the Handbook of Visual Communications published by Academic Press. He is currently a Distinguished Lecturer and Board Member of the Asia-Pacific Signal and Information Processing Association (APSIPA) (2012-2013). He will be a general co-chair of APSIPA ASC 2014. He is a recipient of the IEEE Third Millennium Medal and is a Fellow of IEEE and IET and a member of Sigma Xi.
Title: An introduction to Graph Signal Processing
Speaker: Professor Antonio Ortega
Affiliation: University of Southern California
Graphs have long been used in a wide variety of problems, such analysis of social networks, machine learning, network protocol optimization, decoding of LDPCs or image processing. Techniques based on spectral graph theory provide a “frequency” interpretation of graph data and have proven to be quite popular in multiple applications.
In the last few years, a growing amount of work has started extending and complementing spectral graph techniques, leading to the emergence of “Graph Signal Processing” as a broad research field. A common characteristic of this recent work is that it considers the data attached to the vertices as a “graph-signal” and seeks to create new techniques (filtering, sampling, interpolation), similar to those commonly used in conventional signal processing (for audio, images or video), so that they can be applied to these graph signals.
In this talk, we first introduce some of the basic tools needed in developing new graph signal processing operations. We then introduce our design of wavelet filterbanks of graphs, which for the first time provides a multi-resolution, critically-sampled, frequency- and graph-localized transforms for graph signals. We conclude by providing several examples of how these new transforms and tools can be applied to existing problems. Time permitting, we will discuss applications to image processing, depth video compression, recommendation system design and network optimization.
Speaker Photo and Bio
Antonio Ortega received the Telecommunications Engineering degree from the Universidad Politecnica de Madrid, Madrid, Spain in 1989 and the Ph.D. in Electrical Engineering from Columbia University, New York, NY in 1994. At Columbia he was supported by a Fulbright scholarship.
In 1994 he joined the Electrical Engineering department at the University of Southern California (USC), where he is currently a Professor. He has served as Associate Chair of EE-Systems and director of the Signal and Image Processing Institute at USC. He is a Fellow of the IEEE, and a member of ACM and APSIPA. He has been Chair of the Image and Multidimensional Signal Processing (IMDSP) technical committee and a member of the Board of Governors of the IEEE Signal Processing Society (2002). He has been technical program co-chair of ICIP 2008, MMSP 1998 and ICME 2002. He has been Associate Editor for the IEEE Transactions on Image Processing, the IEEE Signal Processing Letters, the EURASIP Journal on Advances in Signal Processing, and the IEEE Signal Processing Magazine. He is the inaugural Editor-in-Chief of the APSIPA Transactions on Signal and Information Processing, launched by APSIPA and Cambridge University Press in 2012. He received the NSF CAREER award, the 1997 IEEE Communications Society Leonard G. Abraham Prize Paper Award, the IEEE Signal Processing Society 1999 Magazine Award, the 2006 EURASIP Journal of Advances in Signal Processing Best Paper Award, the ICIP 2011 best paper award, and a best paper award at Globecom 2012.
His research interests are in the areas of multimedia compression, communications and signal analysis. His recent work is focusing on distributed compression, multiview coding, error tolerant compression, wavelet-based signal analysis, information representation in wireless sensor networks and graph signal processing. His work at USC has been or is being funded by agencies such as NSF, NASA, DOE, and companies such as HP, Samsung, Chevron or Texas Instruments. Over 30 PhD students have completed their PhD thesis under his supervision at USC and his work has led to over 300 publications in international conferences and journals, as well as several patents.