Catálogo de publicaciones - libros
Advances in Multimedia Information Processing: 8th Pacific Rim Conference on Multimedia, Hong Kong, China, December 11-14, 2007. Proceedings
Horace H.-S. Ip ; Oscar C. Au ; Howard Leung ; Ming-Ting Sun ; Wei-Ying Ma ; Shi-Min Hu (eds.)
En conferencia: 8º Pacific-Rim Conference on Multimedia (PCM) . Hong Kong, China . December 11, 2007 - December 14, 2007
Resumen/Descripción – provisto por la editorial
No disponible.
Palabras clave – provistas por la editorial
Computer Applications; Multimedia Information Systems; Information Storage and Retrieval; Computer Communication Networks; Information Systems Applications (incl. Internet); Image Processing and Computer Vision
Disponibilidad
Institución detectada | Año de publicación | Navegá | Descargá | Solicitá |
---|---|---|---|---|
No detectada | 2007 | SpringerLink |
Información
Tipo de recurso:
libros
ISBN impreso
978-3-540-77254-5
ISBN electrónico
978-3-540-77255-2
Editor responsable
Springer Nature
País de edición
Reino Unido
Fecha de publicación
2007
Información sobre derechos de publicación
© Springer-Verlag Berlin Heidelberg 2007
Cobertura temática
Tabla de contenidos
Real-Time Facial Feature Point Extraction
Ce Zhan; Wanqing Li; Philip Ogunbona; Farzad Safaei
Localization of facial feature points is an important step for many subsequent facial image analysis tasks. In this paper, we proposed a new coarse-to-fine method for extracting 20 facial feature points from image sequences. In particular, the Viola-Jones face detection method is extended to detect small-scale facial components with wide shape variations, and linear Kalman filters are used to smoothly track the feature points by handling detection errors and head rotations. The proposed method achieved higher than 90% detection rate when tested on the BioID face database and the FG-NET facial expression database. Moreover, our method shows robust performance against the variation of face resolutions and facial expressions.
- Session-2: Human Face and Action Recognition | Pp. 88-97
Local Dual Closed Loop Model Based Bayesian Face Tracking
Dan Yao; Hong Lu; Xiangyang Xue; Zhongyi Zhou
This paper presents a new Bayesian face tracking method under particle filter framework. First, two adaptive feature models are proposed to extract face features from image sequences. Then the robustness of face tracking is reinforced via building a local dual closed loop model (LDCLM). Meanwhile, trajectory analysis, which helps to avoid unnecessary restarting of detection module, is introduced to keep tracked faces’ identity as consistent as possible. Experimental results demonstrate the efficacy of our method.
- Session-2: Human Face and Action Recognition | Pp. 98-107
View-Independent Human Action Recognition by Action Hypersphere in Nonlinear Subspace
Jian Zhang; Yueting Zhuang
Though recognizing human action from video is important to applications like visual surveillance, some hurdles still slower the progress of action recognition. One of the main difficulties is view dependency, and this causes the degeneration of many recognition algorithms. In this paper, we propose a template-based view-independent human action recognition approach. The action template comprises a series of “action hyperspheres” in a nonlinear subspace and encodes multi-view information of several typical human actions to facilitate the view-independent recognition. Given an input action from video, we first compute the Motion History Image (MHI) and corresponding polar feature according to the extracted human silhouettes; recognition is achieved by evaluating the distances between the embedding of the polar feature and the virtual centers of the hyperspheres. Experiments show that our approach maintains high recognition accuracy in free viewpoints, and is more computationally efficient compared with classical NN approach.
- Session-2: Human Face and Action Recognition | Pp. 108-117
Efficient Adaptive Background Subtraction Based on Multi-resolution Background Modelling and Updating
Ruijiang Luo; Liyuan Li; Irene Yu-Hua Gu
Adaptive background subtraction (ABS) is a fundamental step for foreground object detection in many real-time video surveillance systems. In many ABS methods, a pixel-based statistical model is used for the background and each pixel is updated online to adapt to various background changes. As a result, heavy computation and memory consumption are required. In this paper, we propose an efficient methodology for implementation of ABS algorithms based on multi-resolution background modelling and sequential sampling for updating background. Experiments and quantitative evaluation are conducted on two open data sets (PETS2001 and PETS2006) and scenarios captured in some public places, and some results are included. Our results have shown that the proposed method requires a significant reduction in memory and CPU usage, meanwhile maintaining a similar foreground segmentation performance as compared with the corresponding single resolution methods.
- Session-2: Human Face and Action Recognition | Pp. 118-127
Towards a Stringent Bit-Rate Conformance for Frame-Layer Rate Control in H.264/AVC
Evan Tan; Jing Chen
This paper presents a novel frame-layer rate control technique that adaptively determines the frame complexity for bit allocation in order to satisfy the target bit-rate constraints without degrading the decoded video significantly. To do this, we first obtain the edge energy of each frame to measure the frame complexity as well as to determine the weighting of a frame for bit allocation. We then present a new bit-rate traffic model for bit allocation to achieve a better conformance to the target bit-rate. Finally, we integrate the edge energy complexity measure into the rate-quantization (R-Q) model. Our results shows robust improvements over the current rate control methods adopted in H.264/AVC in terms of meeting the target bit-rate as well as determining the quality of the decoded video.
- Session-3: H.264 Video Coding | Pp. 128-137
A Quantized Transform-Domain Motion Estimation Technique for H.264 Secondary SP-Frames
Ki-Kit Lai; Yui-Lam Chan; Wan-Chi Siu
The brand-new SP-frame in H.264 facilitates drift-free bitstream switching. Notwithstanding the guarantee of seamless switching, the cost is the bulky size of secondary SP-frames. This induces a significant amount of additional space or bandwidth for storage or transmission. For this reason, a new motion estimation and compensation technique, which is operated in the quantized transform (QDCT) domain, is designed for coding secondary SP-frames in this paper. So far, much investigation has been conducted to evaluate the trade off between the relative sizes of primary and secondary SP-frames by adjusting the quantization parameters. But, our proposed work aims at keeping the secondary SP-frames as small as possible without affecting the size of primary SP-frames by incorporating QDCT-domain motion estimation and compensation in the secondary SP-frame coding. Simulation results demonstrate that the size of secondary SP-frames can be reduced remarkably.
- Session-3: H.264 Video Coding | Pp. 138-147
Efficient Intra Mode Decision Via Statistical Learning
Chiuan Hwang; Shang-Hong Lai
Intra mode selection and motion estimation for spatial and temporal prediction play important roles for achieving high video compression ratio in the latest video coding standards, such as H.264/AVC. However, both components take most of the computational cost in the video encoding process. In this paper, we propose an efficient intra mode prediction algorithm based on using the mode conditional probability learned from a large amount of training video sequences with the ground truth modes of each block to be encoded and its neighboring block modes as well as its associated image content features. By applying the proposed intra-mode selection algorithm into the H.264 reference code, we show significant reduction of the computation time with negligible video quality degradation for H.264 video encoding.
- Session-3: H.264 Video Coding | Pp. 148-157
Fast Mode Decision Algorithms for Inter/Intra Prediction in H.264 Video Coding
Ling-Jiao Pan; Seung-Hwan Kim; Yo-Sung Ho
In this paper, we propose fast mode decision algorithms for both intra prediction and inter prediction in H.264. In order to select the candidate modes for intra4x4 and intra16x16 prediction efficiently, we have used the spatial correlation and directional information. We have also applied an early block size selection method to reduce the searching time further. The fast inter mode decision is achieved by an early SKIP mode decision method, and a fast mode decision method for 16x16 and P8x8 modes. Extensive simulations on different test sequences demonstrate a considerable speed up by saving the encoding time up to 82% for intra prediction and 77% for inter prediction on average, compared to the H.264 standard, respectively. This is achieved at the cost of negligible loss in PSNR values and small increase in bit rates.
- Session-3: H.264 Video Coding | Pp. 158-167
A Novel Fast Motion Estimation Algorithm Based on SSIM for H.264 Video Coding
Chun-ling Yang; Hua-xing Wang; Lai-Man Po
H.264 achieves considerable higher coding efficiency compared with previous video coding standards, whereas the complexity is increased significantly. This paper proposed a novel fast algorithm based on structural similarity (SSIM) in motion estimation (ME) process (FMEBSS), which can greatly reduced the complexity of ME by eliminating the unnecessary search positions and reducing complex prediction modes. Simulation results demonstrate that the proposed method can averagely reduce the coding time by about 50%, and compression ratio is improved at the same time, while the degradation in video quality is negligible.
- Session-3: H.264 Video Coding | Pp. 168-176
Shot Boundary Detection for H.264/AVC Bitstreams with Frames Containing Multiple Types of Slices
Sarah De Bruyne; Wesley De Neve; Davy De Schrijver; Peter Lambert; Piet Verhoeve; Rik Van de Walle
In this paper, a novel shot boundary detection algorithm is introduced for H.264/AVC bitstreams. This algorithm relies on features present in a compressed video bitstream, i.e., macroblock types and reference directions. In contrast to existing algorithms, the proposed technique is able to analyze bitstreams with frames containing different types of slices. To deal with such frames, formulas are presented that work on inter- and intra-coded slices. The results obtained for the different types of slices are combined by calculating a linear combination, taking into account their size. This way, a metric is found that can be used for the automatic detection of shot boundaries. Results show that the proposed algorithm has a high accuracy in terms of recall and precision for video sequences with frames containing multiple slice types.
- Session-4: Video Analysis and Retrieval | Pp. 177-186