Catálogo de publicaciones - libros

Compartir en
redes sociales


Image and Video Retrieval: 5th Internatinoal Conference, CIVR 2006, Tempe, AZ, USA, July 13-15, 2006, Proceedings

Hari Sundaram ; Milind Naphade ; John R. Smith ; Yong Rui (eds.)

En conferencia: 5º International Conference on Image and Video Retrieval (CIVR) . Tempe, AZ, USA . July 13, 2006 - July 15, 2006

Resumen/Descripción – provisto por la editorial

No disponible.

Palabras clave – provistas por la editorial

Computer Graphics; Information Storage and Retrieval; Database Management; Information Systems Applications (incl. Internet); Multimedia Information Systems; Image Processing and Computer Vision

Disponibilidad
Institución detectada Año de publicación Navegá Descargá Solicitá
No detectada 2006 SpringerLink

Información

Tipo de recurso:

libros

ISBN impreso

978-3-540-36018-6

ISBN electrónico

978-3-540-36019-3

Editor responsable

Springer Nature

País de edición

Reino Unido

Fecha de publicación

Información sobre derechos de publicación

© Springer-Verlag Berlin Heidelberg 2006

Tabla de contenidos

Multidimensional Descriptor Indexing: Exploring the BitMatrix

Catalin Calistru; Cristina Ribeiro; Gabriel David

Multimedia retrieval brings new challenges, mainly derived from the mismatch between the level of the user interaction—high-level concepts, and that of the automatically processed descriptors—low-level features. The effective use of the low-level descriptors is therefore mandatory. Many data structures have been proposed for managing the representation of multidimensional descriptors, each geared toward efficiency in some set of basic operations. The paper introduces a highly parametrizable structure called the BitMatrix, along with its search algorithms. The BitMatrix is compared with existing methods, all implemented in a common framework . The tests have been performed on two datasets, with parameters covering significant ranges of values. The BitMatrix has proved to be a robust and flexible structure that can compete with other methods for multidimensional descriptor indexing.

Palabras clave: Recall Rate; Indexing Method; Query Object; Multimedia Object; Query Signature.

- Session P2: Poster II | Pp. 401-410

Natural Scene Image Modeling Using Color and Texture Visterms

Pedro Quelhas; Jean-Marc Odobez

This paper presents a novel approach for visual scene representation, combining the use of quantized color and texture local invariant features (referred to here as visterms ) computed over interest point regions. In particular we investigate the different ways to fuse together local information from texture and color in order to provide a better visterm representation. We develop and test our methods on the task of image classification using a 6-class natural scene database. We perform classification based on the bag-of-visterms (BOV) representation (histogram of quantized local descriptors), extracted from both texture and color features. We investigate two different fusion approaches at the feature level: fusing local descriptors together and creating one representation of joint texture-color visterms, or concatenating the histogram representation of both color and texture, obtained independently from each local feature. On our classification task we show that the appropriate use of color improves the results w.r.t. a texture only representation.

Palabras clave: Support Vector Machine; Interest Point; Natural Scene; Local Descriptor; Fusion Scheme.

- Session P2: Poster II | Pp. 411-421

Online Image Retrieval System Using Long Term Relevance Feedback

Lutz Goldmann; Lars Thiele; Thomas Sikora

This paper describes an original system for content based image retrieval. It is based on MPEG-7 descriptors and a novel approach for long term relevance feedback using a Bayesian classifier. Each image is represented by a special model that is adapted over multiple feedback rounds and even multiple sessions or users. The experiments show its outstanding performance in comparison to often used short term relevance feedback and the recently proposed FIRE system.

Palabras clave: Feature Vector; Image Retrieval; Relevance Feedback; Incremental Learning; Content Base Image Retrieval.

- Session P2: Poster II | Pp. 422-431

Perceptual Distance Functions for Similarity Retrieval of Medical Images

Joaquim Cezar Felipe; Agma Juci Machado Traina; Caetano Traina

A challenge already opened for a long time concerning Content-based Image Retrieval (CBIR) systems is how to define a suitable distance function to measure the similarity between images regarding an application context, which complies with the human specialist perception of similarity. In this paper, we present a new family of distance functions, namely, Attribute Interaction Influence Distances (AID), aiming at retrieving images by similarity. Such measures address an important aspect of psychophysical comparison between images: the effect in the interaction on the variations of the image features. The AID functions allow comparing feature vectors using two parameterized expressions: one targeting weak feature interaction; and another for strong interaction. This paper also presents experimental results with medical images, showing that when the reference is the radiologist perception, AID works better than the distance functions most commonly used in CBIR.

Palabras clave: Feature Vector; Distance Function; Medical Image; Image Retrieval; Attribute Interaction.

- Session P2: Poster II | Pp. 432-442

Using Score Distribution Models to Select the Kernel Type for a Web-Based Adaptive Image Retrieval System (AIRS)

Anca Doloc-Mihu; Vijay V. Raghavan

The goal of this paper is to investigate the selection of the kernel for a Web-based AIRS. Using the Kernel Rocchio learning method, several kernels having polynomial and Gaussian forms are applied to general images represented by color histograms in RGB and HSV color spaces. Experimental results on these collections show that performance varies significantly between different kernel types and that choosing an appropriate kernel is important. Then, based on these results, we propose a method for selecting the kernel type that uses the score distribution models. Experimental results on our data show that the proposed method is effective for our system.

- Session P2: Poster II | Pp. 443-452

Semantics Supervised Cluster-Based Index for Video Databases

Zhiping Shi; Qingyong Li; Zhiwei Shi; Zhongzhi Shi

High-dimensional index is one of the most challenging tasks for content-based video retrieval (CBVR). Typically, in video database, there exist two kinds of clues for query: visual features and semantic classes. In this paper, we modeled the relationship between semantic classes and visual feature distributions of data set with the Gaussian mixture model (GMM), and proposed a semantics supervised cluster based index approach (briefly as SSCI) to integrate the advantages of both semantic classes and visual features. The entire data set is divided hierarchically by a modified clustering technique into many clusters until the objects within a cluster are not only close in the visual feature space but also within the same semantic class, and then an index entry including semantic clue and visual feature clue is built for each cluster. Especially, the visual feature vectors in a cluster are organized adjacently in disk. So the SSCI-based nearest-neighbor (NN) search can be divided into two phases: the first phase computes the distances between the query example and each cluster index and returns the clusters with the smallest distance, here namely candidate clusters; then the second phase retrieves the original feature vectors within the candidate clusters to gain the approximate nearest neighbors. Our experiments showed that for approximate searching the SSCI-based approach was faster than VA^ + -based approach; moreover, the quality of the result set was better than that of the sequential search in terms of semantics.

Palabras clave: High-dimensional index; cluster; video semantics; video database.

- Session P2: Poster II | Pp. 453-462

Semi-supervised Learning for Image Annotation Based on Conditional Random Fields

Wei Li; Maosong Sun

Automatic image annotation (AIA) has been proved to be an effective and promising solution to automatically deduce the high-level semantics from low-level visual features. Due to the inherent ambiguity of image-label mapping and the scarcity of training examples, it has become a challenge to systematically develop robust annotation models with better performance. In this paper, we try to attack the problem based on 2D CRFs (Conditional Random Fields) and semi-supervised learning which are seamlessly integrated into a unified framework. 2D CRFs can effectively capture the spatial dependency between the neighboring labels, while the semi-supervised learning techniques can exploit the unlabeled data to improve the joint classification performance. We conducted experiments on a medium-sized image collection including about 500 images from Corel Stock Photo CDs. The experimental results demonstrated that the annotation performance of this method outperforms standard CRFs, showing the effectiveness of the proposed unified framework and the feasibility of unlabeled data to help the classification accuracy.

Palabras clave: Image Patch; Unlabeled Data; Latent Semantic Analysis; Conditional Random Field; Image Annotation.

- Session P2: Poster II | Pp. 463-472

NPIC: Hierarchical Synthetic Image Classification Using Image Search and Generic Features

Fei Wang; Min-Yen Kan

We introduce NPIC, an image classification system that focuses on synthetic (e.g., non-photographic) images. We use class-specific keywords in an image search engine to create a noisily labeled training corpus of images for each class. NPIC then extracts both content-based image retrieval (CBIR) features and metadata-based textual features for each image for machine learning. We evaluate this approach on three different granularities: 1) natural vs. synthetic, 2) map vs. figure vs. icon vs. cartoon vs. artwork 3) and further subclasses of the map and figure classes. The NPIC framework achieves solid performance (99%, 97% and  85% in cross validation, respectively). We find that visual features provide a significant boost in performance, and that textual and visual features vary in usefulness at the different levels of granularities of classification.

Palabras clave: Textual Feature; Visual Feature; Color Histogram; Synthetic Image; Content Base Image Retrieval.

- Session P2: Poster II | Pp. 473-482

Context-Aware Media Retrieval

Ankur Mani; Hari Sundaram

In this paper we propose a representation framework for dynamic multi-sensory knowledge and user context, and its application in media retrieval. We provide a definition of context, the relationship between context and knowledge and the importance of communication both as a means for the building of context as well as the end achieved by the context. We then propose a model of user context and demonstrate its application in a photo retrieval application. Our experiments demonstrate the advantages of the context-aware media retrieval over other media retrieval approaches especially relevance feedback.

Palabras clave: Context modeling; media retrieval; knowledge representation.

- Session A: ASU Special Session | Pp. 483-486

Estimating the Physical Effort of Human Poses

Yinpeng Chen; Hari Sundaram; Jodi James

This paper deals with the problem of estimating the effort required to maintain a static pose by human beings. The problem is important in developing effective pose classification as wells as in developing models of human attention. We estimate the human pose effort using two kinds of body constraints – skeletal constraints and gravitational constraints. The extracted features are combined together using SVM regression to estimate the pose effort. We tested our algorithm on 55 poses with different annotated efforts with excellent results. Our user studies additionally validate our approach.

Palabras clave: Ground Truth; Support Vector Regression; User Study; Effort Estimation; Supporting Limb.

- Session A: ASU Special Session | Pp. 487-490