Catálogo de publicaciones - libros
Image and Video Retrieval: 5th Internatinoal Conference, CIVR 2006, Tempe, AZ, USA, July 13-15, 2006, Proceedings
Hari Sundaram ; Milind Naphade ; John R. Smith ; Yong Rui (eds.)
En conferencia: 5º International Conference on Image and Video Retrieval (CIVR) . Tempe, AZ, USA . July 13, 2006 - July 15, 2006
Resumen/Descripción – provisto por la editorial
No disponible.
Palabras clave – provistas por la editorial
Computer Graphics; Information Storage and Retrieval; Database Management; Information Systems Applications (incl. Internet); Multimedia Information Systems; Image Processing and Computer Vision
Disponibilidad
| Institución detectada | Año de publicación | Navegá | Descargá | Solicitá |
|---|---|---|---|---|
| No detectada | 2006 | SpringerLink |
Información
Tipo de recurso:
libros
ISBN impreso
978-3-540-36018-6
ISBN electrónico
978-3-540-36019-3
Editor responsable
Springer Nature
País de edición
Reino Unido
Fecha de publicación
2006
Información sobre derechos de publicación
© Springer-Verlag Berlin Heidelberg 2006
Cobertura temática
Tabla de contenidos
doi: 10.1007/11788034_41
Multidimensional Descriptor Indexing: Exploring the BitMatrix
Catalin Calistru; Cristina Ribeiro; Gabriel David
Multimedia retrieval brings new challenges, mainly derived from the mismatch between the level of the user interaction—high-level concepts, and that of the automatically processed descriptors—low-level features. The effective use of the low-level descriptors is therefore mandatory. Many data structures have been proposed for managing the representation of multidimensional descriptors, each geared toward efficiency in some set of basic operations. The paper introduces a highly parametrizable structure called the BitMatrix, along with its search algorithms. The BitMatrix is compared with existing methods, all implemented in a common framework . The tests have been performed on two datasets, with parameters covering significant ranges of values. The BitMatrix has proved to be a robust and flexible structure that can compete with other methods for multidimensional descriptor indexing.
Palabras clave: Recall Rate; Indexing Method; Query Object; Multimedia Object; Query Signature.
- Session P2: Poster II | Pp. 401-410
doi: 10.1007/11788034_42
Natural Scene Image Modeling Using Color and Texture Visterms
Pedro Quelhas; Jean-Marc Odobez
This paper presents a novel approach for visual scene representation, combining the use of quantized color and texture local invariant features (referred to here as visterms ) computed over interest point regions. In particular we investigate the different ways to fuse together local information from texture and color in order to provide a better visterm representation. We develop and test our methods on the task of image classification using a 6-class natural scene database. We perform classification based on the bag-of-visterms (BOV) representation (histogram of quantized local descriptors), extracted from both texture and color features. We investigate two different fusion approaches at the feature level: fusing local descriptors together and creating one representation of joint texture-color visterms, or concatenating the histogram representation of both color and texture, obtained independently from each local feature. On our classification task we show that the appropriate use of color improves the results w.r.t. a texture only representation.
Palabras clave: Support Vector Machine; Interest Point; Natural Scene; Local Descriptor; Fusion Scheme.
- Session P2: Poster II | Pp. 411-421
doi: 10.1007/11788034_43
Online Image Retrieval System Using Long Term Relevance Feedback
Lutz Goldmann; Lars Thiele; Thomas Sikora
This paper describes an original system for content based image retrieval. It is based on MPEG-7 descriptors and a novel approach for long term relevance feedback using a Bayesian classifier. Each image is represented by a special model that is adapted over multiple feedback rounds and even multiple sessions or users. The experiments show its outstanding performance in comparison to often used short term relevance feedback and the recently proposed FIRE system.
Palabras clave: Feature Vector; Image Retrieval; Relevance Feedback; Incremental Learning; Content Base Image Retrieval.
- Session P2: Poster II | Pp. 422-431
doi: 10.1007/11788034_44
Perceptual Distance Functions for Similarity Retrieval of Medical Images
Joaquim Cezar Felipe; Agma Juci Machado Traina; Caetano Traina
A challenge already opened for a long time concerning Content-based Image Retrieval (CBIR) systems is how to define a suitable distance function to measure the similarity between images regarding an application context, which complies with the human specialist perception of similarity. In this paper, we present a new family of distance functions, namely, Attribute Interaction Influence Distances (AID), aiming at retrieving images by similarity. Such measures address an important aspect of psychophysical comparison between images: the effect in the interaction on the variations of the image features. The AID functions allow comparing feature vectors using two parameterized expressions: one targeting weak feature interaction; and another for strong interaction. This paper also presents experimental results with medical images, showing that when the reference is the radiologist perception, AID works better than the distance functions most commonly used in CBIR.
Palabras clave: Feature Vector; Distance Function; Medical Image; Image Retrieval; Attribute Interaction.
- Session P2: Poster II | Pp. 432-442
doi: 10.1007/11788034_45
Using Score Distribution Models to Select the Kernel Type for a Web-Based Adaptive Image Retrieval System (AIRS)
Anca Doloc-Mihu; Vijay V. Raghavan
The goal of this paper is to investigate the selection of the kernel for a Web-based AIRS. Using the Kernel Rocchio learning method, several kernels having polynomial and Gaussian forms are applied to general images represented by color histograms in RGB and HSV color spaces. Experimental results on these collections show that performance varies significantly between different kernel types and that choosing an appropriate kernel is important. Then, based on these results, we propose a method for selecting the kernel type that uses the score distribution models. Experimental results on our data show that the proposed method is effective for our system.
- Session P2: Poster II | Pp. 443-452
doi: 10.1007/11788034_46
Semantics Supervised Cluster-Based Index for Video Databases
Zhiping Shi; Qingyong Li; Zhiwei Shi; Zhongzhi Shi
High-dimensional index is one of the most challenging tasks for content-based video retrieval (CBVR). Typically, in video database, there exist two kinds of clues for query: visual features and semantic classes. In this paper, we modeled the relationship between semantic classes and visual feature distributions of data set with the Gaussian mixture model (GMM), and proposed a semantics supervised cluster based index approach (briefly as SSCI) to integrate the advantages of both semantic classes and visual features. The entire data set is divided hierarchically by a modified clustering technique into many clusters until the objects within a cluster are not only close in the visual feature space but also within the same semantic class, and then an index entry including semantic clue and visual feature clue is built for each cluster. Especially, the visual feature vectors in a cluster are organized adjacently in disk. So the SSCI-based nearest-neighbor (NN) search can be divided into two phases: the first phase computes the distances between the query example and each cluster index and returns the clusters with the smallest distance, here namely candidate clusters; then the second phase retrieves the original feature vectors within the candidate clusters to gain the approximate nearest neighbors. Our experiments showed that for approximate searching the SSCI-based approach was faster than VA^ + -based approach; moreover, the quality of the result set was better than that of the sequential search in terms of semantics.
Palabras clave: High-dimensional index; cluster; video semantics; video database.
- Session P2: Poster II | Pp. 453-462
doi: 10.1007/11788034_47
Semi-supervised Learning for Image Annotation Based on Conditional Random Fields
Wei Li; Maosong Sun
Automatic image annotation (AIA) has been proved to be an effective and promising solution to automatically deduce the high-level semantics from low-level visual features. Due to the inherent ambiguity of image-label mapping and the scarcity of training examples, it has become a challenge to systematically develop robust annotation models with better performance. In this paper, we try to attack the problem based on 2D CRFs (Conditional Random Fields) and semi-supervised learning which are seamlessly integrated into a unified framework. 2D CRFs can effectively capture the spatial dependency between the neighboring labels, while the semi-supervised learning techniques can exploit the unlabeled data to improve the joint classification performance. We conducted experiments on a medium-sized image collection including about 500 images from Corel Stock Photo CDs. The experimental results demonstrated that the annotation performance of this method outperforms standard CRFs, showing the effectiveness of the proposed unified framework and the feasibility of unlabeled data to help the classification accuracy.
Palabras clave: Image Patch; Unlabeled Data; Latent Semantic Analysis; Conditional Random Field; Image Annotation.
- Session P2: Poster II | Pp. 463-472
doi: 10.1007/11788034_48
NPIC: Hierarchical Synthetic Image Classification Using Image Search and Generic Features
Fei Wang; Min-Yen Kan
We introduce NPIC, an image classification system that focuses on synthetic (e.g., non-photographic) images. We use class-specific keywords in an image search engine to create a noisily labeled training corpus of images for each class. NPIC then extracts both content-based image retrieval (CBIR) features and metadata-based textual features for each image for machine learning. We evaluate this approach on three different granularities: 1) natural vs. synthetic, 2) map vs. figure vs. icon vs. cartoon vs. artwork 3) and further subclasses of the map and figure classes. The NPIC framework achieves solid performance (99%, 97% and 85% in cross validation, respectively). We find that visual features provide a significant boost in performance, and that textual and visual features vary in usefulness at the different levels of granularities of classification.
Palabras clave: Textual Feature; Visual Feature; Color Histogram; Synthetic Image; Content Base Image Retrieval.
- Session P2: Poster II | Pp. 473-482
doi: 10.1007/11788034_49
Context-Aware Media Retrieval
Ankur Mani; Hari Sundaram
In this paper we propose a representation framework for dynamic multi-sensory knowledge and user context, and its application in media retrieval. We provide a definition of context, the relationship between context and knowledge and the importance of communication both as a means for the building of context as well as the end achieved by the context. We then propose a model of user context and demonstrate its application in a photo retrieval application. Our experiments demonstrate the advantages of the context-aware media retrieval over other media retrieval approaches especially relevance feedback.
Palabras clave: Context modeling; media retrieval; knowledge representation.
- Session A: ASU Special Session | Pp. 483-486
doi: 10.1007/11788034_50
Estimating the Physical Effort of Human Poses
Yinpeng Chen; Hari Sundaram; Jodi James
This paper deals with the problem of estimating the effort required to maintain a static pose by human beings. The problem is important in developing effective pose classification as wells as in developing models of human attention. We estimate the human pose effort using two kinds of body constraints – skeletal constraints and gravitational constraints. The extracted features are combined together using SVM regression to estimate the pose effort. We tested our algorithm on 55 poses with different annotated efforts with excellent results. Our user studies additionally validate our approach.
Palabras clave: Ground Truth; Support Vector Regression; User Study; Effort Estimation; Supporting Limb.
- Session A: ASU Special Session | Pp. 487-490