Catálogo de publicaciones - libros

Compartir en
redes sociales


Feature Extraction: Foundations and Applications

Isabelle Guyon ; Masoud Nikravesh ; Steve Gunn ; Lotfi A. Zadeh (eds.)

Resumen/Descripción – provisto por la editorial

No disponible.

Palabras clave – provistas por la editorial

No disponibles.

Disponibilidad
Institución detectada Año de publicación Navegá Descargá Solicitá
No detectada 2006 SpringerLink

Información

Tipo de recurso:

libros

ISBN impreso

978-3-540-35487-1

ISBN electrónico

978-3-540-35488-8

Editor responsable

Springer Nature

País de edición

Reino Unido

Fecha de publicación

Información sobre derechos de publicación

© Springer Berlin Heidelberg 2006

Tabla de contenidos

High Dimensional Classification with Bayesian Neural Networks and Dirichlet Diffusion Trees

Radford M. Neal; Jianguo Zhang

Our winning entry in the NIPS 2003 challenge was a hybrid, in which our predictions for the five data sets were made using different methods of classification, or, for the Madelon data set, by averaging the predictions produced using two methods. However, two aspects of our approach were the same for all data sets:

Part II - Feature Selection Challenge | Pp. 265-296

Ensembles of Regularized Least Squares Classifiers for High-Dimensional Problems

Kari Torkkola; Eugene Tuv

It has been recently pointed out that the Regularized Least Squares Classifier (RLSC), continues to be a viable option for binary classification problems. We apply RLSC to the datasets of the NIPS 2003 Feature Selection Challenge using Gaussian kernels. Since RLSC is sensitive to noise variables, ensemble-based variable filtering is applied first. RLSC makes use of the best-ranked variables only. We compare the performance of a stochastic ensemble of RLSCs to a single best RLSC. Our results indicate that in terms of classification error rate the two are similar on the challenge data. However, especially with large data sets, ensembles could provide other advantages that we list.

Part II - Feature Selection Challenge | Pp. 297-313

Combining SVMs with Various Feature Selection Strategies

Yi-Wei Chen; Chih-Jen Lin

This article investigates the performance of combining support vector machines (SVM) and various feature selection strategies. Some of them are filter-type approaches: general feature selection methods independent of SVM, and some are wrapper-type methods: modifications of SVM which can be used to select features. We apply these strategies while participating to the NIPS 2003 Feature Selection Challenge and rank third as a group.

Part II - Feature Selection Challenge | Pp. 315-324

Feature Selection with Transductive Support Vector Machines

Zhili Wu; Chunhung Li

SVM-related feature selection has shown to be effective, while feature selection with transductive SVMs has been less studied. This paper investigates the use of transductive SVMs for feature selection, based on three SVM-related feature selection methods: filtering scores + SVM wrapper, recursive feature elimination (RFE) and multiplicative updates(MU). We show transductive SVMs can be tailored to feature selection by embracing feature scores for feature filtering, or acting as wrappers and embedded feature selectors. We conduct experiments on the feature selection competition tasks to demonstrate the performance of Transductive SVMs in feature selection and classification.

Part II - Feature Selection Challenge | Pp. 325-341

Variable Selection using Correlation and Single Variable Classifier Methods: Applications

Amir Reza Saffari Azar Alamdari

Correlation and single variable classifier methods are very simple algorithms to select a subset of variables in a dimension reduction problem, which utilize some measures to detect relevancy of a single variable to the target classes without considering the predictor properties to be used. In this paper, along with the description of correlation and single variable classifier ranking methods, the application of these algorithms to the NIPS 2003 Feature Selection Challenge problems is also presented. The results show that these methods can be used as one of primary, computational cost efficient, and easy to implement techniques which have good performance especially when variable space is very large. Also, it has been shown that in all cases using an ensemble averaging predictor would result in a better performance, compared to a single stand-alone predictor.

Part II - Feature Selection Challenge | Pp. 343-358

Tree-Based Ensembles with Dynamic Soft Feature Selection

Alexander Borisov; Victor Eruhimov; Eugene Tuv

Tree-based ensembles have been proven to be among the most accurate and versatile state-of-the-art learning machines. The best known are MART (gradient tree boosting) and RF (Random Forest.) Usage of such ensembles in supervised problems with a very high dimensional input space can be challenging. Modelling with MART becomes computationally infeasible, and RF can produce low quality models when only a small subset of predictors is relevant. We propose an importance based sampling scheme where only a small sample of variables is selected at every step of ensemble construction. The sampling distribution is modified at every iteration to promote variables more relevant to the target. Experiments show that this method gives MART a very substantial performance boost with at least the same level of accuracy. It also adds a bias correction element to RF for very noisy problems. MART with dynamic feature selection produced very competitive results at the NIPS-2003 feature selection challenge.

Part II - Feature Selection Challenge | Pp. 359-374

Sparse, Flexible and Efficient Modeling using Regularization

Saharon Rosset; Ji Zhu

We consider the generic regularized optimization problem () = arg min ∑(, )+λ(). We derive a general characterization of the properties of (loss , penalty ) pairs which give piecewise linear coefficient paths. Such pairs allow us to efficiently generate the full regularized coefficient paths.We illustrate how we can use our results to build robust, efficient and adaptable modeling tools.

Part II - Feature Selection Challenge | Pp. 375-394

Margin Based Feature Selection and Infogain with Standard Classifiers

Ran Gilad-Bachrach; Amir Navot

The decision to devote a week or two to playing with the feature selection challenge (FSC) turned into a major effort that took up most of our time a few months. In most cases we used standard algorithms, with obvious modifications for the balanced error measure. Surprisingly enough, the naïve methods we used turned out to be among the best submissions to the FSC.

Part II - Feature Selection Challenge | Pp. 395-401

Bayesian Support Vector Machines for Feature Ranking and Selection

Wei Chu; S. Sathiya Keerthi; Chong Jin Ong; Zoubin Ghahramani

In this chapter, we develop and evaluate a feature selection algorithm for Bayesian support vector machines. The relevance level of features are represented by ARD (automatic relevance determination) parameters, which are optimized by maximizing the model evidence in the Bayesian framework. The features are ranked in descending order using the optimal ARD values, and then forward selection is carried out to determine the minimal set of relevant features. In the numerical experiments, our approach using ARD for feature ranking can achieve a more compact feature set than standard ranking techniques, along with better generalization performance.

Part II - Feature Selection Challenge | Pp. 403-418

Nonlinear Feature Selection with the Potential Support Vector Machine

Sepp Hochreiter; Klaus Obermayer

We describe the “Potential Support Vector Machine” (P-SVM) which is a new filter method for feature selection. The idea of the P-SVM feature selection is to exchange the role of features and data points in order to construct “support features”. The “support features” are the selected features. The P-SVM uses a novel objective function and novel constraints — one constraint for each feature. As with standard SVMs, the objective function represents a complexity or capacity measure whereas the constraints enforce low empirical error. In this contribution we extend the P-SVM in two directions. First, we introduce a parameter which controls the redundancy among the selected features. Secondly, we propose a nonlinear version of the P-SVM feature selection which is based on neural network techniques. Finally, the linear and nonlinear P-SVM feature selection approach is demonstrated on toy data sets and on data sets from the NIPS 2003 feature selection challenge.

Part II - Feature Selection Challenge | Pp. 419-438