Catálogo de publicaciones - libros
Feature Extraction: Foundations and Applications
Isabelle Guyon ; Masoud Nikravesh ; Steve Gunn ; Lotfi A. Zadeh (eds.)
Resumen/Descripción – provisto por la editorial
No disponible.
Palabras clave – provistas por la editorial
No disponibles.
Disponibilidad
Institución detectada | Año de publicación | Navegá | Descargá | Solicitá |
---|---|---|---|---|
No detectada | 2006 | SpringerLink |
Información
Tipo de recurso:
libros
ISBN impreso
978-3-540-35487-1
ISBN electrónico
978-3-540-35488-8
Editor responsable
Springer Nature
País de edición
Reino Unido
Fecha de publicación
2006
Información sobre derechos de publicación
© Springer Berlin Heidelberg 2006
Tabla de contenidos
High Dimensional Classification with Bayesian Neural Networks and Dirichlet Diffusion Trees
Radford M. Neal; Jianguo Zhang
Our winning entry in the NIPS 2003 challenge was a hybrid, in which our predictions for the five data sets were made using different methods of classification, or, for the Madelon data set, by averaging the predictions produced using two methods. However, two aspects of our approach were the same for all data sets:
Part II - Feature Selection Challenge | Pp. 265-296
Ensembles of Regularized Least Squares Classifiers for High-Dimensional Problems
Kari Torkkola; Eugene Tuv
It has been recently pointed out that the Regularized Least Squares Classifier (RLSC), continues to be a viable option for binary classification problems. We apply RLSC to the datasets of the NIPS 2003 Feature Selection Challenge using Gaussian kernels. Since RLSC is sensitive to noise variables, ensemble-based variable filtering is applied first. RLSC makes use of the best-ranked variables only. We compare the performance of a stochastic ensemble of RLSCs to a single best RLSC. Our results indicate that in terms of classification error rate the two are similar on the challenge data. However, especially with large data sets, ensembles could provide other advantages that we list.
Part II - Feature Selection Challenge | Pp. 297-313
Combining SVMs with Various Feature Selection Strategies
Yi-Wei Chen; Chih-Jen Lin
This article investigates the performance of combining support vector machines (SVM) and various feature selection strategies. Some of them are filter-type approaches: general feature selection methods independent of SVM, and some are wrapper-type methods: modifications of SVM which can be used to select features. We apply these strategies while participating to the NIPS 2003 Feature Selection Challenge and rank third as a group.
Part II - Feature Selection Challenge | Pp. 315-324
Feature Selection with Transductive Support Vector Machines
Zhili Wu; Chunhung Li
SVM-related feature selection has shown to be effective, while feature selection with transductive SVMs has been less studied. This paper investigates the use of transductive SVMs for feature selection, based on three SVM-related feature selection methods: filtering scores + SVM wrapper, recursive feature elimination (RFE) and multiplicative updates(MU). We show transductive SVMs can be tailored to feature selection by embracing feature scores for feature filtering, or acting as wrappers and embedded feature selectors. We conduct experiments on the feature selection competition tasks to demonstrate the performance of Transductive SVMs in feature selection and classification.
Part II - Feature Selection Challenge | Pp. 325-341
Variable Selection using Correlation and Single Variable Classifier Methods: Applications
Amir Reza Saffari Azar Alamdari
Correlation and single variable classifier methods are very simple algorithms to select a subset of variables in a dimension reduction problem, which utilize some measures to detect relevancy of a single variable to the target classes without considering the predictor properties to be used. In this paper, along with the description of correlation and single variable classifier ranking methods, the application of these algorithms to the NIPS 2003 Feature Selection Challenge problems is also presented. The results show that these methods can be used as one of primary, computational cost efficient, and easy to implement techniques which have good performance especially when variable space is very large. Also, it has been shown that in all cases using an ensemble averaging predictor would result in a better performance, compared to a single stand-alone predictor.
Part II - Feature Selection Challenge | Pp. 343-358
Tree-Based Ensembles with Dynamic Soft Feature Selection
Alexander Borisov; Victor Eruhimov; Eugene Tuv
Tree-based ensembles have been proven to be among the most accurate and versatile state-of-the-art learning machines. The best known are MART (gradient tree boosting) and RF (Random Forest.) Usage of such ensembles in supervised problems with a very high dimensional input space can be challenging. Modelling with MART becomes computationally infeasible, and RF can produce low quality models when only a small subset of predictors is relevant. We propose an importance based sampling scheme where only a small sample of variables is selected at every step of ensemble construction. The sampling distribution is modified at every iteration to promote variables more relevant to the target. Experiments show that this method gives MART a very substantial performance boost with at least the same level of accuracy. It also adds a bias correction element to RF for very noisy problems. MART with dynamic feature selection produced very competitive results at the NIPS-2003 feature selection challenge.
Part II - Feature Selection Challenge | Pp. 359-374
Sparse, Flexible and Efficient Modeling using Regularization
Saharon Rosset; Ji Zhu
We consider the generic regularized optimization problem () = arg min ∑(, )+λ(). We derive a general characterization of the properties of (loss , penalty ) pairs which give piecewise linear coefficient paths. Such pairs allow us to efficiently generate the full regularized coefficient paths.We illustrate how we can use our results to build robust, efficient and adaptable modeling tools.
Part II - Feature Selection Challenge | Pp. 375-394
Margin Based Feature Selection and Infogain with Standard Classifiers
Ran Gilad-Bachrach; Amir Navot
The decision to devote a week or two to playing with the feature selection challenge (FSC) turned into a major effort that took up most of our time a few months. In most cases we used standard algorithms, with obvious modifications for the balanced error measure. Surprisingly enough, the naïve methods we used turned out to be among the best submissions to the FSC.
Part II - Feature Selection Challenge | Pp. 395-401
Bayesian Support Vector Machines for Feature Ranking and Selection
Wei Chu; S. Sathiya Keerthi; Chong Jin Ong; Zoubin Ghahramani
In this chapter, we develop and evaluate a feature selection algorithm for Bayesian support vector machines. The relevance level of features are represented by ARD (automatic relevance determination) parameters, which are optimized by maximizing the model evidence in the Bayesian framework. The features are ranked in descending order using the optimal ARD values, and then forward selection is carried out to determine the minimal set of relevant features. In the numerical experiments, our approach using ARD for feature ranking can achieve a more compact feature set than standard ranking techniques, along with better generalization performance.
Part II - Feature Selection Challenge | Pp. 403-418
Nonlinear Feature Selection with the Potential Support Vector Machine
Sepp Hochreiter; Klaus Obermayer
We describe the “Potential Support Vector Machine” (P-SVM) which is a new filter method for feature selection. The idea of the P-SVM feature selection is to exchange the role of features and data points in order to construct “support features”. The “support features” are the selected features. The P-SVM uses a novel objective function and novel constraints — one constraint for each feature. As with standard SVMs, the objective function represents a complexity or capacity measure whereas the constraints enforce low empirical error. In this contribution we extend the P-SVM in two directions. First, we introduce a parameter which controls the redundancy among the selected features. Secondly, we propose a nonlinear version of the P-SVM feature selection which is based on neural network techniques. Finally, the linear and nonlinear P-SVM feature selection approach is demonstrated on toy data sets and on data sets from the NIPS 2003 feature selection challenge.
Part II - Feature Selection Challenge | Pp. 419-438