Catálogo de publicaciones - libros

Compartir en
redes sociales


Algorithmic Learning Theory: 17th International Conference, ALT 2006, Barcelona, Spain, October 7-10, 2006, Proceedings

José L. Balcázar ; Philip M. Long ; Frank Stephan (eds.)

En conferencia: 17º International Conference on Algorithmic Learning Theory (ALT) . Barcelona, Spain . October 7, 2006 - October 10, 2006

Resumen/Descripción – provisto por la editorial

No disponible.

Palabras clave – provistas por la editorial

Artificial Intelligence (incl. Robotics); Computation by Abstract Devices; Algorithm Analysis and Problem Complexity; Mathematical Logic and Formal Languages; Document Preparation and Text Processing

Disponibilidad
Institución detectada Año de publicación Navegá Descargá Solicitá
No detectada 2006 SpringerLink

Información

Tipo de recurso:

libros

ISBN impreso

978-3-540-46649-9

ISBN electrónico

978-3-540-46650-5

Editor responsable

Springer Nature

País de edición

Reino Unido

Fecha de publicación

Información sobre derechos de publicación

© Springer-Verlag Berlin Heidelberg 2006

Tabla de contenidos

General Discounting Versus Average Reward

Marcus Hutter

Consider an agent interacting with an environment in cycles. In every interaction cycle the agent is rewarded for its performance. We compare the average reward U from cycle 1 to m (average value) with the future discounted reward V from cycle k to ∞ (discounted value). We consider essentially arbitrary (non-geometric) discount sequences and arbitrary reward sequences (non-MDP environments). We show that asymptotically U for m →∞ and V for k →∞ are equal, provided both limits exist. Further, if the effective horizon grows linearly with k or faster, then the existence of the limit of U implies that the limit of V exists. Conversely, if the effective horizon grows linearly with k or slower, then existence of the limit of V implies that the limit of U exists.

Palabras clave: Average Reward; Bandit Problem; Discount Reward; Reward Sequence; Inconsistent Policy.

- Regular Contributions | Pp. 244-258

The Missing Consistency Theorem for Bayesian Learning: Stochastic Model Selection

Jan Poland

Bayes’ rule specifies how to obtain a posterior from a class of hypotheses endowed with a prior and the observed data. There are three principle ways to use this posterior for predicting the future: marginalization (integration over the hypotheses w.r.t. the posterior), MAP (taking the a posteriori most probable hypothesis), and stochastic model selection (selecting a hypothesis at random according to the posterior distribution). If the hypothesis class is countable and contains the data generating distribution, strong consistency theorems are known for the former two methods, asserting almost sure convergence of the predictions to the truth as well as loss bounds. We prove the first corresponding results for stochastic model selection. As a main technical tool, we will use the concept of a potential: this quantity, which is always positive, measures the total possible amount of future prediction errors. Precisely, in each time step, the expected potential decrease upper bounds the expected error. We introduce the entropy potential of a hypothesis class as its worst-case entropy with regard to the true distribution. We formulate our results in the online classification framework, but they are equally applicable to the prediction of non-i.i.d. sequences.

Palabras clave: Model Class; Current Input; Minimum Description Length; True Distribution; Kolmogorov Complexity.

- Regular Contributions | Pp. 259-273

Is There an Elegant Universal Theory of Prediction?

Shane Legg

Solomonoff’s inductive learning model is a powerful, universal and highly elegant theory of sequence prediction. Its critical flaw is that it is incomputable and thus cannot be used in practice. It is sometimes suggested that it may still be useful to help guide the development of very general and powerful theories of prediction which are computable. In this paper it is shown that although powerful algorithms exist, they are necessarily highly complex. This alone makes their theoretical analysis problematic, however it is further shown that beyond a moderate level of complexity the analysis runs into the deeper problem of Gödel incompleteness. This limits the power of mathematics to analyse and study prediction algorithms, and indeed intelligent systems in general.

Palabras clave: Prediction Algorithm; Kolmogorov Complexity; Input String; Computable Sequence; Short Program.

- Regular Contributions | Pp. 274-287

Learning Linearly Separable Languages

Leonid Kontorovich; Corinna Cortes; Mehryar Mohri

This paper presents a novel paradigm for learning languages that consists of mapping strings to an appropriate high-dimensional feature space and learning a separating hyperplane in that space. It initiates the study of the linear separability of automata and languages by examining the rich class of piecewise-testable languages. It introduces a high-dimensional feature map and proves piecewise-testable languages to be linearly separable in that space. The proof makes use of word combinatorial results relating to subsequences. It also shows that the positive definite kernel associated to this embedding can be computed in quadratic time. It examines the use of support vector machines in combination with this kernel to determine a separating hyperplane and the corresponding learning guarantees. It also proves that all languages linearly separable under a regular finite cover embedding, a generalization of the embedding we used, are regular.

Palabras clave: Support Vector Machine; Weight Vector; Boolean Function; Regular Language; Generalization Error.

- Regular Contributions | Pp. 288-303

Smooth Boosting Using an Information-Based Criterion

Kohei Hatano

Smooth boosting algorithms are variants of boosting methods which handle only smooth distributions on the data. They are proved to be noise-tolerant and can be used in the “boosting by filtering” scheme, which is suitable for learning over huge data. However, current smooth boosting algorithms have rooms for improvements: Among non-smooth boosting algorithms, real AdaBoost or InfoBoost, can perform more efficiently than typical boosting algorithms by using an information-based criterion for choosing hypotheses. In this paper, we propose a new smooth boosting algorithm with another information-based criterion based on Gini index. we show that it inherits the advantages of two approaches, smooth boosting and information-based approaches.

Palabras clave: Loss Function; Gini Index; Training Error; Huge Data; Weak Hypothesis.

- Regular Contributions | Pp. 304-318

Large-Margin Thresholded Ensembles for Ordinal Regression: Theory and Practice

Hsuan-Tien Lin; Ling Li

We propose a thresholded ensemble model for ordinal regression problems. The model consists of a weighted ensemble of confidence functions and an ordered vector of thresholds. We derive novel large-margin bounds of common error functions, such as the classification error and the absolute error. In addition to some existing algorithms, we also study two novel boosting approaches for constructing thresholded ensembles. Both our approaches not only are simpler than existing algorithms, but also have a stronger connection to the large-margin bounds. In addition, they have comparable performance to SVM-based algorithms, but enjoy the benefit of faster training. Experimental results on benchmark datasets demonstrate the usefulness of our boosting approaches.

Palabras clave: Sigmoid Function; Benchmark Dataset; Base Learner; Ordinal Regression; Decision Stump.

- Regular Contributions | Pp. 319-333

Asymptotic Learnability of Reinforcement Problems with Arbitrary Dependence

Daniil Ryabko; Marcus Hutter

We address the problem of reinforcement learning in which observations may exhibit an arbitrary form of stochastic dependence on past observations and actions, i.e. environments more general than (PO) MDPs. The task for an agent is to attain the best possible asymptotic reward where the true generating environment is unknown but belongs to a known countable family of environments. We find some sufficient conditions on the class of environments under which an agent exists which attains the best asymptotic reward for any environment in the class. We analyze how tight these conditions are and how they relate to different probabilistic assumptions known in reinforcement learning and related fields, such as Markov Decision Processes and mixing conditions.

Palabras clave: Optimal Policy; Reinforcement Learning; Markov Decision Process; Countable Family; Average Reward.

- Regular Contributions | Pp. 334-347

Probabilistic Generalization of Simple Grammars and Its Application to Reinforcement Learning

Takeshi Shibata; Ryo Yoshinaka; Takashi Chikayama

Recently, some non-regular subclasses of context-free grammars have been found to be efficiently learnable from positive data. In order to use these efficient algorithms to infer probabilistic languages, one must take into account not only equivalences between languages but also probabilistic generalities of grammars. The probabilistic generality of a grammar G is the class of the probabilistic languages generated by probabilistic grammars constructed on G . We introduce a subclass of simple grammars (SGs), referred as to unifiable simple grammars (USGs), which is a superclass of an efficiently learnable class, right-unique simple grammars (RSGs). We show that the class of RSGs is unifiable within the class of USGs, whereas SGs and RSGs are not unifiable within the class of SGs and RSGs, respectively. We also introduce simple context-free decision processes, which are a natural extension of finite Markov decision processes and intuitively may be thought of a Markov decision process with stacks. We propose a reinforcement learning method on simple context-free decision processes, as an application of the learning and unification algorithm for RSGs from positive data.

- Regular Contributions | Pp. 348-362

Unsupervised Slow Subspace-Learning from Stationary Processes

Andreas Maurer

We propose a method of unsupervised learning from stationary, vector-valued processes. A low-dimensional subspace is selected on the basis of a criterion which rewards data-variance (like PSA) and penalizes the variance of the velocity vector, thus exploiting the short-time dependencies of the process. We prove error bounds in terms of the β -mixing coefficients and consistency for absolutely regular processes. Experiments with image recognition demonstrate the algorithms ability to learn geometrically invariant feature maps.

Palabras clave: Kernel Principal Component Analysis; Stationary Stochastic Process; Batch Algorithm; Slow Feature Analysis; Minimal Subspace.

- Regular Contributions | Pp. 363-377

Learning-Related Complexity of Linear Ranking Functions

Atsuyoshi Nakamura

In this paper, we study learning-related complexity of linear ranking functions from n -dimensional Euclidean space to {1,2,..., k }. We show that their graph dimension , a kind of measure for PAC learning complexity in the multiclass classification setting, is Θ( n + k ). This graph dimension is significantly smaller than the graph dimension Ω( nk ) of the class of {1,2,..., k }-valued decision-list functions naturally defined using k –1 linear discrimination functions. We also show a risk bound of learning linear ranking functions in the ordinal regression setting by a technique similar to that used in the proof of an upper bound of their graph dimension.

Palabras clave: Function Class; Collaborative Filter; Ordinal Regression; Graph Dimension; Decision List.

- Regular Contributions | Pp. 378-392