Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.5555/3041838guideproceedingsBook PagePublication PagesConference Proceedingsacm-pubtype
ICML'03: Proceedings of the Twentieth International Conference on International Conference on Machine Learning
2003 Proceeding
Publisher:
  • AAAI Press
Conference:
Washington, DC USA August 21 - 24, 2003
ISBN:
978-1-57735-189-4
Published:
21 August 2003
Sponsors:
Kluwer Academic Publishers, NSF, Kaidara Software, AAAI, Microsoft Research, HP

Bibliometrics
Abstract

No abstract available.

Proceeding Downloads

Article
Hidden Markov support vector machines
Pages 3–10

This paper presents a novel discriminative learning technique for label sequences based on a combination of the two most successful learning algorithms, Support Vector Machines and Hidden Markov Models which we call Hidden Markov Support Vector Machine. ...

Article
Learning distance functions using equivalence relations
Pages 11–18

We address the problem of learning distance metrics using side-information in the form of groups of "similar" points. We propose to use the RCA algorithm, which is a simple and efficient algorithm for learning a full ranked Mahalanobis metric (Shental ...

Article
Online choice of active learning algorithms
Pages 19–26

This paper is concerned with the question of how to online combine an ensemble of active learners so as to expedite the learning progress during a pool-based active learning session. We develop a powerful active learning master algorithm, based a known ...

Article
Learning logic programs for layout analysis correction
Pages 27–34

Layout analysis is the process of extracting a hierarchical structure describing the layout of a page. In the system WISDOM++, the layout analysis is performed in two steps: firstly, the global analysis determines possible areas containing paragraphs, ...

Article
Multi-objective programming in SVMs
Pages 35–42

We propose a general framework for support vector machines (SVM) based on the principle of multi-objective optimization. The learning of SVMs is formulated as a multi-objective program by setting two competing goals to minimize the empirical risk and ...

Article
Regression error characteristic curves
Pages 43–50

Receiver Operating Characteristic (ROC) curves provide a powerful tool for visualizing and comparing classification results. Regression Error Characteristic (REC) curves generalize ROC curves to regression. REC curves plot the error tolerance on the x-...

Article
Choosing between two learning algorithms based on calibrated tests
Pages 51–58

Designing a hypothesis test to determine the best of two machine learning algorithms with only a small data set available is not a simple task. Many popular tests suffer from low power (5×2 cv [2]), or high Type I error (Weka's 10×10 cross validation [...

Article
Incorporating diversity in active learning with support vector machines
Pages 59–66

In many real world applications, active selection of training examples can significantly reduce the number of labelled training examples to learn a classification function. Different strategies in the field of support vector machines have been proposed ...

Article
The use of the ambiguity decomposition in neural network ensemble learning methods
Pages 67–74

We analyze the formal grounding behind Negative Correlation (NC) Learning, an ensemble learning technique developed in the evolutionary computation literature. We show that by removing an assumption made in the original work, NC can be seen to be ...

Article
Tractable Bayesian learning of tree augmented Naive Bayes models
Pages 75–82

Bayesian classifiers such as Naive Bayes or Tree Augmented Naive Bayes (TAN) have shown excellent performance given their simplicity and heavy underlying independence assumptions. In this paper we introduce a classifier taking as basis the TAN model and ...

Article
AWESOME: a general multiagent learning algorithm that converges in self-play and learns a best response against stationary opponents
Pages 83–90

A satisfactory multiagent learning algorithm should, at a minimum, learn to play optimally against stationary opponents and converge to a Nash equilibrium in self-play. The algorithm that has come closest, WoLF-IGA, has been proven to have these two ...

Article
BL-WoLF: a framework for loss-bounded learnability in zero-sum games
Pages 91–98

We present BL-WoLF, a framework for learnability in repeated zero-sum games where the cost of learning is measured by the losses the learning agent accrues (rather than the number of rounds). The game is adversarially chosen from some family that the ...

Article
Semi-supervised learning of mixture models
Pages 99–106

This paper analyzes the performance of semi-supervised learning of mixture models. We show that unlabeled data can lead to an increase in classification error even in situations where additional labeled data would decrease classification error. We ...

Article
On kernel methods for relational learning
Pages 107–114

Kernel methods have gained a great deal of popularity in the machine learning community as a method to learn indirectly in high-dimensional feature spaces. Those interested in relational learning have recently begun to cast learning from structured and ...

Article
Fast query-optimized kernel machine classification via incremental approximate nearest support vectors
Pages 115–122

Support vector machines (and other kernel machines) offer robust modern machine learning methods for nonlinear classification. However, relative to other alternatives (such as linear methods, decision trees and neural networks), they can be orders of ...

Article
Relational instance based regression for relational reinforcement learning
Pages 123–130

Relational reinforcement learning (RRL) is a Q-learning technique which uses first order regression techniques to generalize the Q-function. Both the relational setting and the Q-learning context introduce a number of difficulties which must be dealt ...

Article
Design for an optimal probe
Pages 131–138

Given a Markov decision process (MDP) with expressed prior uncertainties in the process transition probabilities, we consider the problem of computing a policy that optimizes expected total (finite-horizon) reward. Implicitly, such a policy would ...

Article
Diffusion approximation for Bayesian Markov chains
Pages 139–147

Given a Markov chain with uncertain transition probabilities modelled in a Bayesian way, we investigate a technique for analytically approximating the mean transition frequency counts over a finite horizon. Conventional techniques for addressing this ...

Article
Using the triangle inequality to accelerate k-means
Pages 147–153

The k-means algorithm is by far the most widely used method for discovering clusters in data. We show how to accelerate it dramatically, while still always computing exactly the same result as the standard algorithm. The accelerated algorithm avoids ...

Article
Bayes meets bellman: the Gaussian process approach to temporal difference learning
Pages 154–161

We present a novel Bayesian approach to the problem of value function estimation in continuous state spaces. We define a probabilistic generative model for the value function by imposing a Gaussian prior over value functions and assuming a Gaussian ...

Article
Action elimination and stopping conditions for reinforcement learning
Pages 162–169

We consider incorporating action elimination procedures in reinforcement learning algorithms. We suggest a framework that is based on learning an upper and a lower estimates of the value function or the Q-function and eliminating actions that are not ...

Article
Utilizing domain knowledge in neuroevolution
Pages 170–177

We propose a method called Rule-based ESP (RESP) for utilizing prior knowledge evolving Artificial Neural Networks (ANNs). First, KBANN-likete chniques are used to transform a set of rules into an ANN, then the ANN is trained using the Enforced ...

Article
Boosting lazy decision trees
Pages 178–185

This paper explores the problem of how to construct lazy decision tree ensembles. We present and empirically evaluate a relevance-based boosting-style algorithm that builds a lazy decision tree ensemble customized for each test instance. From the ...

Article
Random projection for high dimensional data clustering: a cluster ensemble approach
Pages 186–193

We investigate how random projection can best be used for clustering high dimensional data. Random projection has been shown to have promising theoretical properties. In practice, however, we find that it results in highly unstable clustering ...

Article
The geometry of ROC space: understanding machine learning metrics through ROC isometrics
Pages 194–201

Many different metrics are used in machine learning and data mining to build and evaluate models. However, there is no general theory of machine learning metrics, that could answer questions such as: When we simultaneously want to optimise two criteria, ...

Article
An analysis of rule evaluation metrics
Pages 202–209

In this paper we analyze the most popular evaluation metrics for separate-and-conquer rule learning algorithms. Our results show that all commonly used heuristics, including accuracy, weighted relative accuracy, entropy, Gini index and information gain, ...

Article
Margin distribution and learning algorithms
Pages 210–217

Recent theoretical results have shown that improved bounds on generalization error of classifiers can be obtained by explicitly taking the observed margin distribution of the training data into account. Currently, algorithms used in practice do not make ...

Article
Perceptron based learning with example dependent and noisy costs
Pages 218–225

Learning algorithms from the fields of artificial neural networks and machine learning, typically, do not take any costs into account or allow only costs depending on the classes of the examples that are used for learning. As an extension of class ...

Article
Hierarchical policy gradient algorithms
Pages 226–233

Hierarchical reinforcement learning is a general framework which attempts to accelerate policy learning in large domains. On the other hand, policy gradient reinforcement learning (PGRL) methods have received recent attention as a means to solve ...

Article
Solving noisy linear operator equations by Gaussian processes: application to ordinary and partial differential equations
Pages 234–241

We formulate the problem of solving stochastic linear operator equations in a Bayesian Ganssian process (GP) framework. The solution is obtained in the spirit of a collocation method based on noisy evaluations of the target function at randomly drawn or ...

Contributors
  • Stanford University
  • Amazon.com, Inc.

Recommendations