Cookies Policy
The website need some cookies and similar means to function. If you permit us, we will use those means to collect data on your visits for aggregated statistics to improve our service. Find out More
Accept Reject
  • Menu
Publications

Publications by Carlos Manuel Soares

2011

Combining Meta-learning and Active Selection of Datasetoids for Algorithm Selection

Authors
Prudencio, RBC; Soares, C; Ludermir, TB;

Publication
HYBRID ARTIFICIAL INTELLIGENT SYSTEMS, PART I

Abstract
Several meta-learning approaches have been developed for the problem of algorithm selection. In this context, it is of central importance to collect a sufficient number of datasets to be used as meta-examples in order to provide reliable results. Recently, some proposals to generate datasets have addressed this issue with successful results. These proposals include datasetoids, which is a simple manipulation method to obtain new datasets from existing ones. However, the increase in the number of datasets raises another issue: in order to generate meta-examples for training, it is necessary to estimate the performance of the algorithms on the datasets. This typically requires running all candidate algorithms on all datasets, which is computationally very expensive. One approach to address this problem is the use of active learning, termed active meta-learning. In this paper we investigate the combined use of active meta-learning and datasetoids. Our results show that it is possible to significantly reduce the computational cost of generating meta-examples not only without loss of meta-learning accuracy but with potential gains.

2009

Metalearning - Applications to Data Mining

Authors
Brazdil, P; Giraud Carrier, CG; Soares, C; Vilalta, R;

Publication
Cognitive Technologies

Abstract

2003

Ranking learning algorithms: Using IBL and meta-learning on accuracy and time results

Authors
Brazdil, PB; Soares, C; Da Costa, JP;

Publication
MACHINE LEARNING

Abstract
We present a meta-learning method to support selection of candidate learning algorithms. It uses a k-Nearest Neighbor algorithm to identify the datasets that are most similar to the one at hand. The distance between datasets is assessed using a relatively small set of data characteristics, which was selected to represent properties that affect algorithm performance. The performance of the candidate algorithms on those datasets is used to generate a recommendation to the user in the form of a ranking. The performance is assessed using a multicriteria evaluation measure that takes not only accuracy, but also time into account. As it is not common in Machine Learning to work with rankings, we had to identify and adapt existing statistical techniques to devise an appropriate evaluation methodology. Using that methodology, we show that the meta-learning method presented leads to significantly better rankings than the baseline ranking method. The evaluation methodology is general and can be adapted to other ranking problems. Although here we have concentrated on ranking classification algorithms, the meta-learning framework presented can provide assistance in the selection of combinations of methods or more complex problem solving strategies.

2007

Rejoinder to letter to the editor from C. Genest and J-F. Plante concerning 'Pinto da Costa, J. & Soares, C. (2005) A weighted rank measure of correlation.'

Authors
da Costa, JP; Soares, C;

Publication
AUSTRALIAN & NEW ZEALAND JOURNAL OF STATISTICS

Abstract

2000

Zoomed anking: Selection of Classification Algorithms Based on Relevant Performance Information

Authors
Soares, C; Brazdil, PB;

Publication
LECTURE NOTES IN COMPUTER SCIENCE <D>

Abstract
Given the wide variety of available classification algorithms and the volume of data today's organizations need to analyze, the selection of the right algorithm to use on a new problem is an important issue. In this paper we present a combination of techniques to address this problem. The first one, zooming, analyzes a given dataset and selects relevant (similar) datasets that were processed by the candidate algoritms in the past. This process is based on the concept of distance, calculated on the basis of several dataset characteristics. The information about the performance of the candidate algorithms on the selected datasets is then processed by a second technique, a ranking method. Such a method uses performance information to generate advice in the form of a ranking, indicating which algorithms should be applied in which order. Here we propose the adjusted ratio of ratios ranking method. This method takes into account not only accuracy but also the time performance of the candidate algorithms. The generalization power of this ranking method is analyzed. For this purpose, an appropriate methodology is defined. The experimental results indicate that on average better results are obtained with zooming than without it.

2012

Combining meta-learning and optimization algorithms for parameter selection

Authors
Gomes, T; Miranda, P; Prudencio, R; Soares, C; Carvalho, A;

Publication
CEUR Workshop Proceedings

Abstract
In this article we investigate the combination of meta-learning and optimization algorithms for parameter selection. We discuss our general proposal as well as present the recent develop-ments and experiments performed using Support Vector Machines (SVMs). Meta-learning was combined to single and multi-objective optimization techniques to select SVM parameters. The hybrid meth-ods derived from the proposal presented better results on predictive accuracy than the use of traditional optimization techniques.

  • 31
  • 46