par Gagliolo, Matteo ;Schmidhuber, Juergen
Référence Lecture Notes in Computer Science, 3697, page (7-12)
Publication Publié, 2005
Article révisé par les pairs
Résumé : One aim of Meta-learning techniques is to minimize the time needed for problem solving, and the effort of parameter hand-tuning, by automating algorithm selection. The predictive model of algorithm performance needed for this task often requires long training times. We address the problem in an online fashion, running multiple algorithms in parallel on a sequence of tasks, continually updating their relative priorities according to a neural model that maps their current state to the expected time to the solution. The model itself is updated at the end of each task, based on the actual performance of each algorithm. Censored sampling allows us to train the model effectively, without need of additional exploration after each task's solution. We present a preliminary experiment in which this new inter-problem technique learns to outperform a previously proposed intra-problem heuristic.