Loading...
Please wait, while we are loading the content...
Similar Documents
Dynamic multi-armed bandits and extreme value-based rewards for adaptive operator selection in evolutionary algorithms (2009).
| Content Provider | CiteSeerX |
|---|---|
| Author | Fialho, Álvaro Costa, Luis Da Schoenauer, Marc Sebag, Michèle |
| Abstract | Abstract. The performance of many efficient algorithms critically depends on the tuning of their parameters, which on turn depends on the problem at hand. For example, the performance of Evolutionary Algorithms critically depends on the judicious setting of the operator rates. The Adaptive Operator Selection (AOS) heuristic that is proposed here rewards each operator based on the extreme value of the fitness improvement lately incurred by this operator, and uses a Multi-Armed Bandit (MAB) selection process based on those rewards to choose which operator to apply next. This Extreme-based Multi-Armed Bandit approach is experimentally validated against the Average-based MAB method, and is shown to outperform previously published methods, whether using a classical Average-based rewarding technique or the same Extreme-based mechanism. The validation test suite includes the easy One-Max problem and a family of hard problems known as “Long k-paths”. 1 |
| File Format | |
| Publisher Date | 2009-01-01 |
| Access Restriction | Open |
| Subject Keyword | Adaptive Operator Selection Evolutionary Algorithm Extreme Value-based Reward Dynamic Multi-armed Bandit Validation Test Suite Judicious Setting Long K-paths Hard Problem Fitness Improvement Easy One-max Problem Operator Rate Selection Process Extreme-based Mechanism Extreme-based Multi-armed Bandit Approach Classical Average-based Rewarding Technique Many Efficient Algorithm Average-based Mab Method Extreme Value Multi-armed Bandit |
| Content Type | Text |
| Resource Type | Article |