- Timestamp:
- 01/13/15 20:02:29 (10 years ago)
- File:
-
- 1 edited
Legend:
- Unmodified
- Added
- Removed
-
branches/HeuristicLab.Problems.GrammaticalOptimization/Main/Program.cs
r11747 r11755 146 146 // TODO: wie kann ich sampler noch vergleichen bzw. was kann man messen um die qualität des samplers abzuschätzen (bis auf qualität und iterationen bis zur besten lösung) => ziel schnellere iterationen zu gutem ergebnis 147 147 // TODO: research thompson sampling for max bandit? 148 // TODO: ausführlicher test von strategien für k-armed max bandit148 // TODO: ausführlicher test von strategien für numCorrectPhrases-armed max bandit 149 149 // TODO: verify TA implementation using example from the original paper 150 150 // TODO: separate policy from MCTS tree data structure to allow sharing of information over disconnected parts of the tree (semantic equivalence) … … 166 166 var random = new Random(); 167 167 168 var phraseLen = 1; 169 var sentenceLen = 25; 170 var numPhrases = sentenceLen / phraseLen; 171 var problem = new RoyalPhraseSequenceProblem(random, 10, numPhrases, phraseLen: 1, k: 1, correctReward: 1, incorrectReward: 0); 172 173 //var problem = new SymbolicRegressionPoly10Problem(); // good results e.g. 10 randomtries and EpsGreedyPolicy(0.2, (aInfo)=>aInfo.MaxReward) 168 //var phraseLen = 3; 169 //var numPhrases = 5; 170 //var problem = new RoyalPhraseSequenceProblem(random, 10, numPhrases, phraseLen: phraseLen, numCorrectPhrases: 1, correctReward: 1, incorrectReward: 0.0, phrasesAsSets: true); 171 172 //var phraseLen = 4; 173 //var numPhrases = 5; 174 //var problem = new FindPhrasesProblem(random, 15, numPhrases, phraseLen, numOptimalPhrases: numPhrases, numDecoyPhrases: 500, correctReward: 1.0, decoyReward: 0.2, phrasesAsSets: true); 175 176 var problem = new SymbolicRegressionPoly10Problem(); // good results e.g. 10 randomtries and EpsGreedyPolicy(0.2, (aInfo)=>aInfo.MaxReward) 174 177 // Ant 175 178 // good results e.g. with var alg = new MctsSampler(problem, 17, random, 1, (rand, numActions) => new ThresholdAscentPolicy(numActions, 500, 0.01)); … … 182 185 //var problem = new EvenParityProblem(); 183 186 // symbreg length = 11 q = 0.824522210419616 184 var alg = new MctsSampler(problem, sentenceLen, random, 0, new BoltzmannExplorationPolicy(200)); 187 //var alg = new MctsSampler(problem, 23, random, 0, new BoltzmannExplorationPolicy(100)); 188 var alg = new MctsSampler(problem, 23, random, 0, new EpsGreedyPolicy(0.1)); 185 189 //var alg = new MctsQLearningSampler(problem, sentenceLen, random, 0, null); 186 190 //var alg = new MctsQLearningSampler(problem, 30, random, 0, new EpsGreedyPolicy(0.2));
Note: See TracChangeset
for help on using the changeset viewer.