Civil Engineering Reference
In-Depth Information
Chen, X., Ankenman, B. E., & Nelson, B. L. (2012). The effects of Common Random Numbers on
stochastic kriging metamodeling. ACM Transactions on Modeling and Computer Simulation ,
22(2). doi: 10.1145/2133390.2133391
Chen, X., Ankenman, B. E., & Nelson, B. L. (2013). Enhancing stochastic kriging metamodels
with gradient estimators. Operations Research , 61(2), 512-528.
Dann, C., Neumann, G., & Peters, J. (2014). Policy evaluation with temporal differences: A survey
and comparison. Journal of Machine Learning Research , 15(1), 809-883.
Diesenroth, M. P., Neumann, G., & Peters, J. (2011). A survey on policy search for robotics.
Foundations and Trends in Robotics , 2(1-2), 1-142.
Embrechts, M. J., Hargis, B. J., & Linton, J. D. (2010). An augmented efficient backpropagation
training strategy for deep autoassociative neural networks. In Proceedings of the 2010 Interna-
tional Joint Conference on Neural Networks (IJCNN), Barcelona, Spain, 18-23 July (pp. 1-6).
doi: 10.1109/IJCNN.2010. 5596828
Gatti, C. J. & Embrechts, M. J. (2014). An application of the temporal difference algorithm to the
truck backer-upper problem. In Proceedings of the 22nd European Symposium on Artificial Neu-
ral Networks, Computational Intelligence and Machine Learning (ESANN), Bruges, Belgium,
23-25 April . Bruges, Belgium: ESANN.
Gatti, C. J., Embrechts, M. J., & Linton, J. D. (2011a). Parameter settings of reinforcement learning
for the game of Chung Toi. In Proceedings of the 2011 IEEE International Conference on
Systems, Man, and Cybernetics (SMC 2011), Anchorage, AK, 9-12 October (pp. 3530-3535).
doi: 10.1109/ICSMC.2011.6084216
Gatti, C. J., Embrechts, M. J., & Linton, J. D. (2013). An empirical analysis of reinforcement
learning using design of experiments. In Proceedings of the 21st European Symposium on Arti-
ficial Neural Networks, Computational Intelligence and Machine Learning (ESANN), Bruges,
Belgium, 24-26 April (pp. 221-226). Bruges, Belgium: ESANN.
Ghory, I. (2004). Reinforcement learning in board games (Technical Report CSTR-04-004,
Department of Computer Science,
University of Bristol).
Retrieved from http://www.cs.
bris.ac.uk/Publications/Papers/2000100.pdf.
Huang, D., Allen, T. T., Notz, W. I., & Zeng, N. (2006). Global optimization of stochastic black-box
systems via sequential kriging meta-models. Journal of Global Optimization , 34(3), 441-466.
Jones, D. R., Schonlau, M., & Welch, W. J. (1998). Efficient global optimization of expensive
black-box functions. Journal of Global Optimization , 13(4), 455-492.
Kalyanakrishnan, S. & Stone, P. (2009). An empirical analysis of value function-based and policy
search reinforcement learning. In Proceedings of the 8th International Conference on Au-
tonomous Agents and Multiagent Systems (AAMAS '09), Budapest, Hungary, 10-15 May (Vol. 2,
pp. 749-756). Richland, SC: International Foundation for Autonomous Agents and Multiagent
Systems.
Kalyanakrishnan, S. & Stone, P. (2011). Characterizing reinforcement learning methods through
parameterized learning problems. Machine Learning , 84(1-2), 205-247.
Kliejnen, J. P. C. (2013). Simulation-optimization via kriging and bootstrapping: A survey
(Technical Report 2013-064, Tilburg University: CentER). Retrieved from https://pure.uvt.nl/
portal/files/1544115/2013-064.pdf.
LeCun, Y., Bottou, L., Orr, G., , & Müller, K. (1998). Efficient backprop. In Orr, G. & Müller, K.
(Eds.), Neural Networks: Tricks of the Trade , volume 1524 (pp. 5-50). Berlin: Springer.
Lin, L.-J. (1992). Self-improving reactive agents based on reinforcement learning, planning and
teaching. Machine Learning , 8(3-4), 293-321.
Loeppky, J. L., Sacks, J., & Welch, W. J. (2009). Choosing the sample size of a computer experiment:
A practical guide. Technometrics , 51(4), 366-376.
Patist, J. P. & Wiering, M. (2004). Learning to play draughts using temporal difference learning
with neural networks and databases. In Proceedings of the 13th Belgian-Dutch Conference
on Machine Learning, Brussels, Belgium, 8-9 January (pp. 87-94). doi: 10.1007/978-3-540-
88190-2_13
Search WWH ::




Custom Search