Civil Engineering Reference
In-Depth Information
Nevmyvaka, Y., Feng, Y., & Kearns, M. (2006). Reinforcement learning for optimized trade execu-
tion. In Cohen, W. W. and Moore, A. (Eds.), Proceedings of the 23rd International Conference
on Machine Learning (ICML), Pittsburgh, PA, 25-29 June (pp. 673-680). New York, NY: ACM.
Ng, A. Y., Coates, A., Diel, M., Ganapathi, V., Schulte, J., Tse, B., Berger, E. & Liang, E. (2004).
Autonomous inverted helicopter flight via reinforcement learning. In International Symposium
on Experimental Robotics (ISER-2004), Singapore, 18-21 June (pp. 363-372). Cambridge,
MA: MIT Press.
Nissen, S. (2007). Large scale reinforcement learning using Q-Sarsa(ʻ) and cascading neu-
ral networks . Unpublished masters thesis, Department of Computer Science, University of
Copenhagen, København, Denmark.
Niv, Y. (2009). Reinforcement learning in the brain. Journal of Mathematical Psychology , 53(3),
139-154.
Ollington, R. B., Vamplew, P. H., & Swanson, J. (2009). Incorporating expert advice into reinforce-
ment learning using constructive neural networks. In Franco, L., Elizondo, D. A., & Jerez, J.
M. (Eds.), Constructive Neural Networks (pp. 207-224). Berlin: Springer.
Orr,
M.
J.
L.
(1996).
Introduction
to
radial
basis
function
networks
(Technical
re-
port,
Centre
For
Cognitive
Science,
University
of
Edinburgh).
Retrieved
from
http://www.cc.gatech.edu/ isbell/tutorials/rbf-intro.pdf.
Osana, Y. (2011). Reinforcement learning using Kohonen feature map probabilistic associative
memory based on weights distribution. In Mellouk, A. (Ed.), Advances in Reinforcement
Learning (pp. 121-136). InTech.
Osentoski, S. (2009). Action-based representation discovery in Markov decision processes .
Unpublished PhD dissertation, University of Massachusetts, Amherst, MA.
Papahristou, N. & Refanidis, I. (2011). Training neural networks to play backgammon variants
using reinforcement learning. In Applications of Evolutionary Computation, Proceedings of
the 11th International Conference on Applications of Evolutionary Computation, Torino, Italy,
27-29 April (pp. 113-122). Berlin: Springer-Verlag.
Papavassiliou, V. A. & Russell, S. (1999). Convergence of reinforcement learning with general
function approximators. In Proceedings of the 17th International Joint Conference on Artificial
Intelligence (IJCAI), Stockholm, Sweden, 31 July-6 August (Vol. 2, pp. 748-755). San Francisco,
CA: Morgan Kaufmann.
Papierok, S., Noglik, A., & Pauli, J. (2008). Application of reinforcement learning in a real environ-
ment using an RBF network. In 1st International Workshop on Evolutionary and Reinforcement
Learning for Autonomous Robot Systems (ERLARS), Patras, Greece, 22 July (pp. 17-22).
Retrieved from http://www.is.uni-due.de/fileadmin/literatur/publikation/papierok08erlars.pdf
Patist, J. P. & Wiering, M. (2004). Learning to play draughts using temporal difference learning
with neural networks and databases. In Proceedings of the 13th Belgian-Dutch Conference
on Machine Learning, Brussels, Belgium, 8-9 January (pp. 87-94). doi: 10.1007/978-3-540-
88190-2_13
Peters, J. & Schaal, S. (2006). Policy gradient methods for robotics. In IEEE/RSJ International
Conference on Intelligent Robots and Systems (IROS), Beijing, China, 9-15 October (pp. 2219-
2225). doi: 10.1109/IROS.2006.282564
Peters, J. & Schaal, S. (2009). Reinforcement learning of motor skills with policy gradients. Neural
Networks , 21(4), 682-697.
Pollack, J. B. & Blair, A. D. (1996). Why did TD-Gammon work? In Mozer, M. C., Jordan, M. I.,
& Petsche, T. (Eds.), Advances in Neural Information Processing Systems 9 . Cambridge, MA:
MIT Press.
Pontrandolfo, P., Gosavi, A., Okogbaa, O. G., & Das, T. K. (2002). Global supply chain manage-
ment: A reinforcement learning approach. International Journal of Production Research , 40(6),
1299-1317.
Powell, W. B. (2007). Approximate Dynamic Programming: Solving the Curse of Dimensionality .
New York, NY: John Wiley & Sons.
Search WWH ::




Custom Search