Information Technology Reference
In-Depth Information
7. C. Seiffert, T. Khoshgoftaar, J. V. Hulse, and A. Napolitano, “RUSBoost: A hybrid
approach to alleviating class imbalance,” IEEE Transactions on Systems, Man, and
Cybernetics - Part A: Systems and Humans , vol. 40, no. 1, pp. 185-197, 2010.
8. H. Guo and H. Viktor, “Learning from imbalanced data sets with boosting and data
generation: The Databoost-IM approach,” ACM SIGKDD Explorations Newsletter ,
vol. 6, no. 1, pp. 30-39, 2004.
9. L. Breiman, “Bagging predictors,” Machine Learning , vol. 24, no. 2, pp. 123-140,
1996.
10. Y. Freund and R. E. Schapire, “A decision-theoretic generalization of on-line learning
and an application to boosting,” Journal of Computer and System Sciences , vol. 55,
no. 1, pp. 119-139, 1997.
11. D. H. Wolpert, “Stacked generalization,” Neural Networks , vol. 5, no. 2, pp. 241-260,
1992.
12. L. Breiman, “Stacked regressions,” Machine Learning , vol. 24, no. 1, pp. 49-64,
1996.
13. P. Smyth and D. Wolper, “Stacked density estimation,” in Advances in Neural Infor-
mation Processing Systems 10 (M. I. Jordan, M. J. Kearns, and S. A. Solla, eds.), pp.
668-674, Cambridge, MA: MIT Press, 1998.
14. A. Krogh and J. Vedelsby, “Neural network ensembles, cross validation, and active
learning,” in Advances in Neural Information Processing Systems 7 (G. Tesauro,
D. S. Touretzky, and T. K. Leen, eds.), pp. 231-238, Cambridge, MA: MIT Press,
1995.
15. Z.-H. Zhou, Ensemble Methods: Foundations and Algorithms . Boca Raton, FL: Chap-
man and Hall/CRC Press, 2012.
16. B. Efron and R. Tibshirani, An Introduction to the Bootstrap . New York: Chapman
& Hall, 1993.
17. M. Kearns and L. G. Valiant, “Cryptographic limitations on learning Boolean formulae
and finite automata,” in Proceedings of the 21st Annual ACM Symposium on Theory
of Computing (Seattle, WA), pp. 433-444, 1989.
18. R. E. Schapire, “The strength of weak learnability,” Machine Learning , vol. 5, no. 2,
pp. 197-227, 1990.
19. L. Breiman, “Random forests,” Machine Learning , vol. 45, no. 1, pp. 5-32, 2001.
20. S. Wang, K. Tang, and X. Yao, “Diversity exploration and negative correlation learn-
ing on imbalanced data sets,” in Proceedings of 2009 International Joint Conference
on Neural Networks , (Atlanta, GA), pp. 3259-3266, 2009.
21. N. V. Chawla, K. W. Bowyer, L. O. Hall, and W. P. Kegelmeyer, “SMOTE: Synthetic
minority over-sampling technique,” Journal of Artificial Intelligence Research , vol.
16, pp. 321-357, 2002.
22. C. Chen, A. Liaw, and L. Breiman, “Using random forest to learn imbalanced data,”
Tech. Rep., University of California, Berkeley, 2004.
23. L. Breiman, J. Friedman, R. Olshen, and C. Stone, Classification and Regression
Trees . Belmont, CA: Wadsworth International Group, 1984.
24. D. Tao, X. Tang, X. Li, and X. Wu, “Asymmetric bagging and random subspace for
support vector machines-based relevance feedback in image retrieval,” IEEE Trans-
actions on Pattern Analysis and Machine Intelligence , vol. 28, no. 7, pp. 1088-1099,
2006.
Search WWH ::




Custom Search