Abstract
Ensembles of learning machines have been formally and empirically shown to outperform (generalise better than) single learners in many cases. Evidence suggests that ensembles generalise better when they constitute members which form a diverse and accurate set. Diversity and accuracy are hence two factors that should be taken care of while designing ensembles in order for them to generalise better. There exists a trade-off between diversity and accuracy. Multi-objective evolutionary algorithms can be employed to tackle this issue to good effect. This chapter includes a brief overview of ensemble learning in general and presents a critique on the utility of multi-objective evolutionary algorithms for their design. Theoretical aspects of a committee of learners viz. the bias-variance-covariance decomposition and ambiguity decomposition are further discussed in order to support the importance of having both diversity and accuracy in ensembles. Some recent work and experimental results, considering classification tasks in particular, based on multi-objective learning of ensembles are then presented as we examine ensemble formation using neural networks and kernel machines.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
H. A. Abbass. A memetic pareto evolutionary approach to artificial neural networks. In Proceedings of the 14th Australian Joint Conference on Artificial Intelligence, pages 1–12, Berlin, 2000. Springer-Verlag.
H. A. Abbass. Pareto neuro-ensemble. In 16th Australian Joint Conference on Artificial Intelligence, pages 554–566, Perth, Australia, 2003. Springer.
H. A. Abbass. Pareto neuro-evolution: Constructing ensemble of neural networks using multi-objective optimization. In The IEEE 2003 Conference on Evolutionary Computation, volume 3, pages 2074–2080. IEEE Press, 2003.
H. A. Abbass. Speeding up backpropagation using multiobjective evolutionary algorithms. Neural Computation, 15(11):2705–2726, November 2003.
H. A. Abbass, R. Sarker, and C. Newton. Pde: A pareto-frontier differential evolution approach for multi-objective optimization problems. In Proceedings of the IEEE Congress on Evolutionary Computation (CEC2001), volume 2, pages 971–978. IEEE Press, 2001.
E. Bauer and R. Kohavi. An empirical comparison of voting classification algorithms: Bagging, boosting, and variants. Machine Learning, 36(1–2):105–139, 1999.
C. M. Bishop. Neural Networks for Pattern Recognition. Oxford University Press, 1995.
A. Blum and R. L. Rivest. Training a 3-node neural network is NP-complete. In Machine Learning: From Theory to Applications, pages 9–28, 1993.
E. Boers, M. Borst, and I. Sprinkhuizen-Kuyper. Evolving artificial neural networks using the “baldwin effect”. Technical Report 95–14, Leiden Unversity, Deptartment of Computer Science, The Netherlands, 1995.
L. Breiman. Bagging predictors. Machine Learning, 24(2):123–140, 1996.
L. Breiman. Using adaptive bagging to debias regressions. Technical Report 547, University of California, Berkeley, 1999.
G. Brown. Diversity in Neural Network Ensembles. PhD thesis, School of Computer Science, University of Birmingham, 2004.
G. Brown, J. Wyatt, R. Harris, and X. Yao. Diversity creation methods: A survey and categorisation. Journal of Information Fusion (Special issue on Diversity in Multiple Classifier Systems), 6:5–20, March 2005.
G. Brown and J. L. Wyatt. The use of the ambiguity decomposition in neural network ensemble learning methods. In T. Fawcett and N. Mishra, editors, 20th International Conference on Machine Learning (ICML’03), Washington DC, USA, August 2003.
A. Chandra. Evolutionary approach to tackling the trade-off between diversity and accuracy in neural network ensembles. Technical report, School of Computer Science, The University of Birmingham, UK, April 2004.
A. Chandra. Evolutionary framework for the creation of diverse hybrid ensembles for better generalisation. Master's thesis, School of Computer Science, The University of Birmingham, Birmingham, UK, September 2004.
A. Chandra and X. Yao. DIVACE: Diverse and Accurate Ensemble Learning Algorithm. In Proc. 5th Intl. Conference on Intelligent Data Engineering and Automated Learning (LNCS 3177), pages 619–625, Exeter, UK, August 2004. Springer-Verlag.
A. Chandra and X. Yao. Evolutionary framework for the construction of diverse hybrid ensembles. In M. Verleysen, editor, Proc. 13th European Symposium on Artificial Neural Networks, pages 253–258, Brugge, Belgium, April 2005. d-side.
A. Chandra and X. Yao. Evolving hybrid ensembles of learning machines for better generalisation. Neurocomputing (submitted), 2005.
A. Chandra and X. Yao. Ensemble learning using multi-objective evolutionary algorithms. Journal of Mathematical Modelling and Algorithms (to appear), 2006.
P. J. Darwen and X. Yao. Every niching method has its niche: Fitness sharing and implicit sharing compared. In Proc. of the 4th International Conference on Parallel Problem Solving from Nature (PPSN-IV), (LNCS-1141), pages 398–407, Berlin, September 1996. Springer-Verlag.
R. de Albuquerque Teixeira, A. P. Braga, R. H. Takahashi, and R. R. Saldanha. Improving generalization of mlps with multi-objective optimization. Neurocomputing, 35:189–194, 2000.
K. Deb. Multi-Objective Optimization Using Evolutionary Algorithms. Chichester, UK : Wiley, 2001.
T. G. Dietterich. Machine-learning research: Four current directions. The AI Magazine, 18(4):97–136, 1998.
T. G. Dietterich and G. Bakiri. Error-correcting output codes: a general method for improving multiclass inductive learning programs. In T. L. Dean and K. McKeown, editors, Proceedings of the Ninth AAAI National Conference on Artificial Intelligence, pages 572–577, Menlo Park, CA, 1991. AAAI Press.
P. Domingos. A unified bias-variance decomposition and its applications. In Proceedings of the Seventeenth International Conference on Machine Learning, pages 231–238, Stanford, CA, USA, 2000.
S. Forrest, R. E. Smith, B. Javornik, and A. S. Perelson. Using genetic algorithms to explore pattern recognition in the immune system. Evolutionary Computation, 1(3):191–211, 1993.
Y. Freund and R. Schapire. A short introduction to boosting. Journal of Japanese Society for Artificial Intelligence, 14(5):771–780, 1999.
Y. Freund and R. E. Schapire. Experiments with a new boosting algorithm. In Proceedings of the 13th International Conference on Machine Learning, pages 148–156. Morgan Kaufmann, 1996.
J. Friedman. On bias, variance, 0/1 loss and the curse of dimensionality. Data Mining and Knowledge Discovery, 1:55–77, 1997.
S. Geman, E. Bienenstock, and R. Doursat. Neural networks and the bias/variance dilemma. Neural Computation, 4(1):1–5, 1992.
S. Gutta, J. Huang, I. F. Imam, and H. Wechsler. Face and hand gesture recognition using hybrid classifiers. In Proceedings of the 2nd International Conference on Automatic Face and Gesture Recognition (FG ’96), pages 164–170. IEEE Computer Society, 1996.
L. K. Hansen and P. Salamon. Neural network ensembles. IEEE Transactions on Pattern Analysis and Machine Intelligence, 12(10):993–1001, 1990.
T. Heskes. Bias/variance decomposition for likelihood-based estimators. Neural Computation, 10:1425–1433, 1998.
T. K. Ho, J. J. Hull, and S. N. Srihari. Decision combination in multiple classifier systems. IEEE Transactions on Pattern Analysis and Machine Intelligence, 16(1):66–75, January 1994.
J. Horn and N. Nafpliotis. Multiobjective Optimization using the Niched Pareto Genetic Algorithm. Technical Report IlliGAl Report 93005, University of Illinois, Urbana-Champaign, July 1993.
Y. Jin, T. Okabe, and B. Sendhoff. Applications of Evolutionary Multi-objective Optimization (Advances in Natural Computation), volume 1, chapter Evolutionary multi-objective approach to constructing neural network ensembles for regression, pages 653–672. World Scientific, 2004.
Y. Jin, T. Okabe, and B. Sendhoff. Neural Network Regularization and Ensembling Using Multi-objective Evolutionary Algorithms. In 2004 Congress on Evolutionary Computation (CEC’2004), volume 1, pages 1–8, Portland, Oregon, USA, June 2004. IEEE Service Center.
E. Kong and T. Dietterich. Error - correcting output coding correct bias and variance. In Proceedings of The XII International Conference on Machine Learning, pages 313–321, San Francisco, CA, USA, 1995.
K. Kottathra and Y. Attikiouzel. A novel multicriteria optimization algorithm for the structure determination of multilayer feedforward neural networks. Journal of Network and Computer Applications, 19:135–147, 1996.
A. Krogh and J. Vedelsby. Neural network ensembles, cross validation, and active learning. Neural Information Processing Systems, 7:231–238, 1995.
M. A. Kupinski and M. A. Anastasio. Multiobjective genetic optimization of diagnostic classifiers with implications for generating receiver operating characteristic curves. IEEE Transactions on Medical Imaging, 18(8):675–685, August 1999.
W. B. Langdon, S. J. Barrett, and B. F. Buxton. Combining decision trees and neural networks for drug discovery. In Genetic Programming, Proceedings of the 5th European Conference, EuroGP 2002, pages 60–70, Kinsale, Ireland, 3–5 April 2002.
B. Littlewood and D. R. Miller. Conceptual modeling of coincident failures in multiversion software. IEEE Transactions on Software Engineering, 15(12):1596–1614, December 1989.
Y. Liu and X. Yao. Ensemble learning via negative correlation. Neural Networks, 12(10):1399–1404, 1999.
Y. Liu and X. Yao. Learning and evolution by minimization of mutual information. In J. J. M. Guervós, P. Adamidis, H.-G. Beyer, J.-L. Fernández-Villacañas, and H.-P. Schwefel, editors, Parallel Problem Solving from Nature VII (PPSN-2002), volume 2439 of LNCS, pages 495–504, Granada, Spain, 2002. Springer Verlag.
Y. Liu, X. Yao, and T. Higuchi. Evolutionary ensembles with negative correlation learning. IEEE Transactions on Evolutionary Computation, 4(4):380, November 2000.
R. Meir and G. Raetsch. An introduction to boosting and leveraging. Advanced lectures on machine learning, pages 118–183, 2003.
C. E. Metz. Basic principles of roc analysis. Seminars in Neuclear Medicine, 8(4):283–298, 1978.
D. Michie, D. Spiegelhalter, and C. Taylor. Machine Learning, Neural and Statistical Classification. Ellis Horwood Limited, 1994.
D. Opitz and R. Maclin. Popular ensemble methods: An empirical study. Journal of Artificial Intelligence Research, 11:169–198, 1999.
D. W. Opitz and J. W. Shavlik. Generating accurate and diverse members of a neural-network ensemble. Neural Information Processing Systems, 8:535–541, 1996.
T. Schnier and X. Yao. Using negative correlation to evolve fault-tolerant circuits. In Proceedings of the 5th International Conference on Evolvable Systems: From Biology to Hardware (ICES’2003), pages 35–46. Springer-Verlag. Lecture Notes in Computer Science, Vol. 2606, March 2003.
A. Sharkey. Multi-Net Systems, chapter Combining Artificial Neural Nets: Ensemble and Modular Multi-Net Systems, pages 1–30. Springer-Verlag, 1999.
A. Sharkey and N. Sharkey. Combining diverse neural networks. The Knowledge Engineering Review, 12(3):231–247, 1997.
K. O. Stanley and R. Miikkulainen. Evolving neural networks through augmenting topologies. Evolutionary Computation, 10(2):99–127, 2002.
K. Tumer and J. Ghosh. Analysis of decision boundaries in linearly combined neural classifiers. Pattern Recognition, 29(2):341–348, February 1996.
N. Ueda and R. Nakano. Generalization error of ensemble estimators. In Proceedings of International Conference on Neural Networks, pages 90–95, 1996.
G. Valentini and T. G. Dietterich. Bias-variance analysis of support vector machines for the development of svm-based ensemble methods. Journal of Machine Learning Research, 5(1):725–775, 2004.
G. Valentini and F. Masulli. Ensembles of learning machines. In R. Tagliaferri and M. Marinaro, editors, Neural Nets WIRN Vietri-2002 (LNCS 2486), pages 3–19. Springer-Verlag, June 2002.
W. Wang, P. Jones, and D. Partridge. Diversity between neural networks and decision trees for building multiple classifier systems. In Proc. Int. Workshop on Multiple Classifier Systems (LNCS 1857), pages 240–249, Calgiari, Italy, June 2000. Springer.
W. Wang, D. Partridge, and J. Etherington. Hybrid ensembles and coincidentfailure diversity. In Proceedings of the International Joint Conference on Neural Networks, (2001), volume 4, pages 2376–2381, Washington, USA, July 2001. IEEE Press.
K. Woods, W. Kegelmeyer, and K. Bowyer. Combination of multiple classiers using local accuracy estimates. IEEE Transactions on Pattern Analysis and Machine Intelligence, 19:405–410, 1997.
X. Yao. Evolving artificial neural networks. In Proceedings of the IEEE, volume 87, pages 1423–1447. IEEE, September 1999.
X. Yao and Y. Liu. Evolving neural network ensembles by minimization of mutual information. International Journal of Hybrid Intelligent Systems, 1(1), January 2004.
W. Yates and D. Partridge. Use of methodological diversity to improve neural network generalization. Neural Computing and Applications, 4(2):114–128, 1996.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer
About this chapter
Cite this chapter
Chandra, A., Chen, H., Yao, X. (2006). Trade-Off Between Diversity and Accuracy in Ensemble Generation. In: Jin, Y. (eds) Multi-Objective Machine Learning. Studies in Computational Intelligence, vol 16. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-33019-4_19
Download citation
DOI: https://doi.org/10.1007/3-540-33019-4_19
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-30676-4
Online ISBN: 978-3-540-33019-6
eBook Packages: EngineeringEngineering (R0)