Multiple Classifier Systems

In undergoing this life, many people always try to do and get the best. New knowledge, experience, lesson, and everything that can improve the life will be done. However, many people sometimes feel confused to get those things. Feeling the limited of experience and sources to be better is one of the lacks to own. However, there is a very simple thing that can be done. This is what your teacher always manoeuvres you to do this one. Yeah, reading is the answer. Reading a book as this multiple classifier systems and other references can enrich your life quality. How can it be?

[1]  Edward H. Adelson,et al.  Belief Propagation and Revision in Networks with Loops , 1997 .

[2]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[3]  Jung-Fu Cheng,et al.  Turbo Decoding as an Instance of Pearl's "Belief Propagation" Algorithm , 1998, IEEE J. Sel. Areas Commun..

[4]  Robert Tibshirani,et al.  Classification by Pairwise Coupling , 1997, NIPS.

[5]  Andreas Stafylopatis,et al.  A Multi-SVM Classification System , 2001, Multiple Classifier Systems.

[6]  Fuad Rahman,et al.  Exploring a hybrid of support vector machines (SVMs) and a heuristic based system in classifying web pages , 2003, IS&T/SPIE Electronic Imaging.

[7]  S. Webb The Physics of Medical Imaging , 1990 .

[8]  Thorsten Joachims,et al.  Making large scale SVM learning practical , 1998 .

[9]  Christopher J. Merz,et al.  Using Correspondence Analysis to Combine Classifiers , 1999, Machine Learning.

[10]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques with Java implementations , 2002, SGMD.

[11]  Volker Roth,et al.  Probabilistic Discriminative Kernel Classifiers for Multi-class Problems , 2001, DAGM-Symposium.

[12]  Jürgen Schürmann,et al.  Pattern classification - a unified view of statistical and neural approaches , 2008 .

[13]  Alain Glavieux,et al.  Reflections on the Prize Paper : "Near optimum error-correcting coding and decoding: turbo codes" , 1998 .

[14]  Lihao Xu,et al.  Soft decoding of several classes of array codes , 2002, Proceedings IEEE International Symposium on Information Theory,.

[15]  B. Parhami Voting algorithms , 1994 .

[16]  Bernt Schiele,et al.  How many classifiers do I need? , 2002, Object recognition supported by user interaction for service robots.

[17]  Shlomo Shamai,et al.  Improved upper bounds on the ML decoding error probability of parallel and serial concatenated turbo codes via their ensemble distance spectrum , 2000, IEEE Trans. Inf. Theory.

[18]  Yoram Singer,et al.  Reducing Multiclass to Binary: A Unifying Approach for Margin Classifiers , 2000, J. Mach. Learn. Res..

[19]  Ludmila I. Kuncheva,et al.  'Change-glasses' approach in pattern recognition , 1993, Pattern Recognit. Lett..

[20]  Josef Kittler,et al.  Improving the performance of the product fusion strategy , 2000, Proceedings 15th International Conference on Pattern Recognition. ICPR-2000.

[21]  Günther Palm,et al.  Tree-Structured Support Vector Machines for Multi-class Pattern Recognition , 2001, Multiple Classifier Systems.

[22]  Eric Bauer,et al.  An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants , 1999, Machine Learning.

[23]  Yoram Singer,et al.  Improved Boosting Algorithms Using Confidence-rated Predictions , 1998, COLT' 98.

[24]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .