Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
2003, Lecture Notes in Computer Science
This study covers weighted combination methodologies for multiple classifiers to improve classification accuracy. The classifiers are extended to produce class probability estimates besides their class label assignments to be able to combine them more efficiently. The leave-one-out training method is used and the results are combined using proposed weighted combination algorithms. The weights of the classifiers for the weighted classifier combination are determined based on the performance of the classifiers on the training phase. The classifiers and combination algorithms are evaluated using classical and proposed performance measures. It is found that the integration of the proposed reliability measure, improves the performance of classification. A sensitivity analysis shows that the proposed polynomial weight assignment applied with probability based combination is robust to choose classifiers for the classifier set and indicates a typical one to three percent consistent improvement compared to a single best classifier of the same set.
IEEE Access
In critical applications, such as medical diagnosis, security related systems, and so on, the cost or risk of action taking based on incorrect classification can be very high. Hence, combining expert opinions before taking decision can substantially increase the reliability of such systems. Such pattern recognition systems base their final decision on evidence collected from different classifiers. Such evidence can be of data type, feature type, or classifier type. Common problems in pattern recognition, such as curse of dimensionality, and small sample data size, among others, have also prompted researchers into seeking new approaches for combining evidences. This paper presents a criteria-based framework for multiclassifiers combination techniques and their areas of applications. The criteria discussed here include levels of combination, types of thresholding, adaptiveness of the combination, and ensemble-based approaches. The strengths and weaknesses of each of these categories are discussed in details. Following this analysis, we provide our perspective on the outlook of this area of research and open problems. The lack of a well-formulated theoretical framework for analyzing the performance of combination techniques is shown to provide a fertile ground for further research. In addition to summarizing the existing work, this paper also updates and complements the latest developments in this area of research.
International Journal on Document Analysis and Recognition, 2003
This paper presents a framework for the analysis of similarity among abstract-level classifiers and proposes a methodology for the evaluation of combination methods. In this paper, each abstract-level classifier is considered as a random variable, and sets of classifiers with different degrees of similarity are systematically simulated, combined, and studied. It is shown to what extent the performance of each combination method depends on the degree of similarity among classifiers and the conditions under which each combination method outperforms the others. Experimental tests have been carried out on simulated and real data sets. The results confirm the validity of the proposed methodology for the analysis of combination methods and its usefulness for multiclassifier system design.
Computational Intelligence, 2017
Classifier combination methods have proved to be an effective tool to increase the performance of classification techniques that can be used in any pattern recognition applications. Despite a significant number of publications describing successful classifier combination implementations, the theoretical basis is still not matured enough and achieved improvements are inconsistent. In this paper, we propose a novel statistical validation technique known as correlation-based classifier combination technique for combining classifier in any pattern recognition problem. This validation has significant influence on the performance of combinations, and their utilization is necessary for complete theoretical understanding of combination algorithms. The analysis presented is statistical in nature but promises to lead to a class of algorithms for rankbased decision combination. The potentials of the theoretical and practical issues in implementation are illustrated by applying it on 2 standard datasets in pattern recognition domain, namely, handwritten digit recognition and letter image recognition datasets taken from UCI Machine Learning Database Repository (http://www.ics. uci.edu/_mlearn). 1 An empirical evaluation using 8 wellknown distinct classifiers confirms the validity of our approach compared to some other combinations of multiple classifiers algorithms. Finally, we also suggest a methodology for determining the best mix of individual classifiers.
Object recognition supported by user interaction for service robots, 2002
In this paper we introduce a novel multiple classifier system that incorporates a global optimization technique based on a genetic algorithm for configuring the system. The system adopts the weighted majority vote approach to combine the decision of the experts, and obtains the weights by maximizing the performance of the whole set of experts, rather than that of each of them separately. The system has been tested on a handwritten digit recognition problem, and its performance compared with those exhibited by a system using the weights obtained during the training of each expert separately. The results of a set of experiments conducted on a 30,000 digits extracted from the NIST database shown that the proposed system exhibits better performance than those of the alternative one, and that such an improvement is due to a better estimate of the reliability of the participating classifiers.
2009
Usage of recognition systems has found many applications in almost all fields. However, Most of classification algorithms have obtained good performance for specific problems; they have not enough robustness for other problems. Combination of multiple classifiers can be considered as a general solution method for pattern recognition problems. It has been shown that combination of classifiers can usually operate better than single classifier provided that its components are independent or they have diverse outputs. It was shown that the necessary diversity of an ensemble can be achieved manipulation of data set features. We also propose a new method of creating this diversity. The ensemble created by proposed method may not always outperforms all classifiers existing in it, it is always possesses the diversity needed for creation of ensemble, and consequently it always outperforms the simple classifier.
International Journal of Pattern Recognition and Artificial Intelligence, 2003
ABSTRACT When several classifiers are brought to contribute to the same task of recognition, various strategies of decisions, implying these classifiers in different ways, are possible. A first strategy consists in deciding using different opinions: it corresponds to the combination of classifiers. A second strategy consists in using one or more opinions for better guiding other classifiers in their training stages, and/or to improve the decision-making of other classifiers in the classification stage: it corresponds to the cooperation of classifiers. The third and last strategy consists in giving more importance to one or more classifiers according to various criteria or situations: it corresponds to the selection of classifiers. The temporal aspect of Pattern Recognition (PR), i.e. the possible evolution of the classes to be recognized, can be treated by the strategy of selection.
Artificial Intelligence Review, 2008
Several real problems involve the classification of data into categories or classes. Given a data set containing data whose classes are known, Machine Learning algorithms can be employed for the induction of a classifier able to predict the class of new data from the same domain, performing the desired discrimination. Some learning techniques are originally conceived for the solution of problems with only two classes, also named binary classification problems. However, many problems require the discrimination of examples into more than two categories or classes. This paper presents a survey on the main strategies for the generalization of binary classifiers to problems with more than two classes, known as multiclass classification problems. The focus is on strategies that decompose the original multiclass problem into multiple binary subtasks, whose outputs are combined to obtain the final prediction.
IEEE Transactions on Pattern Analysis and Machine Intelligence, 2000
In this paper, a theoretical and experimental analysis of linear combiners for multiple classifier systems is presented. Although linear combiners are the most frequently used combining rules, many important issues related to their operation for pattern classification tasks lack a theoretical basis. After a critical review of the framework developed in works by Tumer and Ghosh on which our analysis is based, we focus on the simplest and most widely used implementation of linear combiners, which consists of assigning a nonnegative weight to each individual classifier. Moreover, we consider the ideal performance of this combining rule, i.e., that achievable when the optimal values of the weights are used. We do not consider the problem of weights estimation, which has been addressed in the literature. Our theoretical analysis shows how the performance of linear combiners, in terms of misclassification probability, depends on the performance of individual classifiers, and on the correlation between their outputs. In particular, we evaluate the ideal performance improvement that can be achieved using the weighted average over the simple average combining rule and investigate in what way it depends on the individual classifiers. Experimental results on real data sets show that the behavior of linear combiners agrees with the predictions of our analytical model. Finally, we discuss the contribution to the state of the art and the practical relevance of our theoretical and experimental analysis of linear combiners for multiple classifier systems.
17 th National Conference on Artificial Intelligence ( …, 2000
In recent years, the combination of classifiers has been proposed as a method to improve the accuracy achieved in isolation by a single classifier. We are interested in ensemble methods that allow the combination of heterogeneous sets of classifiers, which ...
1992
Difficult pattern recognition problems involving large class sets and noisy input can be solved by a multiple classifier system, which allows simultaneous use of arbitrary feature descriptors and classification procedures. Independent decisions by each classifier can be combined by methods of the highest rank, Borda count, and logistic regression, resulting in substantial improvement in overall correctness
1997
In the last years, great attention has been devoted to multiple classifier systems. The implementation of such a system implies the definition of a rule (combining rule) for determining the most likely class, on the basis of the class attributed by each single expert. The availability of a criterion to evaluate the credibility of the decision taken by a classifier can be profitable in order to implement the combining rule. We propose a method that, after defining the reliability of a classification on the basis of information directly derived from the output of the classifier, uses this information in the context of a combining rule. The results obtained by combining four handwritten character recognizers on the basis of classification reliability are compared with those obtained by using three different combining criteria. Tests have been performed using a standard handwritten character database.
2014
The One-Class Classifier (OCC) has been widely used for solving the one-class and multi-class classification problems. Its main advantage for multi-class is offering an open system and therefore allows easily extending new classes without retraining OCCs. However, extending the OCC to the multi-class classification achieves less accuracy comparatively to other multiclass classifiers. Hence, in order to improve the accuracy and keep the offered advantage we propose in this paper a Multiple Classifier System (MCS), which is composed of different types of OCC. Usually, the combination is performed using fixed or trained rules. Generally, the static weighted average is considered as straightforward combination rule. In this paper we propose a dynamic weighted average rule that calculates the appropriate weights for each test sample. Experimental results conducted on several real-world datasets proves the effective use of the proposed multiple classifier system where the dynamic weighted average rule achieves the best results for most datasets versus the mean, max, product and the static weighted average rules.
2012 International Conference on Frontiers in Handwriting Recognition, 2012
Classifier combination methods have proved to be an effective tool for increasing the performance in pattern recognition applications. The rationale of this approach follows from the observation that appropriately diverse classifiers make uncorrelated errors. Unfortunately, this theoretical assumption is not easy to satisfy in practical cases, thus reducing the performance obtainable with any combination strategy. In this paper we propose a new weighted majority vote rule which try to solve this problem by jointly analyzing the responses provided by all the experts, in order to capture their collective behavior when classifying a sample. Our rule associates a weight to each class rather than to each expert and computes such weights by estimating the joint probability distribution of each class with the set of responses provided by all the experts in the combining pool. The probability distribution has been computed by using the naive Bayes probabilistic model. Despite its simplicity, this model has been successfully used in many practical applications, often competing with much more sophisticated techniques. The experimental results, performed by using three standard databases of handwritten digits, confirmed the effectiveness of the proposed method.
2000
In many applications of computer vision, combination of decisions from mulitiple sources is a very important way of achieving more accurate and robust classification. Many such techniques can be used, two of which are the Majority Voting and the Divide and Conquer techniques. The former achieves decision combination by measuring consensus among the participating classifiers and the latter achieves the same by dividing the problem into smaller problems and solving each of these sub-problems more e f iciently. Both these approaches have their advantages and disadvantages. In this paper, a novel approach t o combining these two techniques is presented. Although the success of the approach has been demonstrated in a typical application area of computer vision (recognition of complex and highly variable image data), the approach is completely generalised and is applicable to other task domains.
This paper presents a method for combining classifiers that uses estimates of each individual classifier's local accuracy in small regions of feature space surrounding an unknown test sample. An empirical evaluation using five real data sets confirms the validity of our approach compared to some other Combination of Multiple Classifiers algorithms. We also suggest a methodology for determining the best mix of individual classifiers.
2004
The aim of this paper is to investigate the role of the a-priori knowledge in the process of classifier combination. For this purpose three combination methods are compared which use different levels of a-priori knowledge. The performance of the methods are measured under different working conditions by simulating sets of classifier with different characteristics. For this purpose, a random variable is used to simulate each classifier and an estimator of stochastic correlation is used to measure the agreement among classifiers. The experimental results, which clarify the conditions under which each combination method provides better performance, show to what extend the a-priori knowledge on the characteristics of the set of classifiers can improve the effectiveness of the process of classifier combination.
Information Fusion, 2001
This paper proposes a novel algorithm for multiple classi®ers combination based on clustering and selection technique (called M3CS), which can ®nd in the feature space the regions where each classi®er has best classi®cation performance. The proposed method may be divided into two steps: clustering and selection (operation). At clustering step, the feature space is partitioned into several regions by clustering separately the correctly and incorrectly classi®ed training samples from each classi®er, and the performances of the classi®er in each region are calculated. In the selection step, the most accurate classi®er in the vicinity of the input sample is nominated to provide the ®nal decision of the committee. The performance comparison between M3CS and Kuncheva's CS DT method, as well as some simple aggregation methods such as maximum, minimum, average, and majority vote, con®rms the validity of the proposed scheme.
2002
A new approach to serial multi-stage combination of classifiers is proposed. Each classifier in the sequence uses a smaller subset of features than the subsequent classifier. The classification provided by a classifier is rejected only if its decision is below a pre-defined confidence level. The approach is tested on a two-stage combination of k-Nearest Neighbor classifiers. The features to be used by the first classifier in the combination are selected by two stand-alone algorithms (Relief and Info-Fuzzy Network, or IFN) and a hybrid method, called "IFN + Relief." The feature-based approach is shown empirically to provide a substantial decrease in the computational complexity, while maintaining the accuracy level of a single-stage classifier or even improving it.
Neurocomputing, 2018
Classifier ensembles are pattern recognition structures composed of a set of classification algorithms (members), organized in a parallel way, and a combination method with the aim of increasing the classification accuracy of a classification system. In this study, we investigate the application of a generalized mixture (GM) functions as a new approach for providing an efficient combination procedure for these systems through the use of dynamic weights in the combination process. Therefore, we present three GM functions to be applied as a combination method. The main advantage of these functions is that they can define dynamic weights at the member outputs, making the combination process more efficient. In order to evaluate the feasibility of the proposed approach, an empirical analysis is conducted, applying classifier ensembles to 25 different classification data sets. In this analysis, we compare the use of the proposed approaches to ensembles using traditional combination methods as well as the state-of-the-art ensemble methods. Our findings indicated gains in terms of performance when comparing the proposed approaches to the traditional ones as well as comparable results with the state-of-the-art methods.
2014
In this work we were interested in investigating the predictive accuracy of one of the most popular learning schemes for the combination of supervised classification methods: the Stacking Technique proposed by Wolpert (1992) and consolidated by Ting and Witten, (1999) and Seewald (2002). In particular, we made reference to the StackingC (Seewald 2002) as a starting point for our analysis, to which some modifications and extensions were made. Since most of the research on ensembles of classifiers tends to demonstrate that this scheme can perform comparably to the best of the base classifiers as selected by cross-validation, if not better, this motivated us to investigate the performance of the Stacking empirically. An analysis of the results obtained by applying the our Stacking scheme, which includes differences and characteristic implementations compared to what is proposed by the literature, to the set of the dataset generated by means of an experimental design does not lead us to...
Loading Preview
Sorry, preview is currently unavailable. You can download the paper by clicking the button above.