Journal of Software Engineering and Applications, 2011, 4, 426-432 doi:10.4236/jsea.2011.47049 Published Online July 2011 (http://www.SciRP.org/journal/jsea) Copyright © 2011 SciRes. JSEA Mapping Software Metrics to Module Complexity: A Pattern Classification Approach Nick John Pizzi Department of Computer Science, University of Manitoba, Winnipeg, Canada. Email: pizzi@cs.umanitoba.ca Received June 3rd, 2011; revised June 30th, 2011; accepted July 1st, 2011. ABSTRACT A desirable software engineering goal is the prediction of software module complexity (a qualitative concept) using automatically generated software metrics (quantitative measurements). This goal may be couched in the language of pattern classification; namely, given a set of metrics (a pattern) for a software module, predict the class (level of com- plexity) to which the module belongs. To find this mapping from metrics to complexity, we present a classification strategy, stochastic metric selection, to determine the subset of software metrics that yields the greatest predictive pow- er with respect to module complexity. We demonstrate the effectiveness of this strategy by empirically evaluating it us- ing a publicly available dataset of metrics compiled from a medical imaging system and comparing the prediction re- sults against several classification system benchmarks. Keywords: Software Metrics, Pattern Classification, Feature Selection, Software Complexity 1. Introduction Software systems are used to solve or model increas- ingly complex problems in a variety of application do- mains. However, problem complexity should not mani- fest itself as software module complexity. That is, prob- lem complexity should only produce software complex- ity as an emergent property and not as an intrinsic prop- erty of individual software modules (functions, objects, templates, and so on). Intrinsic module complexity is an undesirable characteristic that is often subjectively iden- tified by software developers for some form of remedial action. However, this subjective process is slow, tedious, and error-prone. Automating (to some extent) this proc- ess would be desirable especially if coupled with a more rigorous quantitative approach. One possible strategy is the prediction of software module complexity using auto- matically generated quantitative measurements, namely software metrics [1,2]. A software metric is a mapping from a software mod- ule to a set of numerical values to quantify one or more software attributes [3]. These metrics are commonly re- garded as important factors reflecting the nature of soft- ware systems including, among other characteristics, the property of complexity [4-7]. Software complexity has a direct influence on its quality, clarity, extensibility, and maintainability. The analysis of complexity may be cast as a problem of pattern classification: find a mapping that predicts the class label (level of software complexity) of a pattern (software module), given only its features (software metrics). Such a classification system could serve as a “complexity filter” for software modules. For instance, a project manager or software developer could use such a filter to predict module complexity to identify highly complex software modules for review and possi- ble revision. While some metrics may have discriminatory power, that is, predictive utility vis-à-vis complexity, other met- rics may have none. Moreover, it is possible for some metrics to have a potential deleterious effect on the clas- sification process [8-11]; these confounding metrics dis- tort class boundaries in the metric space (for instance, delineating between high complexity and low complexity modules), thereby reducing classification performance (accuracy). Therefore, it is important not only to identify discriminatory metrics but to also cull confounding ones. To this end, stochastic metric selection is investigated as a pattern classification strategy to determine the subset of software metrics that best predict software module com- plexity. The strategy, which stochastically examines subsets of software metrics for their predictive power, is empirically evaluated using a database of software met- rics from a large medical image processing system, and
Mapping Software Metrics to Module Complexity: a Pattern Classification Approach427 the strategy’s classification performance is compared against several classification system benchmarks. Section 2 presents a brief overview of the dataset’s software met- rics. Section 3 describes the classification strategy for finding a mapping between software metrics and module complexity. Details relating to the design of the experi- ments and the results compared against the benchmarks are presented in Section 4 and Section 5, respectively, with concluding remarks in Section 6. 2. Classification & Software Measurements Couched in classification terms, we are interested in finding a mapping between patterns (metrics) and class labels (complexity); formally, X={(xk, ωk), k = 1,2,…,N} is a set of N patterns, xk∈ Ը n, with respective class labels, ωk ∈={1,2,…,c}. X is randomly allocated to a design subset (often referred to as a training set), XD, comprising ND patterns, or a validation subset (often referred to as a test set), XV, comprising NV patterns (ND + NV = N). A classification system finds a mapping, f: XD . Sub- sequently, f is validated using XV, f: XV . In our spe- cific case, X is the dataset described in Section 2.1 where xk is the list of software metrics (see Section 2.2) for software module k. As defined in Section 2.3, we will use c = 2 class labels for complexity: low, ω1, and high, ω2. It is important to note at this point that the choices for metrics and class labels described below were made for the sake of expediency; specifically, these were the pa- rameters that were supplied in the publicly available da- taset. Class labels, based on change counts (see Section 2.3), are unlikely to be the best delineators of software module complexity, nor are the supplied metrics neces- sarily the best predictors of complexity. No claim is made as to the optimality of these parameters vis-à-vis complexity prediction, rather we make the simple prag- matic assertion that they will serve as an experimental test bed to examine the efficacy of treating the assess- ment of software module complexity as a problem of pattern classification. 2.1. Medical Imaging System This investigation considers the Medical Imaging System (MIS) dataset [12] that has been used in some other stu- dies dealing with software quality [13,14]. MIS is a commercial software system consisting of approximately 4500 modules and comprising approximately 400000 lines of code. The dataset consists of N = 390 software modules. Each module is described using n = 11 software metrics as well as its change count, which is a reflection of its complexity [7]. 2.2. Software Metrics As the software engineering literature provides thorough and extensive discussions of the software metrics used in this investigation, including their relative advantages and disadvantages [13,15-17], we restrict ourselves to brief summaries. The first software metrics collected for each MIS module are the ubiquitous “lines of code” (L1), which is the total number of lines of code including pro- grammer comments, and the number of lines of code excluding comments (L2). The next set of software measures includes the number of code characters (C1), comments (C2), comment characters (C3), and code characters (C4), respectively. Another set of software measures pertain to more complex computations based on the length of a program: Halstead’s measure of program length (P1), which is the sum of the total number of program operators and the total number of program operands [18]; Halstead’s soft- ware measure of estimated program length (P2), P2 = η1logη1 + η2logη2, where η1 and η2 are the total number of unique program operators and program operands, re- spectively; and, Jensen’s program length (P3), P3 = (logη1)! + (logη2)! [16]. The final two software measures concern program control flow graphs: McCabe’s cyclomatic number (F1), which is one more than the total number of decision nodes in the control flow graph [19]; and Belady’s bandwidth metric (F2), F2 = a–1 i (i ai), where ai is the number of nodes at level i in a nested control flow graph of a nodes. 2.3. Change Count In this investigation, the change count (CC) measure is used as the class label. While the MIS dataset has 42 unique CC labels, it is unrealistic to expect reliable classi- fication results using so many class labels with only 390 modules. As a result, the CC class labels are aggregated into two qualitative classes, low (ω1) complexity versus high (ω2) complexity, using two different points of de- lineation, DS3 (no more than three programmer changes) and DS4 (no more than four programmer changes). The delineations were arbitrarily chosen, before the classifica- tion experiments were conducted, so that the two com- plexity classes were well distributed (roughly an equal number of software modules in each class). With DS3, the number of software modules in the low complexity class is N11 = 192 and the number of software modules in the high class is N12 = 198, respectively. With DS4, the number of software modules in the low class is N21 = 220 and the number of modules in the high class is N22 = 170. 3. Classification Method 3.1. Stochastic Metric Selection The motivation for pre-processing strategies exploiting Copyright © 2011 SciRes. JSEA
Mapping Software Metrics to Module Complexity: a Pattern Classification Approach 428 pattern feature selection [20-22] is to simplify the deter- mination and construction of optimal class boundaries in the feature space. An example of such a strategy is sto- chastic metric selection (SMS), which is based on previ- ous work described in [11] but modified to deal with software metrics. Figure 1 presents a flowchart of the iterative SMS algorithm, which individual steps we now describe below. The investigator must select the classifier types to be used; SMS operates with homogeneous or heterogeneous collections of pattern classifiers e.g., linear discriminant analysis, artificial neural networks, genetic algorithms, and support vector machines). The algorithm terminates upon exceeding a maximum number of iterations, τ, or a performance (classification accuracy) threshold, PS. A number of different performance measures may be used (see Section 4.2). The subsets of metrics are constrained by four parameters: the minimum and maximum number of metric subsets, and the minimum and maximum car- dinality for a metric subset. Metric subsets may also be quadratically combined to produce “new” metrics (see below). Using b-fold validation, X is randomly allocated to either XD or XV. During the design phase, an instantia- tion of one of the pattern classifier types is used with randomly sampled metric subsets (satisfying the con- straints above) from XD to design a mapping. Subse- quently, during the validation phase, the classification performance (accuracy), P, is measured using XV. This process iterates with many pairs of classifier instances and metric subsets. SMS exploits the quadratic combination of metric subsets. The intent is that if the original metric space had non-linear class boundaries, the new (quadratic) parame- ter space may end up having class boundaries that are more linear. SMS has three categories of quadratic com- binations: using the original metric subset; squaring the SMS algorithm (1) Select parameter values, classifiers types, and P. (2) Repeat while iteration count and P' PS: i) Select metric subsets using histogram. ii) Perform quadratic transformation of subsets. iii) Instantiate a classifier type. iv) Repeat b times (b-fold validation): a) Randomly allocate X to XD and XV. b) Design mapping using classifier and XD. c) Assess performance using XV. v) If P' PH, update frequency histogram. vi) If P' exceeds current best performance, update. (3) Report best P' (with standard deviation). Figure 1. Flowchart for SMS algorithm (P' is the median performance result of a set of b runs). metric values for a particular metric subset; and using all pair-wise cross products of metrics from two subsets. The probabilities (set by the investigator) of selecting one of these quadratic metric combination categories must sum to 1. The stochastic nature of SMS is controlled by a metric frequency histogram whereby the performance of each classification task is assessed using the selected per- formance measure. If the performance exceeds the histo- gram fitness threshold, PH, the histogram is incremented at those metric indices corresponding to the subsets used by the classification task. This metric histogram is used to generate an ad hoc cumulative distribution function, which is used when randomly sampling new metric sub- sets. So, rather than each metric having an equal likeli- hood of being selected for a new classification task (a new classifier instance), those software metrics that were used in previous “successful” classification tasks have a greater likelihood of being chosen. A temperature term, 0 t 1, provides additional control over this metric sam- pling process. If t = 0, the cumulative distribution func- tion is used but, as t 1, the randomness becomes more uniform (when t = 1 a strict uniform distribution is used). 3.2. Underlying Classifier As previously mentioned, the SMS algorithm can operate with any kind of underlying pattern classifier. In this investigation, we chose a simple linear pattern classifier using linear discriminant analysis (LDA) [8], which de- termines linear class boundaries between c classes (in our case, complexity) while simultaneously taking into ac- count between-class and within-class variances. If the error distributions for each class are the same, it can be shown that LDA finds the optimal set of linear class boundaries. In real-world pattern classification situations, this optimality is seldom achieved since different classes of patterns typically give rise to different (sometimes, significantly different) distributions. The LDA algorithm allocates a pattern, x, to class k for which the probability distribution, pk(x), is greatest, that is, x is allocated to class k, if qk pk(x) qi pi(x) (i k), where qk are the prior (or proportional) probabilities. Lk(x) = logqk + mk T W–1(x–½mk) is the discriminant function for class k where mk is the mean for class k and W is the co- variance matrix. The metric space hyperplane, Fki(x), separating class k from class i is defined as Fki(x) = Lk(x) – Li(x) = 0. 4. Experiment Design 4.1. Benchmarks To assess the effectiveness of SMS for the prediction of module complexity, we compare it against the perform- Copyright © 2011 SciRes. JSEA
Mapping Software Metrics to Module Complexity: a Pattern Classification Approach429 ance of two classification system benchmarks, the sup- port vector machine and the radial basis function net- work. 4.1.1. Support Vector Machine The support vector machine (SVM) [23,24] is an impor- tant family of supervised learning algorithms that select models that maximize the error margin of a training (de- sign) subset. This approach has been successively used in a wide range of pattern classification problems [25,26]. Given a set of patterns that are assigned to one of either two classes, an SVM finds the hyperplane leaving the largest possible fraction of patterns of the same class on the same side while maximizing the distance of either class from the hyperplane. The SVM approach is usually formulated as a constrained optimization problem and solved using constrained quadratic programming. While the original SVM approach [26] could only be used for linearly separable problems, it may be extended by employing a “kernel trick” [27] that exploits the fact that a non-linear mapping of sufficiently high dimension can project the patterns to a new parameter space in which classes can be separated by a hyperplane. In gen- eral, it cannot be determined a priori which kernel will produce good classification results for a given dataset so one must rely on heuristic (trial and error) experiments. Common kernel functions, K(x, y), for patterns, x, y, in- clude (x × y)d, (ax × y + b)d, tanh(ax × y + b), and exp(–½|x – y|2/σ). 4.1.2. Radial Basis Function Network A radial basis function network (RBF) [28], which may be used for pattern classification, has an internal repre- sentation of nodes that possess radial symmetry f(x) = φ(||x–μ||) where: μ is its centre; ||·|| is a distance metric that determines how far an input pattern is from μ; and, the transfer function, φ, must output high values when the distance from a pattern to μ is small, and low values otherwise. While RBF train quickly [29], selecting the number of receptive fields is strictly ad hoc. If μi is a column vector representing the centre of pattern layer node i and σi is its receptive region diameter, then zi(x) = exp[–(x – μi)T(x – μi)/(2σi 2) is the output for a given pat- tern, x. The values μ and σ may analogously be viewed as the mean and standard deviation of the response curve, respectively. The response function of an RBF node diminishes ra- pidly as patterns deviate from its mean. Here, the pattern layer weights are trained as well as the location and shape of the response curves. Standard k-means cluster- ing is used to compute a μ set. The radius of the normal- ized zi is determined by σi. If μi is widely separated then σi should be large to cover the gaps. If they are tightly packed then σi should be small enough to accurately re- tain the distinctiveness of each receptive field. P-nearest neighbor is a standard heuristic algorithm used to deter- mine σi is. Given a receptive region’s centre, σi, let i1,…,ip be the indices of the P centres nearest to μi. Then σi = (P–1p||μi – μip||2)½ (here, P=1). 4.2. Measuring Performance Selecting a method for measuring the performance (ac- curacy) of a classification system is often taken for granted, which often leads to overly optimistic classifica- tion performance results. Given an n×n confusion matrix of desired versus predicted class labels, classification performance is typically measured using the observed agreement, Po = N–1i nii (i=1,2,…,c), which is the ratio of patterns that lie in regions associated with their corre- sponding classes to the total number of patterns regard- less of where they lie (nii are the diagonal elements of the confusion matrix). While often used to measure classifi- cation performance, Po does not take into account the classification agreement that might be due to chance, Pr = N–2i (j nij j nji) (i, j=1 ,2 ,…,c) [30]. A more conserva- tive classification performance measure is the κ score [31,32], a chance-corrected measure of agreement be- tween the desired and predicted class assignments, κ = (Po – Pr)/(1 – Pr). If the classification agreement is due strictly to chance, κ = 0. If the agreement is greater than chance κ>0; κ = 1 indicates complete (perfect) agreement. If the agreement is less than chance then κ < 0 (the min- imum value depends upon the marginal distributions). 4.3. Parameter Settings The following parameter values were used with the SMS algorithm: underlying classifier, LDA; frequency histo- gram threshold, 0.4; temperature threshold, t = 0.5; number of metric subsets, 1–11; metric subset cardinality, 1–11; validation, b = 5; and classification performance measure, κ. For all experiments, five receptive fields were used for the RBF algorithm and the Gaussian kernel was used for SVM. Concerning SMS quadratic metric combinations, 30% were squared terms, pair-wise cross products were utilized 30% of the time, and the remain- ing 40% used subsets of the original metrics. For SMS, each classification experiment was allowed to run for 105 iterations. 5. Results and Discussion 5.1. Predictive Power Tables 1 and 2 list the breakdown of software module patterns into design (XD) and validation (XV) subsets of metrics for the respective datasets, DS3 and DS4, as des Copyright © 2011 SciRes. JSEA
Mapping Software Metrics to Module Complexity: a Pattern Classification Approach 430 Table 1. Low/high complexity breakdown: Design and vali- dation subsets for DS3. Metric Patterns Class Label XD XV Total Low (1) 128 64 N11 = 192 High (2) 128 70 N12 = 198 Table 2. Low/high complexity breakdown: Design and vali- dation subsets for DS4. Metric Patterns Class Label XD XV Total Low (1) 114 106 N21 = 220 High (2) 114 56 N22 = 170 cribed in Section 2.3. The SMS, SVM, and RBF algo- rithms all use XD to design their respective classification mappings and XV to validate the classification perform- ance. Table 3 lists the mean (with standard deviations) clas- sification performance results for DS3 with all classifiers using the validation subset of metrics, XV. It is clear that the SMS algorithm produced significantly better classi- fication results than either benchmark (recall that both benchmarks used all of the original metrics). Compared against the best classifier benchmark (the SVM algo- rithm), SMS produced a 29% increase in the κ score (0.71 ± 0.02 versus 0.55 ± 0.02) and a 10% increase in the standard Po agreement measure (0.86 ± 0.01 versus 0.78 ± 0.01). At this point, it is important to note the greater rate of improvement with the conservative κ score. This demon- strates that the classification performance increase with SMS is more significant than Po would suggest. This is because the SMS algorithm had a concomitant improve- ment in both the sensitivity and specificity results. Fi- nally, the class-wise accuracy (0.90 ± 0.01) for high (ω2) complexity software modules is an excellent result as one of the motivations for this classification system is to have high predictive power for problematic software modules (that is, those considered to be of high complexity). Table 4 lists the mean (with standard deviations) clas- sification performance results for DS3 with all classifiers using the validation subset of metrics, XV. Again, SMS produced better results than the best benchmark (SVM): a 12% increase in the κ score (0.58 ± 0.02 versus 0.52 ± 0.02) and a 4% increase for Po (0.81±0.01 versus 0.78 ± 0.01). It is interesting to note that the SMS Po values for DS3 and DS4 are similar, 0.86 and 0.81, respectively, (only a 6% difference) but the κ scores are sig nificantly Table 3. DS3 Validation Subset (XV) results with κ, PO, and class-wise accuracies. Accuracies Classifier Low (1) High (2) Po SMS 0.81 ± 0.010.90 ± 0.01 0.71 ± 0.02 0.86 ± 0.01 RBF 0.67 ± 0.030.81 ± 0.01 0.49 ± 0.03 0.75 ± 0.02 SVM 0.70 ± 0.010.84 ± 0.02 0.55 ± 0.02 0.78 ± 0.01 Table 4. DS3 Validation Subset (XV) results with κ, PO, and class-wise accuracies. Accuracies Classifier Low (1) High (2) Po SMS 0.86 ± 0.010.71 ± 0.02 0.58 ± 0.02 0.81 ± 0.01 RBF 0.80 ± 0.010.70 ± 0.02 0.49 ± 0.01 0.77 ± 0.01 SVM 0.81 ± 0.010.71 ± 0.02 0.52 ± 0.02 0.78 ± 0.01 different, 0.71 and 0.58, respectively (a 22% difference). This is evidenced by the poorer balance between the sen- sitivity and specificity in the case of DS4. Again, while Po suggests little difference in the classification per- formance results using the different datasets, κ strongly suggests that DS3 is a much better delineated dataset. 5.2. Predictive Software Metrics In the case of DS3, two software metric subsets were selected as being highly discriminatory. One subset used eight of the original software metrics: C1, C2, C3, C4, P1, P2, P3, and F1. The other subset was a pair-wise cross product of the metrics C1, C2, C3, and C4 with C1, C2, and C3. In the case of DS4, two metric subsets were also selected. The first used ten of the original software met- rics excluding only F2. The other metric subset was the squared values of all software metrics except F2 and F1. Both experiments demonstrate that quadratic combina- tions of these software metrics effect greater predictive power than any combination of the original software me- trics. 6. Conclusions The stochastic metric selection algorithm has been shown to be an effective strategy for the prediction of software module complexity, as expressed by module change count, given a set of software metrics. Compared to the conven- tional classification system benchmarks, this strategy produced significantly better overall classification results, as well as improved sensitivity and specificity, using only a subset of the original metrics. Finally, it was par- ticularly effective in predicting the occurrence of high Copyright © 2011 SciRes. JSEA
Mapping Software Metrics to Module Complexity: a Pattern Classification Approach431 complexity modules, which is an important characteristic for software project managers and program developers when maintaining and updating software modules. 7. Acknowledgements The author would like to thank A. Demko and C. Wiebe for their valuable contributions in implementing the sto- chastic metric selection algorithm in Scopira. The author would also like to thank M. R. Lyu for making the MIS dataset available for this pattern classification investiga- tion. This work was supported in part by the Natural Sci- ences and Engineering Research Council of Canada (NSERC). REFERENCES [1] S. R. Chidamber and C. F. Kemerer, “A Metrics Suite for Object-Oriented Design,” IEEE Transactions on Software Engineering, Vol. 20, No. 6, 1994, pp. 476-493. doi:10.1109/32.295895 [2] B. A. Kitchenham, R. T. Hughes and S. G. Kinkman, “Modeling Software Measurement Data,” IEEE Transac- tions on Software Engineering, Vol. 27, No. 9, 2001, pp. 788-804. doi:10.1109/32.950316 [3] N. E. Fenton and A. A. Kaposi, “Metrics and Software Structure,” Information and Software Technology, Vol. 29, No. 6, 1987, pp. 301-320. doi:10.1016/0950-5849(87)90029-2 [4] N. E. Fenton and M. Neil, “A Critique of Software Defect Prediction Models,” IEEE Transactions on Software En- gineering, Vol. 25, No. 5, 1999, pp. 675-689. doi:10.1109/32.815326 [5] G. Poels and G. Dedene, “Distance-Based Software Measurement: Necessary and Sufficient Properties for Software Measures,” Information and Software Technol- ogy, Vol. 42, No. 1, 2000, pp. 35-46. [6] E. J. Weyuker, “Evaluating Software Complexity Meas- ures,” IEEE Transactions on Software Engineering, Vol. 14, No. 9, 1988, pp. 1357-1365. doi:10.1109/32.6178 [7] M. Reformat, W. Pedrycz and N. J. Pizzi, “Software Quality Analysis with the Use of Computational Intelli- gence,” Information and Software Technology, Vol. 45, No. 7, 2003, pp. 405-417. doi:10.1016/S0950-5849(03)00012-0 [8] G. A. F. Seber, “Multivariate Observations,” John Wiley & Sons Ltd., Hoboken, 1984. doi:10.1002/9780470316641 [9] R. O. Duda, P. E. Hart and D. G. Stork, “Pattern Classifi- cation,” 2nd Edition, Wiley-Interscience, New York, 2000. [10] R. Gnanadesikan, “Methods for Statistical Data Analysis of Multivariate Observations,” 2nd Edition, John Wiley & Sons Ltd., New York, 1997. doi:10.1002/9781118032671 [11] N. J. Pizzi and W. Pedrycz, “Effective Classification Us- ing Feature Selection and Fuzzy Integration,” Fuzzy Sets and Systems, Vol. 159, No. 21, 2008, pp. 2859-2872. doi:10.1016/j.fss.2008.03.015 [12] M. R. Lyu, “Data Directory in the CD-ROM,” http://www.cse.cuhk.edu.hk/~lyu/book/reliability/data.ht ml [13] J. C. Munson and T. M. Khoshgofthaar, “Software Met- rics for Software Reliability Assessment,” In: M. R. Lyu, Ed., Handbook of Software Reliability Engineering, McGraw-Hill, New York, 1996, pp. 493-529. [14] R. Lind and K. Vairavan, “An Experimental Investigation of Software Metrics and Their Relationship to Software Development Effort,” IEEE Transactions on Software Engineering, Vol. 15, No. 5, 1989, pp. 649-653. doi:10.1109/32.24715 [15] R. S. Pressman and R. Pressman, “Software Engineering: A Practitioner’s Approach,” McGraw-Hill, New York, 2000. [16] N. E. Fenton and S. L. Pfleeger, “Software Metrics: A Rigorous and Practical Approach,” PWS Publishing, Boston, 1997. [17] J. F. Peters and W. Pedrycz, “Software Engineering: An Engineering Approach,” John Wiley & Sons Ltd., Hobo- ken, 1999. [18] M. H. Halstead, “Elements of Software Science,” Elsevier, New York, 1977. [19] T. J. McCabe, “A Complexity Metric,” IEEE Transac- tions on Software Engineering, Vol. 2, No. 4, 1976, pp. 308-320. doi:10.1109/TSE.1976.233837 [20] E. K. Tang, P. N. Suganthan and X. Yao, “Gene Selection Algorithms for Microarray Data Based on Least Square Support Vector Machine,” BMC Bioinformatics, Vol. 7, No. 95, 2006. doi:10.1186/1471-2105-7-95 [21] Q. Liu, A. Sung, Z. Chen and J. Xu, “Feature Mining and Pattern Classification for LSB Matching Steganography in Grayscale Images,” Pattern Recognition, Vol. 41, No. 1, 2008, pp. 56-66. [22] N. Kasabov and Q. Song, “DENFIS: Dynamic Evolving Neural-Fuzzy Inference System and Its Application for Time-Series Prediction,” IEEE Transactions on Fuzzy Systems, Vol. 10, No. 2, 2002, pp. 144-154. doi:10.1109/91.995117 [23] B. Schölkopf and A. J. Smola, “Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond,” MIT Press, Cambridge, 2002. [24] V. Vapnik, “Statistical Learning Theory,” John Wiley & Sons Ltd., New York, 1998. [25] L. Wang, “Support Vector Machines: Theory and Appli- cations,” Springer-Verlag, Berlin, 2005. [26] V. Vapnik and A. Lerner, “Pattern Recognition Using Generalized Portrait Method,” Automation and Remote Control, Vol. 24, 1963, pp. 774-780. [27] W. H. Press, S. A. Teukolsky, W. T. Vetterling and B. P. Flannery, “Numerical Recipes: The Art of Scientific Computing,” 3rd Edition, Cambridge University Press, Cambridge, 2007. [28] J. Moody and C. J. Darken, “Fast Learning Networks of Copyright © 2011 SciRes. JSEA
Mapping Software Metrics to Module Complexity: a Pattern Classification Approach Copyright © 2011 SciRes. JSEA 432 Locally–Tuned Processing Units,” Neural Computation, Vol. 1, No. 2, 1989, pp. 281-294. [29] S. M. Weiss and C. A. Kulikowski, “Computer Systems that Learn: Classification and Prediction Methods from Statistics, Neural Nets, Machine Learning and Expert Systems,” Morgan Kaufmann Publishing, San Mateo, 1991. [30] B. S. Everitt, “Moments of the Statistics Kappa and Weighted Kappa,” British Journal of Mathematical and Statistical Psychology, Vol. 21, 1968, pp. 97-103. [31] J. L. Fleiss, “Measuring Agreement between Two Judges on the Presence or Absence of a Trait,” Biometrics, Vol. 31, No. 3, 1975, pp. 651-659. doi:10.2307/2529549 [32] T. McGinn, P. C. Wyer, T. B. Newman, S. Keitz, R. Leipzig and G. Guyatt, “Tips for Learners of Evi- dence-Based Medicine: 3. Measures of Observer Vari- ability (Kappa Statistic),” Canadian Medical Association Journal, Vol. 171, No. 11, 2004, pp. 1369-1373. doi:10.1503/cmaj.1031981
|