Sunteți pe pagina 1din 6

A Comparative Analysis of Evolutionary

Algorithms and Mahalanobis-Taguchi System


for Breast Cancer Classification
Benjamin Daniels, Steven M. Corns, and Elizabeth A. Cudney
Abstract – Classification for medical diagnosis is an important group) to that of an unknown set of values (abnormal group).
problem in the field of pattern recognition. We introduce a new It has been successfully used in a broad range of cases, largely
method for classification based on repeated analysis of due to its ability to identify inter-variable changes in data.
information tailored to small data sets – the Rote Evolutionary Also, because the MD is measured in terms of standard
Algorithm. Using the Wisconsin Breast Cancer study, this
deviations from the mean of the samples, it provides a
method was compared to three other methods of classification:
Mahalanobis-Taguchi Systems, Finite State Classifiers, and statistical measure of how well an unknown sample matches a
Neural Networks. It was found that for the given data set, the known sample set.
Rote Evolutionary Algorithm outperformed the other methods of Finite state classifiers are Finite State Machines (FSMs)
classification. This new algorithm correctly classified over 92% of where a vote to classify the data set is performed in each state
the data set. based on the data being input to drive the state machine [3, 4].
The input being analyzed by the FSM is used to drive the
I. INTRODUCTION machine between states, gathering information from the states
In the field of computational intelligence, it has been as it progresses. The FSC were created using an evolutionary
demonstrated that no problem solving method dominates all algorithm that treated the state descriptions as an array which
others in all problems [1]. In this paper we compare three was used as the genome representation for the problem. These
different prediction methods for breast cancer. The first algorithms have been shown to be useful as classifiers on
method is an evolutionary algorithm (EA) developed to Polymerase Chain Reaction (PCR) primer classification [5, 6]
classify breast cancer patient referred to as a Rote and image classification [7], which contain a full description
Evolutionary Algorithm (REA). This method is compared to of how these classifiers are constructed and implemented.
two other methods for data classification: the Mahalanobis- The data set used for this analysis is the Wisconsin Breast
Taguchi System (MTS) [2] and finite state classifiers (FSC) Cancer study, consisting of nine attributes and one class
[3]. Both of these methods have been successfully applied to (positive or negative for cancer). The classification is
other classification problems, and therefore are appropriate performed based on the values of these nine separate
methods to compare and contrast for accuracy. attributes, with values ranging from 1 to 10. The nine
The Rote Evolutionary Algorithm is an EA developed to attributes taken under consideration are:
perform classification on data sets where a relatively small
amount of training data is available. It is designed to address 1. Clump Thickness
problems in which the input parameters consist of discrete 2. Uniformity of Cell Size
points in one or more bounded ranges of known size. The 3. Uniformity of Cell Shape
problem input parameters for the REA method consist of an 4. Marginal Adhesion
arbitrary number of ranges to be evaluated for a Boolean 5. Single Epithelial Cell Size
output. The REA method is particularly adept at training itself 6. Bare Nuclei
well in cases where only very small training populations exist. 7. Bland Chromatin
The REA method was created specifically for evolving breast 8. Normal Nucleoli
cancer prediction, although it should have equivalent 9. Mitoses
performance with similar problems.
The Mahalanobis-Taguchi System (MTS) was developed More information on these attributes can be found at the
by Genichi Taguchi to provide a means to establish a reference University of California Irvine Machine Learning Repository
group and a means to define a measure the degree of Website [8].
abnormality of individual observations when using the Evolutionary algorithms are generally an appropriate
Mahalanobis distance (MD) to determine class membership method for this type of problem; however, there are typically
[2]. Mahalanobis distance is a discriminant analysis approach many unique ways to apply EA operators to a given problem.
that uses a distance measure based on correlations between In the course of this paper, a new approach is taken and then
variables and the different patterns that can be identified and compared with the use of finite-state classifiers, and the
analyzed with respect to a reference population. MD is used Mahalanobis-Taguchi System. For additional consideration,
to determine the similarity of a known set of values (normal results for a neural network approach [9] are also considered.
1

corresponding value in C1 (Chromosome 1). In this case, it is


II. METHODOLOGY ‘greater than’. Chromosome 2 has a value of false. This in
conjunction with the fact that the patient’s value is greater than
A) Basic Approach the C1 value means that there is a negative diagnosis on this
The representation of the REA method is separated into indicator for this patient on this solution. Since the diagnosis is
four chromosomes to which the traditional EA operators negative, we look at the weighting coefficient for this indicator
(parent selection, recombination, mutation, and natural when the diagnosis is negative. The value is 8, so -8 is the
selection) are applied. The information contained in these four output of this indicator for the patient.
chromosomes is as follows: This is performed for each of the nine attributes, and then
the results are summed. If the value is positive, then the
• Chromosome 1 contains an integer value from 1-10 prediction is positive for cancer, otherwise the prediction is
which is compared with the patient’s test results. negative.
Each test will be compared to a unique location in the
chromosome. These values will be compared to B) Sensitivity Function
determine which is greater to drive the classification. The above example uses a single point which is chosen in
• Chromosome 2 contains a Boolean operator, or a each range, provided by Chromosome 1. However, there is
value which can be converted into a Boolean. If true, one final main addition to the REA method: in order to
the hypothesis is that the patient has cancer if the test achieve more ‘sensitivity,’ or ‘granularity,’ multiple points
score lies above the number in Chromosome 1, and within each range are examined simultaneously. This requires
does not if otherwise. If false, the hypothesis is that adding four more chromosomes for each increment in
the patient has cancer if the test score lies below the sensitivity. For example, if the sensitivity factor is 2, then two
number in Chromosome 1, and does not if otherwise. points on each range of each indicator are chosen to evaluate.
• Chromosome 3 holds the weighting factor or number Figures 1 and 2 illustrate how this sensitivity factor is used in
of points earned towards a positive diagnosis. It is evaluating the attribute values. There are a total of eight
counted for a patient only if Chromosomes 1 and 2 chromosomes used for this representation. The range for
point to a positive diagnosis. clump thickness is now split twice and each of the two
• Chromosome 4 holds the weighting factor or number corresponding chromosome sets are evaluated independently.
of points earned towards a negative diagnosis. It is The sums of each are then added for the final diagnosis
counted for a patient only if Chromosomes 1 and 2 decision. For simplicity, the sensitivity factor is chosen at the
negate a positive diagnosis. onset of a simulation for the REA method; all indicators of all
solutions will have the same sensitivity factor. The sensitivity
To better understand this approach, consider the following factor may be of any arbitrary size, but as the sensitivity factor
example in Table 1. The test results for a single patient are increases, marginal improvement in EA effectiveness per
evaluated by the chromosomes of a single potential solution. increase of the sensitivity factor reasonably approaches zero.
The patient’s test scores are on the left. The information in the
C) Application of EA
four chromosomes determines how this data is evaluated.
For this method, truncation selection is used as the natural
Finally, the results for each indicator are summed for a final
selection operator. As it is known that the selective pressure
determination. A negative number indicates no cancer, while a
for truncation selection is extremely high, the parent selection
positive one indicates a positive diagnosis. To evaluate the
method used in the REA method was chosen to be random,
first attribute (Clump Thickness), it is first compared to the
patient’s value for this indicator with the

Table 1: Example evaluation of breast cancer patient information using REA. A negative sum indicates no cancer predicted.
2

since no further selective pressure seems likely to be needed.


An n-point recombination operator was selected for crossover,
with the number of points selected by the user before a run.
All subsequent runs will then utilize the same n-point value.
Mutation occurs by randomly selecting a child, then randomly
Figure 2, Information contained by Chromosome 1for clump thickness, with a
selecting a chromosome, and mutating it to a random value
sensitivity factor of 2.
within the valid range of possible solutions. There will be a
number of mutations equal to the value input by the user for The training data was broken into two parts: 485 for
‘Mutation Rate’ when the simulation is started. A population training and 199 for evaluation. The initial comparisons are
size of 10 was used for the REA. conducted with previous work, so to compare the performance
of the REA method with the other methods it was necessary to
D) Fitness Evaluation train the data on the data set sizes of 20, 30, 50, and 100. All
Type-I (false positives) and type-II error (false negatives) used data points were chosen at random from the training data
are distinguished within this method. Specifically, type-II set of 485. A random sample set was chosen at the beginning
error is prioritized to decrease the likelihood that a patient with of each run. For each training data size, the REA method was
a malignant growth will be diagnosed with a benign growth. tested on 10 runs at seven different levels of sensitivity, and
While a false positive for cancer may be stressful for the the sensitivity levels used are 1, 2, 3, 4, 5, 7, 10, and 30.
patient, it is less likely to put the patient at risk than an The REA method had mediocre classification accuracy
instance of type-II error. Type II errors were discouraged when a data set of 20 was used for sensitivities of one to four
through a fitness weighting scheme. The fitness weighting (fig. 3), with classification accuracy typically of 83-90%.
criteria are as follows: When the sensitivity was increased to five or higher, the
algorithm classification accuracy increased to 95% or better.
• A correct positive is worth +4 As the training data set size increased, this trend continued,
• A correct negative is worth +3 although the classification accuracy was always above 92%
• An incorrect positive is worth -2 and usually above 95%. For most data set sizes, the accuracy
• An incorrect negative is worth -6 continues to improve for three of the five data sets. This may
indicate the beginning of overtraining on the data set.
To determine fitness, each solution in the Plus (parents The type I and type II errors were also evaluated to
plus children) survival population is evaluated using data from determine how well the algorithm classified the two possible
500 actual patients. The prediction made by the solution is outcomes. The type I errors were calculated by dividing the
compared with the real diagnosis and fitness points are number of false positives by the total number of biopsies that
awarded to solutions accordingly based on predictive were negative (fig. 4). The type II errors were calculated by
accuracy. dividing the number of false positives by the number of
biopsies found positive for cancer (fig. 5). Comparing the
III. RESULTS results, it can be seen that there is a distinct bias against type II
The REA method was first tested for breast cancer errors, with only about one percent of errors falling in this
prediction accuracy at various levels of its sensitivity function. category for all positive biopsy results.
It was then tested at its highest level of sensitivity and
compared versus other prediction methods.

A) Training
The data set used for this project is the breast cancer data
from the UCI machine-learning repository, which was
collected at the University of Wisconsin by W. H. Wolberg
[10]. They used this data to predict whether a tissue sample
taken from a patient’s breast is malignant or benign. There are
one class, nine numerical attributes, and 699 observations.
Sixteen instances contain a single missing attribute value and
were removed from the analysis [11].

Figure 1, Information contained by Chromosome 1for clump thickness, with a Figure 3, total number of errors vs. sensitivity for the REA approach with
sensitivity factor of 1. training set sizes of 20, 30, 50, 100, and 500.
3

Figure 6, Comparison of correct classifications vs. training data size for REA,
Figure 4, Percentage of type I errors vs. sensitivity for the REA approach with
MTS, and NN.
training set sizes of 20, 30, 50, 100, and 500.

Table 2 – Results for Mahalanobis - Taguchi System [11]

Figure 5, Percentage of type II errors vs. sensitivity for the REA approach Table 3 – Results from Neural Network Classification [11]
with training set sizes of 20, 30, 50, 100, and 500.

The best performing sensitivity level was then compared


against Mahalanobis-Taguchi System and the neural net
methods. The information for this comparison was extracted
from previous work [11] in which the Mahalanobis-Taguchi
system was compared to a neural network approach. Data
from exactly ten runs was collected and averaged, and the
results plotted and compared with ten runs of the neural
network and ten applications of the Mahalanobis-Taguchi
system. Figure six shows the comparison between these
methods. The data collected from each of the experiments here
are given in tables 2, 3, and 4.
4

Table 4 – Results for the REA method The best run of 10 runs with a sensitivity coefficient of 30
data20 data30 data50 data100 from the REA method was compared. The two EAs were
compared on training data set size of 20, 30, 50, 100, and 485.
1 0.974 0.924 0.936 0.985
Figure 7 shows that the REA method had a better success rate
2 0.947 0.990 0.952 0.985 on classification of the data for all sizes of training data.
3 0.842 0.947 0.900 0.990
IV. DISCUSSION
4 0.995 0.985 0.969 0.985 A) Sensitivity Factor and Performance
5 0.856 0.963 0.990 0.936 As expected, higher levels of sensitivity performed better in
preliminary testing before comparison with other methods. As
6 0.888 0.930 0.990 0.952
the sensitivity increased to 10, error distribution between the
7 0.849 0.952 0.888 0.936 different population sizes stabilized and generally decreased.
8 0.990 1.000 0.958 0.969 Since a sensitivity factor of 30 seemed to be the highest
performing, it was chosen to be compared with the other
9 0.969 0.979 0.947 0.963
methods. It performed well at smaller population sizes versus
10 0.906 0.909 0.990 0.974 the Mahalanobis-Taguchi System and neural network
REA 0.921 0.958 0.952 0.967 methods, and slightly better at larger population sizes. The
REA method dominated the other methods on all data set
The next step was to compare the REA method with a sizes.
finite-state classifier (FSC) method. The FSC used for the It is interesting to note that overall performance of the
comparison were the best classifier from 100 different runs, REA method actually decreased slightly as population size
and was constructed in a similar manner to that used in increased, and then began to increase again. It is possible that
previous work [7]. Each state had am attribute number, a real this is due to randomness in chosen population members. It is
value assigned to it, and a comparison operator. When also possible that this may be due to some effect of the high
entering the state, the value stored in the state was compared sensitivity level used with the REA method for the
to the value of the attribute specified in the state. Based on comparison. Additional experiments with a larger number of
whether the statement was true or false, a response instructed runs will be necessary to evaluate this fully.
the state incremented, decremented or in some cases ignored a Although a sensitivity factor of 30 was the highest used,
running counter. When the information was fully evaluated, a the sensitivity factor for the REA method may be set
positive result indicated the growth was malignant and a arbitrarily high. Results have shown that as the sensitivity
negative valued indicated benign. 64 states were used for each factor is increased, general accuracy of the EA is also
of the FSC, and they were evolved using a generational increased, at all training sizes. The downside is computation
algorithm using two point crossover and a mutation operator time; the EA requires more computation time as the sensitivity
that changed the initial state (10%), the comparison operator factor is increased. There was no study done to determine
(10%), the transition destination (20%), the response (10%), or what this rate of increase in computation time is, or whether it
the value used to make the comparison (50%). This algorithm is linear or exponential. Also, the marginal benefit yielded by
was run for 1000 generations using training data set sizes of an increase of the sensitivity factor should eventually approach
20, 30, 50, 100, and 485. zero. In terms of general computation time, the Mahalanobis-
Taguchi System has the lowest; it is virtually instantaneous.
The REA method had much better computation times than the
FSC, which required a wall clock time of approximately one
hour.

B) Further Study
In order to further investigate the performance of the REA
method with larger training populations, the REA method was
trained with a population size of 485 observations, the
maximum possible. At this level, all sensitivity levels
discovered better solutions. The best solution found had a 99%
success rate, with zero instances of type-II error, and only two
instances of type-I error after testing on 199 observations.

C) Possible Sources of Error in Comparison


The first possible source of error comes from the fact that
sample size (ten runs) used to compare the Mahalanobis-
Figure 7, Comparison of correct classifications vs. training data size for REA, Taguchi System, neural network method, and the REA method
and FSC is quite small. For this paper, it was necessary to use ten runs
5

because only ten runs were performed for the Mahalanobis- It may be useful to look for other applications where the
Taguchi System and neural network prediction methods. As a REA method might be successful. It may be beneficial to
rule of thumb, it is typically advisable to have at least 30 investigate further the use of different levels of sensitivity in
samples for comparisons of the mean between different the training of the REA method, or to further refine the
populations. It may also be possible that variation from the technique. It would also be advantageous to apply this method
sampling of the data sets has introduced some uncertainty in to other data sets to gain more information on the benefits of
the results, as there is no available data on the particular the weighting scheme here and to explore any potential areas
portion of the data set the Mahalonbis-Taguchi System and the where overtraining would be a serious issue.
neural network approaches used. The problem solving techniques which the REA method
Another possible source of error is due to the distinction was compared against are generally good at a wide variety of
which the REA method makes between type-I and type-II problems. It could be valuable to compare the REA method
error. Unlike the other methods to which the REA method was against other problem solving techniques which are specially
compared, the REA method penalizes type-II error formulated for range searching, such as simplex, halfspace, or
significantly. This may lead to more total errors in a solution if orthogonal range searching methods.
it is necessary to have more instances of type-I error in order
to avoid type-II errors with a greater success rate. REFERENCES
One final possible source of error in comparison with the
Mahalanobis-Taguchi and neural net methods is in the size of [1] Wolpert, D.H.; Macready, W.G., "No free lunch theorems for
the pool of training observations used for creating a training optimization," Evolutionary Computation, IEEE Transactions on , vol.1,
no.1, pp.67-82, Apr 1997
population versus that used for testing. Although the same data [2] Taguchi, G and R. Jugulum, “New Trends in Multivariate Diagnosis”,
was used by all methods compared, the other methods may Indian Journal of Statistics, 62, Series B, 2 233-248 (2000).
have allocated the use of data differently. For example, the [3] Stanley, E. A., Ashlock, D. A. and Tesfatsion, L., “Iterated prisoner's
Mahalanobis-Taguchi System analyzes the data directly and dilemma with choice and refusal,” Artificial Life III, vol. 17 of Santa Fe
Institute Studies in the Sciences of Complexity, edited by Christopher
does not have a training step. Langton, Reading, 1994, Addison-Wesley, pp. 131-176.
[4] Fogel, D. B., “Evolving behaviors in the iterated prisoner’s dilemma”.
V. CONCLUSIONS Evolutionary Computation, Vol. 1, No. 1, 1993.
The REA method was extremely successful at solving this [5] Ashlock, D. A., Bryden, K. M., Corns, S. M., Schnable, P. S., and Wen,
T.J., “Training Finite State Classifiers to Improve PCR Primer Design,”
problem, and may be capable of regularly out-performing 10th Annual AIAA/ISSMO Multidisciplinary Analysis and Optimization
other known methods at solving range searching problems. Conference, Albany, NY, 2004.
The REA method is very efficient in terms of computation [6] Yadav, S. and Corns, S.M., “Improved PCR Design for Mouse DNA by
time, with very low training time compared with the FSC. The Training Finite State Machines,” 2010 IEEE Symposium on
Computational Intelligence in Bioinformatics and Computational
REA method dominated all other prediction methods at all Biology, pages 249-254, 2010.
levels of population size. [7] Cudney, E.A.; Corns, S.M.; , "A comparison of Finite State Classifier
Repetition within the algorithm on the attributes of the and Mahalanobis-Taguchi System for multivariate pattern recognition in
data set allows the algorithm more opportunity to fine tune the skin cancer detection," Computational Intelligence in Bioinformatics and
Computational Biology (CIBCB), 2011 IEEE Symposium on , vol., no.,
weighting of the different attributes. This allows for a gradual pp.1-7, 11-15 April 2011
build up of weights and makes it more likely that [8] University of California Irvine Machine Learning Repository, Wisconsin
complimentary combinations of information can be found. Breast Cancer Study, http://archive.ics.uci.edu/ml/datasets/Breast+
There is a danger in this approach, in that it seems that over Cancer+Wisconsin+%28Original%29, last accessed 12/3/2011.
[9] Jugulum, R., and Monplaisir, L. “Comparison between Mahalanobis-
training of the algorithm could easily be reached. However, in Taguchi- System and Artificial Neural Networks”, Journal of
this test case there was no evidence of overtraining. QualityEngineering Society, Vol. 10, No.1, pp.60-73, 2002.
The outcome of the REA method may be promising, but [10] Woodall W. H., Koulelik R., Tsui K. L., Kim S. B., Stoumbos Z. G.,
the data set used has a limited number of data points. The Carvounis C. P. “A review and analysis of the Mahalanobis”,
Technometrics, 45(1); 1-30. 2003.
attributes used are those that medical professionals consider to [11] Cudney E. A., Hong J., Jugulum R., Paryani K., Ragsdell R. M., Taguchi
be the leading indicators of whether a tumor is malignant or G. “An Evaluation of Mahalanobis-Taguchi System and Neural Network
benign. Nine variables with a value of 0-9 gives a search space for Multivariate Pattern Recognition”, Journal of Industrial and Systems
with 109 (one billion) possible combinations, which would still Engineering, Vol. 1, No. 2, pp 139-150, 2007.
be viewed as a test problem as the size does not warrant a
heuristic approach. Future work could include the
development of a nine dimensional response surface that may
be used to indicate combinations of variables that indicate
positive results for malignant growths.

S-ar putea să vă placă și