Upload
maen
View
216
Download
0
Embed Size (px)
Citation preview
The Automatic Identification of Melanoma by Wavelet and Curvelet Analysis:
Study Based on Neural Network Classification
Md. Khalad Abu Mahmoud Adel Al-Jumaily
School of Electrical, Mechanical and Mechatronic Systems
Faculty of Engineering
School of Electrical, Mechanical and Mechatronic Systems
Faculty of Engineering
University of Technology, Sydney University of Technology, Sydney
PO Box 123 Broadway NSW 2007 Australia PO Box 123 Broadway NSW 2007 Australia
mdkhaled.abumahmoud @ student.uts.edu.au [email protected]
Maen Takruri
School of Electrical, Mechanical and Mechatronic Systems
Faculty of Engineering
University of Technology, Sydney
PO Box 123 Broadway NSW 2007 Australia
Abstract—this paper proposes an automatic skin cancer
(melanoma) classification system. The input for the prosed
system is a collected data images, it followed by different image
processing procedures to enhance the image properties. Two
segmentation methods used to identify the normal skin cancer
from malignant skin and to extract the useful information
from these images that passed to the classifier for training and
testing. The features used for classification is the coefficients
created by Wavelet decompositions and simple wrapper
curvelet. Curvelet is suitable for the image that contains
oriented texture and cartoon edges. Recognition accuracy of
the three layers back-propagation neural network classifier
with wavelet is 51.1% and with curvelet is 75. 6% in digital
images database.
Keywords: segmentation; Wavelet decompositions; simple
wrapper Curvelet; Back-Propagation Neural Network (BNN)
Classifier.
I. INTRODUCTION
Australia has one of the highest skin cancer rates in the
world at nearly four times the rates in Canada, the US and
the UK. it has been estimated 115,000 new cases of cancer
diagnosed and more than 43,000 people are expected to die
from cancer according to Cancer council of Australian 2010
[1], The Chair of Public Health Committee pronounced that,
more than 430,000 cases treated for non-melanoma, and
more than 10,300 people are treated for melanoma, with
1430 people dying each year[1].
The skin cancer malignant melanoma is the deadliest form of
skin cancer. It can be removed by simple surgery if it has not
entered the blood stream. Melanoma can be recovered if
diagnosed and treated in early stages. Early diagnosis is
obviously dependent upon patient attention and accurate
assessment by a medical practitioner. Several published
classification systems show accuracy rates ranging from
60%- 92% [2] which coincides with the estimated rates
obtained by general practitioners. These techniques were
aiming to be able to provide recommendation for non-
specialized users. But the variations of diagnosis are
sufficiency large and there are lacks of detail of the test
methods. One commercial product, Solar Scan by Polar
technics, has an accuracy rate of 92%. Solar Scan is a
complex system, taking high quality Epiluminescence Light
Microscopy (ELM) images and using advanced image
analysis techniques to extract a number of features for
classification which it make not suitable for normal person
usage . The traditional imaging is just a recording of what
the human eye can see by digital camera while the
Dermoscopy known as Epiluminescence Light Microscopy
(ELM) images need a professional experience to get the
required image.
Different approaches have been established to improve
image processing result. The image processing for skin
cancer detection consists of 6 main steps namely; (1) image
acquisition step which used to collect database from different
resources, (2) pre-processing to remove the fine hairs, noise
and air bubbles, (3) Post-processing to enhance the shape of
the image, (4) Segmentation to remove the healthy skin from
the image and find the region of interest, usually the cancer
cell that remains in the image (5) Feature extraction which
extracts the useful information or image properties from the
segmented images (6) classification system to classify the
image if it cancer or not. The classification system is usually
supported by an intelligent classifier, such as neural network
and support vector machine. The element of all these ideas
will extend in the following sections.
680978-1-4577-2152-6/11/$26.00 c©2011 IEEE
Our previous work [3] has studied the automatically early
detection of skin cancer detection based on 3-layers back-
propagation and auto-associative neural network
classification. The experiment has conducted with both
methods for better performance and accuracy. The wavelet
based feature extraction method had used in this comparison.
The result showed that back-propagation neural network has
better accuracy than that of auto-associative neural network.
The use of wavelet decomposition provides useful extracted
feature for image processing but it cannot develop the
regularity of the edge curves [4, 5]. This is becoming one of
the major short comings for feature extraction and thus
finding of better feature extracting method while maintaining
the 3-layers back-propagation neural network as a
classification method is one of the main interests in this
work. This paper include: The different image processing
algorithms for skin cancer detection, which are explained in
section 2, 3 and 4. The segmentation methods have been
clarified in section 5. The feature extraction approach for
wavelet and curvelet transforms is provided in section 6.
Back-Propagation Neural Network (BNN) in section 7.
Classification results are presented in section 8, and finally
the conclusion is made in section 9.
II. IMAGE ACQUISITION
The database images we used, contains both digital photo
and dermoscopy images. These images collected from
Sydney Melanoma Diagnostic Centre in Royal Prince Alfred
Hospital and internet websites. Total 448 images are
collected; these images include digital malignant melanoma
images and digital benign melanoma images. These images
separated into four groups to test the accuracy of each group.
The database images are obtained from different sources and
the size of the images is non-standard. The first step in the
process is to resize the image to have a fixed width (360) but
variable size of height. The second step is to remove the
background noise from the pictures to increase the accuracy
of the detected images.
III. IMAGE PRE-PROCESSING
In order to remove unwanted features, firstly some pre-
processing has applied to remove the fine hairs, noise and air
bubbles on the skin and facilitating image segmentation by
using wiener and median filters. Wiener filter purposed to
reduce the amount of noise presented in a signal by
comparison with an estimation of the desired noiseless
signal. It work based on neighbor hooding pixels of size m-
by-n to estimate the local image mean and standard
deviation. The additive noise (Gaussian white noise) power
is assumed to be noise [6]. The median filter is one of the
most common smoothing filters in the literature. Median
filtering with a mask of appropriate size can eliminate most
of the artifacts in a dermoscopy image [7]. Median filtering
is a standard nonlinear signal processing technique
developed by Turkey [8] for suppressing the noise in the
image by removing the outliers that are the extreme pixel
values.
IV. IMAGE POST-PROCESSING
The post- processing enhances the shape of the image. In
this work, Histogram equalization (HE) algorithm was
utilized. It is a smart contrast enhancement technique which
stretches the dynamic range of the image histogram and
resulting in overall contrast improvement. Fig. 1 shows the
result of segmentation by threshold with Gray scale
Histogram Equalization (GHE) (bottom right) and Local
Histogram Equalization (LHE). The LHE (bottom left)
sharpen the lesion but also reduce the surrounding detail.
V. IMAGE SEGMENTATION
Segmentation means the process of partitioning a digital
image into multiple segments (sets of pixels). It is used to
locate objects and boundaries (lines, curves, etc.) in images.
In this paper, two segmentation methods have been
implemented, thresholding Region of Interest (ROI) and
Statistical Region Merging (SRM). It work based on the
information from the clinical result that noticed lesion is
usually darker than the human skin.
A. Segmentation by Region of Interest (ROI)
Region of interest (ROI) is extracted by segmentation.
The interesting features of melanoma are included within the
border since most of the cancer cells are lump structures. The
border structure provides vital information for accurate
diagnosis. Many clinical features including asymmetry and
border irregularity are calculated from the border. Threshold
is an advanced method for segmentation to compute the
intensity value from grey images. Fig. 2 shows the result of
threshold segmentation where the white color represents the
area of interest (ROI).
B. Segmentation by Region Growing (SRM)
Statistical region merging (SRM) algorithm is based on
region growing and merging. The performance of SRM has a
better segmentation result compare to other popular
segmentation methods [9]. The theory of region merging
starts at a seed point and compares its four neighbor points or
pixels. The region is grown when the neighbors pixel share
the same homogenous properties. SRM usually works with a
statistical test to decide the merging of regions. Fig. 3 shows
the result of SRM segmentation where white color represents
the area of interest, which more accurate in segmentation. It
includes more useful areas and does not put the outer region
as part of the area of interest.
VI. FEATURE EXTRACTION
Feature extraction is the procedure that takes out the
hidden properties or the raw data from the image that is
2011 11th International Conference on Hybrid Intelligent Systems (HIS) 681
further used in the classifier. In this section we look at a
modern method of feature extraction, wavelet, and curvelet
transformation which are discussed and used in this paper.
Thus approaches, such as discrete wavelet and discrete
curvelet, transforms the frequency information at different
scales and locations are obtained [10].
Fig 1. Result of segmentation by threshold with GHE (bottom right) and
LHE (bottom left), Top: shown histogram Local and Global results
Fig2.Segmentation by Thresholding
(ROI) Fig 3. Segmentation by SRM
A. Wavelet Transform
Wavelet analysis uses the terms approximations and
details. Approximations are the high scale low frequency
components while details are the low scale high frequency
components. Approximations are what give the signal its
identity. Wavelet decomposition can be a repeated process
occurring over several levels. The original signals are
decomposed into approximations feeding the next level of
decomposition, creating a decomposition tree. The approach
of this feature extraction is similar to our previous work in[3]
, the image energy is distributed according to the resolution.
Each of the nodes is represent one feature, which then can be
used as an input for classification stage. That paper
mentioned the second level decomposition can generate 16
nodes or features. Two-dimensional wavelet packet returns
the coefficients in 2 dimensions matrix. The features are
calculated by their mean, maximum, minimum, median,
standard deviation and variance. Therefore, 144 features for
each image (16 nodes x 9 features) for 448 images are
produced to be input for classification system. This would
increase the computation time without any improvement to
the performance [3]. Therefore Principal Component
Analysis (PCA) is suggested to be used before training
process (by chosen PCA factor as 0.002, it reduce the 144
features into 60 features). Fig 4. Shown the original image
and its transform (wavelet coefficients). This method allow
useful cancer’s feature to be extracted from the images
without clinical knowledge. In this system, 2-D wavelet
packet is used to enhance the image in grey scaled as an
input.
Fig 4. Display the image and its transform (wavelet coefficients)
Unfortunately, wavelet ignores the geometric properties of
objects with edges in higher dimensions and do not exploit
the regularity of the edge curves [4, 5] which lead to
unsatisfying results in the existing wavelet. To prevent edges
to be smooth, the curvelet transform is proposed to solve this
problem.
B. Curvelet Transform
In the last decade, there were a lot of activities in the
development of new mathematical and computational tools based
on multiscale ideas permeated in fields of science and
technology. A special member of this emerging family of
multiscale geometric transforms is the curvelet transform [11-
13], which was developed in the last few years in an attempt to
overcome inherent limitations of traditional multiscale
representations such as wavelets. As explained in [3], curvelets
are waveforms which are highly anisotropic at fine scales, with
effective support obeying the parabolic principle (length
width2). Just as for wavelets, there is both a continuous and a
discrete curvelet transform. Fig. 5 Showmen a single curvelet and
Fig. 6 displays representation of this joint localization.
Fig. 5: A single curvelet with width and length
682 2011 11th International Conference on Hybrid Intelligent Systems (HIS)
Conceptually, the curvelet transform is a multi-scale non-
standard pyramid with many directions and positions at each
length scale, and at high scales, the curvelet waveform
becomes so fine that it looks like a needle shaped element
(left images of Fig. 7 (b). Whereas, the curvelet is non
directional at the coarsest scale (left image of Fig. 7(a)) [14].
With increase in the resolution level the curvelet becomes
finer and smaller in the spatial domain and shows more
sensitivity to curved edges which enables it to effectively
capture the curves in an image (Fig.7).
Figure 6: Schematic representation of the support of a curvelet in both
space and frequency.
(a)
(b)
Fig. 7: Curvelets (absolute value) at scales different scales at a single
direction are shown in the spatial domain (left) and in the frequency domain
(right).
Curvelets have useful geometric features that set them
apart from wavelets and the likes. If we combine the
frequency responses of curvelets at different scales and
orientations, we get a rectangular frequency tiling that covers
the whole image in the spectral domain as shown in Fig 8.
Thus, the curvelet spectra completely cover the frequency
plane and there is no loss of spectral information like the
Gabor filters.
Fig 8. Rectangular frequency (basic digital) tiling of an image with 5 level
curvelet
To achieve higher level of efficiency, curvelet transform is
usually implemented in the frequency domain. The product
is then inverse Fourier transformed to obtain the curvelet
coefficients. The process can be described as:
Curvelet transform = IFFT [FFT (Curvelet) x FFT (Image)
And the product from the multiplication is a wedge.
The trapezoidal wedge in the spectral domain is not suitable
for use with the inverse Fourier transform which is the next
step in collecting the curvelet coefficients using IFFT.
The wedge data cannot be accommodated directly into a
rectangle of size 2 j 2 j / 2. To overcome this problem, The
researcher in [14] have formulated a wedge wrapping
procedure , where a parallelogram with sides 2 j and 2 j / 2 is
chosen as a support to the wedge data. (Fig. 9)
Fig. 9: Wrapping wedge around the origin by periodic tiling of the wedge
data.The angle is in the range ( / 4, 3 / 4).
The wrapping is done by periodic tiling of the spectrum
inside the wedge and then collecting the rectangular
coefficient area in the center. The center rectangle of size 2 j
× 2 j / 2 successfully collect all the information in that
parallelogram (Fig. 9).
The researcher in [10] defined the product of above equation
is a wedge. The trapezoidal wedge in the spectral domain is
not suitable for use with the inverse Fourier transform which
is the next step in collecting the curvelet coefficients using
IFFT. Researchers in [15] Have shown that wrapping based
fast discrete curvelet transform is much more efficient and
2011 11th International Conference on Hybrid Intelligent Systems (HIS) 683
provides better transform result than ridgelet based curvelet
transform.
(a)
(b)
Fig. 10: Fast discrete curvelet transform to generate curvelet coefficients.
In the spatial domain, a curvelet has an envelope strongly
aligned along a specified `ridge' while in the frequency
domain; it is supported near a box whose orientation is
aligned with the co-direction of the ridge. (Note that these
formulae allow us to analyze and synthesize arbitrary
functions in ( ) as a superposition of curvelets in a
stable and concrete way. [16] This micro local behavior
turns out to be key to understanding the properties of
curvelet-propagation. [5].
The feature extraction of curvelet has done with Matlab
software and the code consists of two parts. The first part
computes the texture features from the curvelet coefficients
of each image in the database. The second part returns the
curvelet coefficients of a selected image to the first part.
Curvelet transform returns the set of curvelet coefficients
indexed by scale, orientation and location parameters.
Calculate feature vector for each texture image in the
database and insert those in order to another file. Then the
database images are indexed in the feature database to be
ready to calculate the mean and standard deviation of these
coefficients. (Fig. 10)
Computation of curvelet coefficients from image function is
used to calculate the coefficients at the different scales and
orientations of an image. Basically, we use the curvelet
coefficients in the cells for feature extraction. Fast Discrete
Curvelet Transform via wedge wrapping used to find the
features.
For this wedge wrapping process, this approach of curvelet
transform is known as the ‘wrapping based curvelet
transform. The wrapping is illustrated in Fig. 10(b) and
explained as following. As shown in Fig. 10, in order to
do IFFT on the FT wedge, the wedge have to be arranged as
a rectangle? The idea is to replicate the wedge on a 2-D grid,
so a rectangle in the center captures all the components a, b,
and c of the wedge.
We have to calculate the mean and standard deviation for the
first half of the total cells at each scale. The test on
melanoma image using curvelet method is shown in Fig 11.
Fig 11. Curvelet denoising image that contains oriented texture and
cartoon edges
(a) Original (b) Noisy, and (c) Curvelet
VII. BACK-PROPAGATION NEURAL NETWORK (BNN)
Wide ranges of classifiers are available and each one of them
has its strengths and weaknesses. Classifier performance
depends on the characteristics of the data to be classified and
there is no single classifier that works best on all given
problems [3]. In this work, back-propagation neural network
(BNN) has utilized as neural classifier as in our previous
work [3] fig. 12 has proven that BNN has better performance
and accuracy.
VIII. COMPARISON AND RESULT
In our experiment, we used 60% of the total images for
Training, 20% for testing and 20% for validation. The best
result with highest overall accuracy is 70.44% for curvelet
with BNN with three layers [40 25 10 neurons] as shown in
table 1. The accuracy for neural network as a classifier is
684 2011 11th International Conference on Hybrid Intelligent Systems (HIS)
Figure 12: A Multilayer feed-forward network
increased with the number of neuron in the hidden layer.
However, the number of hidden layers could reduce the
probability of over-fitting; it found that the curvelet method
has 14.5% higher accuracy than wavelet in the classification
test as illustrated in table 1.
Method No of Neurons
No of Layers
Training Validation Testing
Wavelet 40 25 10 3 76.29 55.67 55.94
Curvelet 40 25 10 3 79.76 68.06 70.44
Table l. Displayed the comparison between wavelet and carvelet features results.
IX. Conclusion
Building an image classification system for early skin
cancer detection includes different stages: pre- and post-
processing, segmentation, feature extraction and classifier. The
pre-processing resizes the image that improves the speed
performance and removes the extra feature such as the noise
and fine hair. Post-processing enhances the image quality and
sharpens the outline of the cancer cell. Segmentation is used to
find the region of interest by removing the healthy skins from
the image and keep the cancer cell; while the feature extraction
decomposes the useful feature without prior clinical knowledge.
The curvelet properties and advantages have been
discussed in detail above. Principles of wrapping based discrete
curvelet transform have also been described. The classifier uses
neural networks to improve the predicted new image. However
in our paper the overall result, based on mixed type of images
database, is 55.94% for back-propagation neural network using
wavelet and 70.44 for using curvelet. This paper concludes that
there are some possible factors of low classification result. A
possible improvement and generalization is to deal with the
effect of large variation and overcome it by using larger
databases. For future direction we will develop the curvelet and
use other methods to increase accuracy of classification results.
REFERENCES
[1] C. C. Australia. (2010). Cancer Council to launch new
research / Failure to monitor highlights cancer risk. Available:
<http://www.cancer.org.au/cancersmartlifestyle/SunSmart/Skinc
ancerfactsandfigures.htm>
[2] J. Sikorski, "Identification of malignant melanoma by
wavelet analysis," in Student/Faculty Research Day, CSIS, Pace
University., 2004.
[3] H. T. Lau, Al-Jumaily, A., "Automatically Early
Detection of Skin Cancer: Study Based on Neural Network
Classification," presented at the International Conference of Soft
Computing and Pattern Recognition, 2009.
[4] J. He, et al., "A Joint Multiscale Algorithm with: Auto-
adapted Threshold for Image Denoising," Fifth International
Conference on Information Assurance and Security, 2009.
[5] E. J. Candes, Donoho, D.L., "The Curvelet Transform
for Image Denoising," IEEE Trans. Image Processing, vol. 6, pp.
670-684, 2002.
[6] J. S. Lim, "Two-Dimensional Signal and Image
Processing," Englewood Cliffs, 1990.
[7] P. Schmid, "Segmentation and symmetry measure for
image analysis: application to digital dermatoscopy," pp. 1-192,
1999.
[8] J. Tukey, W., "“Exploratory Data Analysis”," Addison-
Wesley, Reading, 1971.
[9] M. E. Celebi, et al., "Border detection in dermoscopy
images using statistical region merging," Skin Research and
Technology Journal, vol. 14, pp. 1-7, 2008.
[10] I. J. Sumana, "Image Retrieval Using Discrete Curvelet
Transform," Master of Information Technology Master of
Information Technology, Gippsland School of Information
Technology, Monash University, Australia, 2008.
[11] E. J. Cand`es, Donoho, D.L., Cohen, A.R.A.,
Schumaker, L.L., "Curvelets - a surprisingly e�ective
nonadaptive representation for objects with edges," Curves and
Surfaces, Vanderbilt University Press, pp. 105-120, 2000.
[12] E. J. Cand`es, Guo, F. , "New multiscale transforms,
minimum total variation synthesis: application to edge-
preserving image reconstruction.," Sig. Process., special issue on
Image and Video Coding Beyond Standards vol. 82, pp. 1519-
1543, 2002.
[13] E. J. Cand`es, Donoho, D.L., "New tight frames of
curvelets and optimal representations of objects with piecewise-
C 2 singularities," Comm. on Pure and Appl. Math, vol. 57, pp.
219-26, 2004.
[14] E. L. Candès, et al., "Fast Discrete Curvelet
Transforms," Multiscale Modeling and Simulation, vol. 5, pp.
861-899, 2005.
[15] M. J. Fadili, Starck, J.L., "Curvelets and Ridgelets,"
Encyclopedia of Complexity and System Science , in press,
2007.
[16] J. Emmanuel, et al., "Curvelets and Wave Equations,"
in Tutorial lecture notes, I. MGA Program, UCLA, Ed., 2004.
2011 11th International Conference on Hybrid Intelligent Systems (HIS) 685