6
2015 IEEE International Conference on Bioinformatics d Biomedicine (BTBM) Probabilistic Visual Search for Masses Within Mammography Images using Deep Leaing Mehmet Gnhan Ertosun, PhD Department of Radiology Stanford School of Medicine Stanford, CA USA gunha[email protected] Abstract- We developed a deep learning-based visual search system for the task of automated search and localization of masses in whole mammography images. The system consists of two modules: a classification engine and a localization engine. It first classifies mammograms as containing a mass or no mass using a deep learning classifier, and then localizes the mass(es) within the image using a regional probabilistic approach based on a deep learning network. We obtained 85% accuracy for the task of identifying images that contain a mass, and we were able to localize 85% of the masses at an average of 0.9 false positives per image. Our system has the advantages of being able to work with an entire mammography image as input without the need for image segmentation or other pre-processing steps, such as cropping or tiling the image, and it is based on deep learning with unsupervised feature discovery, so it does not require pre-defined and hand-craſted image features. Keywords- Mammography, Visual Search, CAD, Breast Cancer, Deep Learning, Classcation, Detection I. INTRODUCTION Breast cancer is the most common cancer among women, except for skin cancers, and breast cancer is the second leading cause of cancer death in women, exceeded only by lung cancer [1]. About 1 in 8 U.S. women (about 12%) will develop invasive breast cancer over the course of her lifetime, and in 2015, an estimated 231,840 new cases of invasive breast cancer are expected to be diagnosed in women in the U.S., along with 60,290 new cases of non-invasive (in situ) breast cancer [1]. Mammography screening has helped to increase the early-stage breast cancer detection rate [2], yet it has been shown that there is significant amount of variability in radiologists' mammography evaluations [3], [4]. A confounding factor is that breast density varies widely among women, and cancer is more difficult to detect in mammograms of women with radiographically dense breasts. Mammographic density refers to the relative lighess of a mammogram, determined by the number of x-ray photons that peneate the breast [5]. Breast density is a risk factor for missed cancers, and both false-positive and false-negative mammographic interpretations are more likely with dense breasts [6]. In a study of more than 11,000 women with no clinical symptoms of breast cancer, the sensitivity of mammography was only 48 percent for the subset of women with exemely dense breasts compared to 78 percent sensitivity for the entire sample of women in the study [7]. 978-1-4673-6799-8/15/$31.00 ©2015 IEEE 1310 Daniel L. Rubin, , MS Departments of Radiology and Medicine (Biomedical Informatics) Stanford School of Medicine Stanford, CA USA [email protected] Because using mm ography to detect cancer in nmal breast tissue is ndamentally a signal-to-noise exercise, it is particularly suited to CADe (Computer Aided Detection) technology [8]. Since factors such as radiologist fatigue and disaction, the complexity of breast sucture, and the subtle characteristics of early stage disease make intpreting mammograms challenging and conibute to both false-positive and false-negative results, the use of CADe with mammography becomes particularly atactive, offering experienced radiologists the option of a "double read." [5] A completely automated CADe system for mass detection was recently reported [9], consisting of three components, (1) Image segmentation (2) ROI characterization by means of textural features, and (3) ROI classification by means of a neural network. produced 4.23 false positives per image at 80% sensitivity for mass detection. The task of classiing mammograms and detecting masses using image features is not unlike other types of automated image feature machine leing problems in which a set of pre- defined features is used to characterize the image and predict the classification label and detect regions of interests, and the prior work on CADe in mm ography has in general been approached using pre-defmed features. However, a substantial disadvantage of pre-defmed features is the need to ow those features that are most informative in the classification and detection task. Oſten the best features are not own, and a method of automatic feature discovery could be advantageous, particularly if abundant data is available. Pre-defmed features are not only limited to those understood by humans, but they are formulated for specific tasks and they are expensive to produce. Instead of using pre-defmed features, one could l the most informative image features om the data itself using deep leing techniques, given that there is enough data. Deep leing has various closely related defmitions or high-level descriptions, one of which is: "A class of machine ling techniques, where many layers of information processing stages in hierarchical supervised architectures are exploited for unsupervised feature leing and for patte analysis/classification." [10] The essence of deep leing is to compute hierarchical features or representations of the observational data, where the higher-level features or factors

Probabilistic Visual Search for Masses Within …...mammography images using deep learning. As illustrated in Fig. 1, we fust c1assity input mammography images as containing a mass

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Probabilistic Visual Search for Masses Within …...mammography images using deep learning. As illustrated in Fig. 1, we fust c1assity input mammography images as containing a mass

2015 IEEE International Conference on Bioinformatics and Biomedicine (BTBM)

Probabilistic Visual Search for Masses Within

Mammography Images using Deep Learning

Mehmet Giinhan Ertosun, PhD Department of Radiology

Stanford School of Medicine Stanford, CA USA

[email protected]

Abstract- We developed a deep learning-based visual search system for the task of automated search and localization of masses in whole mammography images. The system consists of two modules: a classification engine and a localization engine. It first classifies mammograms as containing a mass or no mass using a deep learning classifier, and then localizes the mass(es) within the image using a regional probabilistic approach based on a deep learning network. We obtained 85% accuracy for the task of identifying images that contain a mass, and we were able to localize 85% of the masses at an average of 0.9 false positives per image. Our system has the advantages of being able to work with an entire mammography image as input without the need for image segmentation or other pre-processing steps, such as cropping or tiling the image, and it is based on deep learning with unsupervised feature discovery, so it does not require pre-defined and hand-crafted image features.

Keywords- Mammography, Visual Search, CAD, Breast Cancer, Deep Learning, Classification, Detection

I. INTRODUCTION

Breast cancer is the most common cancer among women, except for skin cancers, and breast cancer is the second leading cause of cancer death in women, exceeded only by lung cancer [1]. About 1 in 8 U.S. women (about 12%) will develop invasive breast cancer over the course of her lifetime, and in 2015, an estimated 231,840 new cases of invasive breast cancer are expected to be diagnosed in women in the U.S., along with 60,290 new cases of non-invasive (in situ) breast cancer [1]. Mammography screening has helped to increase the early-stage breast cancer detection rate [2], yet it has been shown that there is significant amount of variability in radiologists' mammography evaluations [3], [4].

A confounding factor is that breast density varies widely among women, and cancer is more difficult to detect in mammograms of women with radiographically dense breasts. Mammographic density refers to the relative lightness of a mammogram, determined by the number of x-ray photons that penetrate the breast [5]. Breast density is a risk factor for missed cancers, and both false-positive and false-negative mammographic interpretations are more likely with dense breasts [6]. In a study of more than 11,000 women with no clinical symptoms of breast cancer, the sensitivity of mammography was only 48 percent for the subset of women with extremely dense breasts compared to 78 percent sensitivity for the entire sample of women in the study [7].

978-1-4673-6799-8/15/$31.00 ©2015 IEEE 1310

Daniel L. Rubin, MD, MS Departments of Radiology and Medicine (Biomedical

Informatics) Stanford School of Medicine

Stanford, CA USA [email protected]

Because using mammography to detect cancer in normal breast tissue is fundamentally a signal-to-noise exercise, it is particularly suited to CADe (Computer Aided Detection) technology [8]. Since factors such as radiologist fatigue and distraction, the complexity of breast structure, and the subtle characteristics of early stage disease make interpreting mammograms challenging and contribute to both false-positive and false-negative results, the use of CADe with mammography becomes particularly attractive, offering experienced radiologists the option of a "double read." [5] A completely automated CADe system for mass detection was recently reported [9], consisting of three components, (1) Image segmentation (2) ROI characterization by means of textural features, and (3) ROI classification by means of a neural network. It produced 4.23 false positives per image at 80% sensitivity for mass detection.

The task of classifYing mammograms and detecting masses using image features is not unlike other types of automated image feature machine learning problems in which a set of pre­defined features is used to characterize the image and predict the classification label and detect regions of interests, and the prior work on CADe in mammography has in general been approached using pre-defmed features. However, a substantial disadvantage of pre-defmed features is the need to know those features that are most informative in the classification and detection task. Often the best features are not known, and a method of automatic feature discovery could be advantageous, particularly if abundant data is available. Pre-defmed features are not only limited to those understood by humans, but they are formulated for specific tasks and they are expensive to produce.

Instead of using pre-defmed features, one could learn the most informative image features from the data itself using deep learning techniques, given that there is enough data. Deep learning has various closely related defmitions or high-level descriptions, one of which is: "A class of machine learning techniques, where many layers of information processing stages in hierarchical supervised architectures are exploited for unsupervised feature learning and for pattern analysis/classification." [10] The essence of deep learning is to compute hierarchical features or representations of the observational data, where the higher-level features or factors

Page 2: Probabilistic Visual Search for Masses Within …...mammography images using deep learning. As illustrated in Fig. 1, we fust c1assity input mammography images as containing a mass

are defmed from lower-level ones. The family of deep learning methods has been growing increasingly richer, encompassing neural networks, hierarchical probabilistic models, and a variety of unsupervised and supervised feature learning algorithms [10].

Convolutional Neural Networks (CNNs or ConvNets), are a type of discriminative deep architecture in which layers include a convolutional layer and a pooling layer that are often stacked up to form a deep model [11]. The convolutional layer shares many weights, and the pooling layer subsamples the output of the convolutional layer and reduces the dimensionality of the data. The weight sharing in the convolutional layer, together with appropriately chosen pooling schemes, endows the CNN with some "invariance" properties (e.g., translation invariance). CNNs have been found to be highly effective in non-medical tasks, and they are commonly used in computer vision and image recognition [10]. CNNs have the advantage of automatically learning the appropriate image features, as opposed to traditional machine learning approaches that use hand-crafted features.

In this work, we created and evaluated a probabilistic visual search engine to look for and locate masses within mammography images using deep learning. As illustrated in Fig. 1, we fust c1assity input mammography images as containing a mass or not, and then we localize the masses within the images, and the output is a probabilistic map for masses within the image.

Fig.l. Probabilistic visual search engine for finding masses within mammography images

II. DATA & METHODS

A. Dataset

We used the Digital Database for Screening Mammography (DDSM) dataset [12], which includes images from approximately 2500 mammography studies, where each study includes two views of each breast, and having regions of interest (ROIs) labeled as mass or calcification. Our dataset contained mammograms where 2420 ROIs contained a mass, and the dataset was randomly divided into training/ validation/test splits with 80%/10%/10% split.

The input image size to the deep learning networks was selected as 256x256 pixels. Image patches were extracted from the entire image; those within a mass ROI were labeled as "mass," and those within non-ROI portions of the images were labeled as "non-mass." Since deep networks have many

1311

parameters, it was essential to make use of extensive data augmentation to prevent overfitting. This data augmentation included cropping, translation, rotation, flipping and scaling of the image tiles, and it was performed on the training set, resulting in 1.8 million samples.

B. Convolutional Neural Networks

CNNs are representatives of the multi-stage Hubel-Wiesel architecture, which extract local features at a high resolution and successively combine these into more complex features at lower resolutions. CNNs include two kinds of basic layers: convolutional layers (C layers), which resemble the simple cells, and pooling layers (P layers), which model the behavior of complex cells. Each convolutional layer performs a discrete 2D convolution operation on its source image with a filter kernel and applies a non-linear transfer function. The pooling layers reduce the size of the input by sununarizing neurons from a small spatial neighborhood [13].

The standard way to model a neuron's output f as a function of its input x is with f(x) = tanh (x) or f(x) = (1 + e-xrl. In terms of training time with gradient descent, these saturating nonlinearities are much slower than the non-saturating nonlinearity f(x) = max(O;x). The neurons with this nonlinearity are referred as Rectified Linear Units (ReLUs) [14]. Deep convolutional neural networks with ReLUs train several times faster than their equivalents with tanh units, [15] hence we used rectified linear units in this work.

The fmal layer of the CNN is fully connected to the preceding layer, which is a loss layer. The loss layer drives learning by comparing an output to a target and assigning a cost to be minimized. We use a softmax-Ioss as the loss layer, so the output of the fully-connected layer acts as input for the softmax classifier. The softmax loss layer computes the multinomial logistic loss of the softmax of its inputs, and it is conceptually identical to a softmax layer followed by a multinomial logistic loss layer, but provides a more numerically stable gradient [16].

C. Implementing and Training the CNNs

The CNNs are trained with the back-propagation algorithm [11]. We are using Caffe [16] for implementing and training the CNNs. Caffe trains models by the standard stochastic gradient descent algorithm. The CNNs are discriminatively trained via back-propagation through layers of convolutional filters and other operations such as rectification and pooling. Layers have two key responsibilities for the operation of the network as a whole: a forward pass that takes the inputs and produces the outputs, and a backward pass that takes the gradient with respect to the output, and computes the gradients with respect to the parameters and to the inputs, which are, in turn, back-propagated to earlier layers [17].

For the classification task we trained three different convolutional network architectures. For the localization task, to generate the output maps we made use of the classification module by transforming it into a fully convolutional network to generate a classification map for dense inference. This

Page 3: Probabilistic Visual Search for Masses Within …...mammography images using deep learning. As illustrated in Fig. 1, we fust c1assity input mammography images as containing a mass

conversion enables us to slide the original network across a larger image in a single forward pass, hence very efficiently.

D. Evaluation

We evaluated the accuracy of our classification task using the validation data set, and we evaluated the localization of masses using the test data set. For the classification task, we computed the accuracy of correctly c1assitying images as containing a mass. For the localization task, we analyzed the probability values that were assigned to the mass ground-truth ROI windows for each mass within the test data set. The mean and maximum values of the probabilities that were assigned to the mass ground-truth ROI window was compared against a probabilistic threshold value (i.e. a threshold value between 0 and 1), and a decision was made whether or not that particular mass was successfully localized depending on the comparison result. The percentage of the masses that were correctly localized (i.e. sensitivity) was computed.

III. RESULTS

A. CNN Architectures

We tried different CNN architectures that are known in the literature as AlexNet, VGG Net and GoogLeNeT. AlexNet [15] was the first work that popularized convolutional networks in computer vision [20]. It had a similar architecture basic as LeNet [11], but it was bigger, deeper and it had convolutional layers stacked on top of each other. AlexNet was the ImageNet ILSVRC challenge winner in 2012. GoogLeNet [18] was the winner of the ILSVRC in 2014. With the development of a module called Inception Module, it managed to dramatically reduce the number of parameters in the network (4M, compared to AlexNet with 60M) [18]. VGG Net [19] was the runner-up in the ILSVRC challenge in 2014. It demonstrated that the depth of the network is a critical component in achieving a good performance [20]. The VGG-16 network contains 16 convolution/fully-connected layers with an extremely homogeneous architecture, which only performs 3x3 convolutions and 2x2 pooling from the beginning to the end. Despite its slightly weaker classification performance, the VGG Net features outperform those of GoogLeNet in multiple transfer learning tasks [20]. A disadvantage of the VGG Net is that it is more expensive to evaluate and uses much more memory and parameters (140M) [20].

B. MasslNo-mass Classification Task

As a first step, we c1assity a mammography image as either containing at least one mass or having no mass. To achieve this goal, we trained these three CNN architectures on our data set for a mass/no-mass classification task. The validation accuracies achieved for this task is given in Table 1, along with number of layer and number of parameter values for each network architecture.

1312

TABLE I. DIFFERENT CNN ARCHITECTURES AND THE VALIDATION ACCURACIES OBTAINED FOR THE MAss/No-MAss CLASSIFICATION TASK

Number of Number of Validation Layers Parameters Accuracy

AlexNeT 8 60M 84%

VGG Net-16 16 140M 82%

GoogLeNeT 22 4M 85%

Fig 2 shows the validation accuracies for the mass/no-mass classification task during the training of the AlexNet.

(;'0.8 [::; __ �i!iiiiii�".jii . ..... jij·1Tji;"l �0.7r 0

o �O.6

.. >0.5 ° 3 4

Iteration

Fig. 2. Validation accuracies during the training of A1exNet

We achieved 84%, 85%, and %82 percent accuracies on the validation set for the mass/no-mass classification task with the AlexNet, GoogLeNet and VGG Net, respectively.

C. Mass Localitization Task

After identitying those images as containing at least one mass, we then seek to localize the mass (or masses, for the cases with multiple masses) within the image. To achieve this goal, after training the CNNs for the classification task (the one labeled as CNN-I in Fig. 1), we transformed one of the networks into a fully convolutional net (CNN-II in Fig. 1) to generate a probabilistic classification map for dense inference on the full images in an efficient manner. In this model, rather than outputting single classification label, the network outputs a classification map for the given input, and this classification map can be used for mass localization within the images. This also lets us visualize what portions of the image are leading to the classification of an input sample as having a mass, and provides a better understanding of the reasoning mechanism of the trained CNN. Providing such explanation of output results is particularly important in medical domains compared to other fields, since physician acceptance of decision support systems is enhanced when they provide explanation [21].

Fig 3 and Fig 4 show an example of a case when we feed an entire mammography image (including the side and view labels, such as the 'R' and 'CC' in Fig. 4(a)) into our search system. Fig 3 (a) shows the original mammography image with the mass marked with an ROI window, and (b) shows the version of it that is scaled appropriately to be fed into the network, and (c) shows the classification map as the output. We can clearly see that the output probability map identifies the region where the actual mass is located by assigning very high probabilities to that region. By this approach we not only can c1assity the image as containing a mass or not, but also locate the regions within the image that are identified as being mass based on high probabilities.

Page 4: Probabilistic Visual Search for Masses Within …...mammography images using deep learning. As illustrated in Fig. 1, we fust c1assity input mammography images as containing a mass

Fig. 3. (a) Mammography image with mass depicted within the ROI box (b) Scaled version of the mammography image that is input to the system ( c) Output showing the mass localization within the image

Fig. 4. Another example of a (a) mammography image (b) scaled version of the mammography image as input to the network (c) Output mass probability map

Fig 5 shows a case where we have multiple masses within the image, and the deep learning network successfully identifies all three mass regions.

Fig. 5. A case with multiple masses within one image

Dense breasts seem to be more of a challenging case to identity the lesions, which is also the case for human observers. Fig 6 shows a case where a small mass within the image is detected along with a false positive that has higher probability assignment. Further analysis of the wrongly identified region reveals that it surprisingly shares some of the visual characteristics of a mass. Also, it has to be noted that the networks are doing these classifications by only looking at one image, whereas radiologists have access to the images belonging to the multiple views of the same lesion, which helps them to reduce false positives. From a computer vision perspective, the wrongly identified region shares striking similarities to a mass.

1313

Fig. 6. An example with a denser breast where the ROI was detected with lower probabil ities compared to a non-annotated region that has a visual resemblance to a mass.

Dense breasts also cause a problem, where the background probabilities assigned to non-mass regions are higher compared to a non-dense breast tissue. Fig 7 shows an example of this case where the mass is correctly identified by a region of very high probability assignment, yet the background probabilities assigned to non-mass regions of dense breast tissue are also high, compared to cases in which breasts are not dense, as seen in Fig 3 and 4.

Fig.7. Background probabilities are more pronounced for the case of dense breasts

Fig 8 shows another case of a very dense breast where the network did fmd the location of the mass successfully, but Fig 9 shows a case where it was not able to identity a very small mass on a very dense breast, as the probabilities assigned to the mass are not significant due to high background density. That being said, this case is an example of one of the most challenging cases due to mass being small in size on a dense breast.

Fig.S. Example of a case that the system successfully located a mass on a very dense breast

Page 5: Probabilistic Visual Search for Masses Within …...mammography images using deep learning. As illustrated in Fig. 1, we fust c1assity input mammography images as containing a mass

Fig. 9 An example where the system failed to find a very small mass on a very dense breast

Fig 10 shows another case where a mass is located successfully on breast with dense tissue.

Fig 10. An example of a breast with dense tissue where mass is located successfully

It is important to notice that the ROI bounding boxes within the DDSM dataset unfortunately do not tightly circumscribe the masses (as it can be seen in Fig. 5). Fig. 11 shows the histogram of the mean ROI probabilities, i.e. the mean values of the probabilities that are assigned to the area labeled by the mass ROI window (i.e. ground truth window) for each case within the test data set. Due to the coarseness and the non­uniformity of the degree of coarseness of ROI bounding boxes within the dataset, the mean ROI probabilities have a dispersed di stributi on.

30 Histogram of Mean Values of ROI Probabilities

25 20 .

§ 15 o

v

Il-rh 10

IIII H1 °0�LUo�.2�� 0.4�� 0�.6��0�.8LLn� Mean value of the ROI probability

Fig.11. Mean values of the probabilities assigned to individual mass ROI windows across all cases within the test dataset

Fig. 12 shows the maximum values of the ROI probabilities, i.e. the maximum probability value that is assigned within a mass ROI window (i.e. ground truth

1314

window) for each case within the test data set. It can be seen that the system correctly assigned high probabilities for majority of the cases that contained a mass.

100 Histogram of Max Values of ROI Probabilities

80 � 60 o

V 40 20 00

Max value of the ROI probability

Fig. 12. Maximum values of the probabilities assigned to individual mass ROI windows across all cases within the test dataset

Fig 13 shows the scatter plot of mean and max values of the ROI probabilities (probabilities assigned to each mass ground­truth ROI window) and their histograms. It can be seen that some cases with low mean ROI probabilities indeed have high maximum ROI probabilities.

Fig. 13 Mean and max values of the ROI probabilities

When we use the maximum value of ROI probabilities as a decision criterion (i.e. the maximum value of the probabilities that are assigned within the ROI window is compared against this threshold value for each mass to be counted as being correctly located), Fig 14 shows the percentage of masses that are correctly found as function of the probability decision threshold value. When this decision threshold probability was set to 0.4, %85 of the masses were correctly located (i.e. sensitivity was 85%). The system produced an average of 0.9 false positives per image at 85% sensitivity for mass detection.

� 100 c---------------� � � 90

• .<: � ... III C 80 �o �.� o � 70 �.., E� 60 c ... �.= 50 • .<: � ... "'j; 40 L-______________ --.J 0.2 0.4 0.6 0.8

Decision Threshold

Fig. 14. Percentage of the masses that are found as a function of probabilistic decision threshold

Page 6: Probabilistic Visual Search for Masses Within …...mammography images using deep learning. As illustrated in Fig. 1, we fust c1assity input mammography images as containing a mass

IV. CONCLUSIONS

We developed a visual search engine conslstmg of classification and localization modules for the task of searching and localizing masses within mammography images. The first module classifies mammography images as containing a mass or not, and a second module localizes masses within mammography images that contain a mass. We used deep learning techniques with convolutional neural networks to undertake this task using unsupervised feature learning. By separating the classification and localization modules, we have the freedom of independently fine-tuning individual modules for different accuracy and sensitivity requirements, such as making localization module more sensitive so as to reduce the chances of missing a mass location. We could also use different CNN architectures for each individual module. Also, such a modular structure gives us the freedom of utilizing individual modules independently, if needed.

For the localization module, we also made use of the classification module by transforming it into a fully convolutional network to generate a probabilistic classification map for dense inference on large images in an efficient manner. This approach is advantageous in that it not only saves time during network training, but it also yields a very fast and one-step image localization by taking advantage of the internal structure of the network-we use the entire mammography image as input to the network and get the output map in one step, rather than creating multiple windowed images from one mammography image and feeding them into the network separately (such as using an external sliding-window or region­proposal approach). This approach also enables the system to provide some insight into the reasoning mechanisms of the network; deep learning networks can be criticized as being too much of a "black box" mechanism, especially for applications to medicine. Such visualization methods indeed help us understand what the network is doing and provide explanation regarding the single classification label it outputs. Providing such explanation of output results is particularly important in medical domains compared to other fields, since physician acceptance of decision support systems is enhanced when they provide explanation [21].

We obtained 85% accuracy for the task of identifying images that have a mass, and we were able to localize 85% of the masses when the probabilistic decision threshold was set to 0.4. The system produced an average of 0.9 false positives per image at 85% sensitivity for mass detection.

ACKNOWLEDGMENT

We acknowledge, and would like to thank Francisco Jimenez for scientific discussions on mammography, and Sidhanth Grover for his assistance in compiling the literature review on mammography. We

also would like to thank NVIDIA corporation for GPU donation. Mehmet Glinhan Ertosun is a fellow at Stanford Cancer Imaging

Training (SCIT) Program, and acknowledges its support (Nlli T32 CA009695). This work was supported in part by a grant from the

I3IS

National Cancer Institute, National Institutes of Health, UOICA190214.

REFERENCES

[I] http://www . breastcancer.orglsymptoms/understand _ bc!statistics

[2] C.R Smart, et aI., "Twenty-year follow-up of the breast cancers diagnosed during the Breast Cancer Detection Demonstration Project," CA A Cancer Journal for Clinicians, vol. 47, pp. 134-149, 1997.

[3] lG. Elmore, et al. "Variability in interpretive performance at screening mammography and radiologists' characteristics associated with accuracy." Radiology, vol. 253, pp. 641-651, 2009.

[4] W.E. Barlow, et al. "Accuracy of screening mammography interpretation by characteristics of radiologists." Journal Of The National Cancer Institute, vol. 96, pp. 1840-50, 2004.

[5] lE. Joy et al. "Saving Women's Lives: Strategies for Improving Breast Cancer Detection and Diagnosis", Institute of Medicine and NRC, ISBN: 0-309-53209-4, 2005

[6] S. W. Fletcher, JG Elmore. "Clinical practice. Mammographic screening for breast cancer". N Engl J Med. 2003;348(17):1672-1680.

[7] T.M. Kolb, et al. "Comparison of the performance of screening mammography, physical examination, and breast". Radiology. 2002;225( I): 165-175.

[8] C.l D'Orsi . "Computer-aided detection: there is no free lunch". Radiology. 200 I ;221(3):585-586.

[9] R Bellotti, et al. "A completely automated CAD system for mass detection in a large mammographic database.", Med Phys. 2006 Aug;33(8):3066-75.

[10] L. Deng, D. Yu, "Deep Learning: Methods and applications foundations and trends in signal processing", Now Publishers Incorporated, 2014

[II] Y. LeCun, et al. "Gradient-based learning applied to document recognition". Proceedings of the IEEE, 86:2278-2324, 1998.

[12] M. Heath , et al."The digital database for screening mammography", Proc. Int. Workshop Dig. Mammography, pp.212 -218 2000

[13] D. Scherer, A.Mliller, S. Behnke, "Evaluation of pooling operations in convolutional architectures for object recognition", Artificial Neural Networks-ICANN 2010, 92-101.

[14] V. Nair, G.E. Hinton, "Rectified linear units improve restricted boltzmann machines." In Proc. 27th Int. Conference on Machine Learning, 2010.

[15] A. Krizhevsky, et al,"Imagenet classification with deep convolutional neural networks, Advances in neural information processing systems, pp. 1097-1105, 2012.

[16] Caffe: Deep learning framework by the Berkeley Vision and Learni ng Center, http://caffe.berkeleyvi si on.orgl

[17] l Yangqing, et al. "Caffe: Convolutional Architecture for Fast Feature Embedding", Proc. of ACM Int. Con£ on Multimedia,pp 675-678, 2014

[18] C. Szegedy , et al. "Going deeper with convolutions." arXiv preprint arXiv: I 409.4842 (2014).

[19] K. Simonyan, A. Zisserman, "Very Deep Convolutional Networks for Large-Scale Image Recognition" arXiv:1409.1556

[20] A. Karpathy, "Convolutinoal Neural Networks for Visual Recognition" www.cs23In.stanford.edu

[21] RD. Shankar, et al. "Building an explanation function for a hypertension decision-support system." Stdy Health Tech. In£ 200 I ;84(Pt 1):538-42.